Italia-GPT / README.md
XythicK's picture
Update README.md
c7718e6 verified
metadata
language:
  - it
license: llama3.2
base_model: meta-llama/Llama-3.2-1B-Instruct
tags:
  - llama-3.2
  - italian
  - sft
  - text-generation
  - safetensors
model_name: Italia-GPT ๐Ÿ‡ฎ๐Ÿ‡น

Italia-GPT

Italia-GPT is a 1.2B parameter model fine-tuned for native Italian instruction following. By focusing on linguistic nuances and cultural context, this model provides superior fluency compared to standard base models.

Model Card Model Size


๐Ÿ’Ž Performance Overview

Below are the target benchmarks for the CALAMITA and Evalita-LLM frameworks:

Metric Score Description
Logic & Reasoning % Native Italian sentence logic
Grammar Accuracy % Gender/Number agreement precision
Sentiment (ITA) % Detection of Italian irony and tone
Cultural Q&A % Localized knowledge and trivia

๐Ÿ›  Technical Specifications

  • Base Architecture: Llama 3.2
  • Precision: BFloat16 ($BF16$)
  • Weights: Merged Safetensors (Standalone)
  • Language Support: Primary: Italian ๐Ÿ‡ฎ๐Ÿ‡น, Secondary: English ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿš€ Usage Guide

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "XythicK/Italia-GPT"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id, 
    torch_dtype=torch.bfloat16, 
    device_map="auto"
)

# Native Italian Chat Template
messages = [{"role": "user", "content": "Come si prepara una vera carbonara?"}]
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to("cuda")
outputs = model.generate(inputs, max_new_tokens=256)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))