Update README.md
Browse files
README.md
CHANGED
|
@@ -14,7 +14,7 @@ model_name: "Italia-GPT 🇮🇹"
|
|
| 14 |
|
| 15 |
# Italia-GPT <img src="https://flagcdn.com/w40/it.png" width="35" style="display: inline; vertical-align: middle; margin-bottom: 44px;">
|
| 16 |
|
| 17 |
-
**Italia-GPT** is a
|
| 18 |
|
| 19 |

|
| 20 |

|
|
@@ -29,10 +29,10 @@ Below are the target benchmarks for the **CALAMITA** and **Evalita-LLM** framewo
|
|
| 29 |
|
| 30 |
| Metric | Score | Description |
|
| 31 |
| :--- | :--- | :--- |
|
| 32 |
-
| **Logic & Reasoning** | **
|
| 33 |
-
| **Grammar Accuracy** | **
|
| 34 |
-
| **Sentiment (ITA)** | **
|
| 35 |
-
| **Cultural Q&A** | **
|
| 36 |
|
| 37 |
---
|
| 38 |
|
|
@@ -66,4 +66,5 @@ messages = [{"role": "user", "content": "Come si prepara una vera carbonara?"}]
|
|
| 66 |
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to("cuda")
|
| 67 |
outputs = model.generate(inputs, max_new_tokens=256)
|
| 68 |
|
| 69 |
-
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|
|
|
|
|
| 14 |
|
| 15 |
# Italia-GPT <img src="https://flagcdn.com/w40/it.png" width="35" style="display: inline; vertical-align: middle; margin-bottom: 44px;">
|
| 16 |
|
| 17 |
+
**Italia-GPT** is a 1.2B parameter model fine-tuned for native Italian instruction following. By focusing on linguistic nuances and cultural context, this model provides superior fluency compared to standard base models.
|
| 18 |
|
| 19 |

|
| 20 |

|
|
|
|
| 29 |
|
| 30 |
| Metric | Score | Description |
|
| 31 |
| :--- | :--- | :--- |
|
| 32 |
+
| **Logic & Reasoning** | **%** | Native Italian sentence logic |
|
| 33 |
+
| **Grammar Accuracy** | **%** | Gender/Number agreement precision |
|
| 34 |
+
| **Sentiment (ITA)** | **%** | Detection of Italian irony and tone |
|
| 35 |
+
| **Cultural Q&A** | **%** | Localized knowledge and trivia |
|
| 36 |
|
| 37 |
---
|
| 38 |
|
|
|
|
| 66 |
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to("cuda")
|
| 67 |
outputs = model.generate(inputs, max_new_tokens=256)
|
| 68 |
|
| 69 |
+
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 70 |
+
```
|