Update README.md
Browse files
README.md
CHANGED
|
@@ -2,4 +2,16 @@
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- kurakurai/Luth-1.7B-Instruct
|
| 5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- kurakurai/Luth-1.7B-Instruct
|
| 5 |
+
language:
|
| 6 |
+
- fr
|
| 7 |
+
- en
|
| 8 |
+
pipeline_tag: text-generation
|
| 9 |
+
library_name: transformers
|
| 10 |
+
tags:
|
| 11 |
+
- text-generation-inference
|
| 12 |
+
---
|
| 13 |
+
|
| 14 |
+
# **Luth-Instruct-GGUF**
|
| 15 |
+
|
| 16 |
+
Luth-1.7B-Instruct is a French fine-tuned variant of the Qwen3-1.7B model, enhanced using the Luth-SFT dataset to significantly improve its capabilities in French instruction following, mathematics, and general knowledge while maintaining and even boosting its English performance. It was trained by full fine-tuning with Axolotl and later merged with the base Qwen3-1.7B, thus preserving its English competencies alongside marked improvements in French benchmarks. The model demonstrates strong performance on selected French and English benchmarks, including ifeval, gpqa-diamond, mmlu, math-500, arc-chall, and hellaswag, showing notable gains over comparable models in both languages. It is designed for tasks requiring bilingual proficiency with pronounced strength in French and is supported by available evaluation, training, and data scripts on GitHub. The model is suitable for instruction-following applications in contexts demanding enhanced French language understanding without compromising English language capabilities. It is openly accessible under an appropriate license for research and usage.
|
| 17 |
+
|