Update README.md
Browse files
README.md
CHANGED
|
@@ -13,7 +13,7 @@ tags:
|
|
| 13 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
This model is a prototype of a large language model specifically fine-tuned for Fortran90 code generation. It is based on the Qwen 2.5 Coder 3B Instruct model and has been refined using Supervised Fine-Tuning.
|
| 15 |
|
| 16 |
-
**There is a more powerful version of this model, which has also been fine-tuned using Reinforcement Learning with Verifiable Rewards (via GRPO).**
|
| 17 |
|
| 18 |
This model was fine-tuned briefly, without any human-labeled data and using only a single consumer GPU. Despite these clear constraints, the training process led to a 400% boost in performance on tasks involving simple to moderately complex fortran program generation (HumanEval-like). Compilation errors dropped as well, and the model now performs close to much larger general-purpose models that weren’t specifically trained for this task.
|
| 19 |
|
|
|
|
| 13 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
This model is a prototype of a large language model specifically fine-tuned for Fortran90 code generation. It is based on the Qwen 2.5 Coder 3B Instruct model and has been refined using Supervised Fine-Tuning.
|
| 15 |
|
| 16 |
+
**There is a more powerful version of this model at https://huggingface.co/GiuLeo01/FortranCodeGen-3B-SynthData, which has also been fine-tuned using Reinforcement Learning with Verifiable Rewards (via GRPO).**
|
| 17 |
|
| 18 |
This model was fine-tuned briefly, without any human-labeled data and using only a single consumer GPU. Despite these clear constraints, the training process led to a 400% boost in performance on tasks involving simple to moderately complex fortran program generation (HumanEval-like). Compilation errors dropped as well, and the model now performs close to much larger general-purpose models that weren’t specifically trained for this task.
|
| 19 |
|