GiuLeo01 commited on
Commit
52e6c5b
·
verified ·
1 Parent(s): 5765c17

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -13,7 +13,7 @@ tags:
13
  <!-- Provide a quick summary of what the model is/does. -->
14
  This model is a prototype of a large language model specifically fine-tuned for Fortran90 code generation. It is based on the Qwen 2.5 Coder 3B Instruct model and has been refined using Supervised Fine-Tuning.
15
 
16
- **There is a more powerful version of this model, which has also been fine-tuned using Reinforcement Learning with Verifiable Rewards (via GRPO).**
17
 
18
  This model was fine-tuned briefly, without any human-labeled data and using only a single consumer GPU. Despite these clear constraints, the training process led to a 400% boost in performance on tasks involving simple to moderately complex fortran program generation (HumanEval-like). Compilation errors dropped as well, and the model now performs close to much larger general-purpose models that weren’t specifically trained for this task.
19
 
 
13
  <!-- Provide a quick summary of what the model is/does. -->
14
  This model is a prototype of a large language model specifically fine-tuned for Fortran90 code generation. It is based on the Qwen 2.5 Coder 3B Instruct model and has been refined using Supervised Fine-Tuning.
15
 
16
+ **There is a more powerful version of this model at https://huggingface.co/GiuLeo01/FortranCodeGen-3B-SynthData, which has also been fine-tuned using Reinforcement Learning with Verifiable Rewards (via GRPO).**
17
 
18
  This model was fine-tuned briefly, without any human-labeled data and using only a single consumer GPU. Despite these clear constraints, the training process led to a 400% boost in performance on tasks involving simple to moderately complex fortran program generation (HumanEval-like). Compilation errors dropped as well, and the model now performs close to much larger general-purpose models that weren’t specifically trained for this task.
19