Update README.md
Browse files
README.md
CHANGED
|
@@ -21,9 +21,11 @@ Single A100 was used for fine-tuning and evaluation.
|
|
| 21 |
|
| 22 |
The [TRL](https://github.com/huggingface/trl) library was used with SFT/full-rank options:
|
| 23 |
|
|
|
|
| 24 |
python trl/scripts/sft.py --model_name_or_path Qwen/Qwen3-0.6B --dataset_name openai/gsm8k --dataset_config main --learning_rate 2e-5 \
|
| 25 |
--num_train_epochs 1 --per_device_train_batch_size 2 --gradient_checkpointing --eos_token '<|im_end|>' --eval_strategy steps \
|
| 26 |
--eval_steps 100 --completion_only_loss True --report_to wandb --output_dir /path/to/the/finetuned/model
|
|
|
|
| 27 |
|
| 28 |
The dataset was preprocessed to the conversational format:
|
| 29 |
|
|
@@ -47,8 +49,10 @@ dataset = dataset.map(preprocess_function)
|
|
| 47 |
|
| 48 |
Evaluation was done with lm_eval on the test split of gsm8k:
|
| 49 |
|
| 50 |
-
|
| 51 |
-
|
|
|
|
|
|
|
| 52 |
|
| 53 |
### Results
|
| 54 |
|
|
|
|
| 21 |
|
| 22 |
The [TRL](https://github.com/huggingface/trl) library was used with SFT/full-rank options:
|
| 23 |
|
| 24 |
+
```bash
|
| 25 |
python trl/scripts/sft.py --model_name_or_path Qwen/Qwen3-0.6B --dataset_name openai/gsm8k --dataset_config main --learning_rate 2e-5 \
|
| 26 |
--num_train_epochs 1 --per_device_train_batch_size 2 --gradient_checkpointing --eos_token '<|im_end|>' --eval_strategy steps \
|
| 27 |
--eval_steps 100 --completion_only_loss True --report_to wandb --output_dir /path/to/the/finetuned/model
|
| 28 |
+
```
|
| 29 |
|
| 30 |
The dataset was preprocessed to the conversational format:
|
| 31 |
|
|
|
|
| 49 |
|
| 50 |
Evaluation was done with lm_eval on the test split of gsm8k:
|
| 51 |
|
| 52 |
+
```bash
|
| 53 |
+
python -m lm_eval --model vllm --model_args pretrained=${model},tensor_parallel_size=1,dtype=auto,gpu_memory_utilization=0.9,data_parallel_size=1 \
|
| 54 |
+
--tasks gsm8k --batch_size 1 --apply_chat_template=True --confirm_run_unsafe_code --trust_remote_code
|
| 55 |
+
```
|
| 56 |
|
| 57 |
### Results
|
| 58 |
|