Improve model card for QFFT model

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +26 -27
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
 
2
  library_name: transformers
3
  license: other
4
- base_model: Qwen/Qwen2.5-7B-Instruct
5
  tags:
6
  - llama-factory
7
  - full
@@ -9,22 +9,23 @@ tags:
9
  model-index:
10
  - name: 7b_isntruct_pretrain
11
  results: []
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- # 7b_isntruct_pretrain
18
 
19
- This model is a fine-tuned version of Qwen2.5-7B-Instruct on the S1_QFFT dataset.
20
 
21
  ## Model description
22
 
23
- More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
28
 
29
  ## Training and evaluation data
30
 
@@ -35,31 +36,29 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 1e-05
39
- - train_batch_size: 1
40
- - eval_batch_size: 8
41
- - seed: 42
42
- - distributed_type: multi-GPU
43
- - num_devices: 8
44
- - gradient_accumulation_steps: 1
45
- - total_train_batch_size: 8
46
- - total_eval_batch_size: 32
47
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
- - lr_scheduler_type: cosine
49
- - lr_scheduler_warmup_ratio: 0.1
50
- - num_epochs: 6
51
-
52
- ### Training results
53
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
 
 
55
 
56
  ### Framework versions
57
 
58
- - Transformers 4.48.2
59
- - Pytorch 2.6.0+cu124
60
- - Datasets 3.2.0
61
- - Tokenizers 0.21.0
62
-
63
 
64
  ## 📖 Citation
65
 
@@ -70,6 +69,6 @@ The following hyperparameters were used during training:
70
  year={2025},
71
  eprint={2506.12860},
72
  archivePrefix={arXiv},
73
- primaryClass={cs.CL},
74
  url={https://arxiv.org/abs/2506.12860},
75
- }
 
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-7B-Instruct
3
  library_name: transformers
4
  license: other
 
5
  tags:
6
  - llama-factory
7
  - full
 
9
  model-index:
10
  - name: 7b_isntruct_pretrain
11
  results: []
12
+ pipeline_tag: text-generation
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
+ # 7b\_isntruct\_pretrain
19
 
20
+ This model is a fine-tuned version of [Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the S1\_QFFT dataset. It was trained using the Question-Free Fine-Tuning (QFFT) method, as described in the paper [QFFT, Question-Free Fine-Tuning for Adaptive Reasoning](https://huggingface.co/papers/2506.12860).
21
 
22
  ## Model description
23
 
24
+ This model adapts a base LLM to leverage both concise and detailed reasoning patterns by fine-tuning it without input questions, learning directly from long chain-of-thought responses.
25
 
26
  ## Intended uses & limitations
27
 
28
+ This model is intended for research purposes in the domain of adaptive reasoning and chain-of-thought generation. It may exhibit improved efficiency in generating reasoning steps compared to standard fine-tuning but may also inherit limitations from the base model and the fine-tuning dataset.
29
 
30
  ## Training and evaluation data
31
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39
 
40
+ * learning\_rate: 1e-05
41
+ * train\_batch\_size: 1
42
+ * eval\_batch\_size: 8
43
+ * seed: 42
44
+ * distributed\_type: multi-GPU
45
+ * num\_devices: 8
46
+ * gradient\_accumulation\_steps: 1
47
+ * total\_train\_batch\_size: 8
48
+ * total\_eval\_batch\_size: 32
49
+ * optimizer: Use OptimizerNames.ADAMW\_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer\_args=No additional optimizer arguments
50
+ * lr\_scheduler\_type: cosine
51
+ * lr\_scheduler\_warmup\_ratio: 0.1
52
+ * num\_epochs: 6
53
 
54
+ ### Training results
55
 
56
  ### Framework versions
57
 
58
+ * Transformers 4.48.2
59
+ * Pytorch 2.6.0+cu124
60
+ * Datasets 3.2.0
61
+ * Tokenizers 0.21.0
 
62
 
63
  ## 📖 Citation
64
 
 
69
  year={2025},
70
  eprint={2506.12860},
71
  archivePrefix={arXiv},
 
72
  url={https://arxiv.org/abs/2506.12860},
73
+ }
74
+ ```