gsmyrnis commited on
Commit
02daf28
·
verified ·
1 Parent(s): ac7bae4

Model save

Browse files
README.md CHANGED
@@ -1,10 +1,9 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: Qwen/Qwen2.5-1.5B-Instruct
5
  tags:
6
  - llama-factory
7
- - full
8
  - generated_from_trainer
9
  model-index:
10
  - name: qwen2-5_multiple_samples_ground_truth_openr1_llm_verifier_clean
@@ -16,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # qwen2-5_multiple_samples_ground_truth_openr1_llm_verifier_clean
18
 
19
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the mlfoundations-dev/instruction_filtering_scale_up_code_base_gemini_length_1K dataset.
20
 
21
  ## Model description
22
 
@@ -40,8 +39,10 @@ The following hyperparameters were used during training:
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - distributed_type: multi-GPU
 
43
  - gradient_accumulation_steps: 3
44
- - total_train_batch_size: 3
 
45
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_ratio: 0.1
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: Qwen/Qwen2.5-7B-Instruct
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: qwen2-5_multiple_samples_ground_truth_openr1_llm_verifier_clean
 
15
 
16
  # qwen2-5_multiple_samples_ground_truth_openr1_llm_verifier_clean
17
 
18
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on an unknown dataset.
19
 
20
  ## Model description
21
 
 
39
  - eval_batch_size: 8
40
  - seed: 42
41
  - distributed_type: multi-GPU
42
+ - num_devices: 32
43
  - gradient_accumulation_steps: 3
44
+ - total_train_batch_size: 96
45
+ - total_eval_batch_size: 256
46
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: cosine
48
  - lr_scheduler_warmup_ratio: 0.1
generation_config.json CHANGED
@@ -6,7 +6,7 @@
6
  151643
7
  ],
8
  "pad_token_id": 151643,
9
- "repetition_penalty": 1.1,
10
  "temperature": 0.7,
11
  "top_k": 20,
12
  "top_p": 0.8,
 
6
  151643
7
  ],
8
  "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
  "temperature": 0.7,
11
  "top_k": 20,
12
  "top_p": 0.8,
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76304d214fae0a740bb4fbc3c23abb9300610273f5b3d5ca132dcce44092b64f
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ae1634941827c62065f475455180268fe186cd4e4b8b88358f597c1221e980e
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:079bfb3f0ed47587461c74f3be997fcc232d133f9476971ec76d71f7a98e25ea
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddc8bab0965783d6373234ebda9eaec1fa1b9cdc96ef53a1e2e788dbe1f88b31
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73745049d80771cf6f864bceb36015477879a88a966bc6b47906479f79b34525
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0d72b343eb118b97c15e46376b6f14a373380e4c8532fa92091c50fdc306a5b
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a8d94aa13d1614a7b4523b345cde3fcce160cd5190a9d5deb809a3a69d7557c
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108b6c338d3c55d8ebe18adb23f40f05dd52862f828a525c5406abc68bc21de3
3
  size 1089994880