cam-1000 commited on
Commit
fa6b2e1
·
verified ·
1 Parent(s): 54e0d1e

End of training

Browse files
Files changed (4) hide show
  1. README.md +6 -7
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,7 +1,6 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
- base_model: Qwen/Qwen3-0.6B-Base
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,9 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # MNLP_M3_rag_model
16
 
17
- This model is a fine-tuned version of [Qwen/Qwen3-0.6B-Base](https://huggingface.co/Qwen/Qwen3-0.6B-Base) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.0000
20
 
21
  ## Model description
22
 
@@ -50,9 +49,9 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:------:|:----:|:---------------:|
53
- | 0.0 | 0.0228 | 100 | 0.0000 |
54
- | 0.0 | 0.0456 | 200 | 0.0000 |
55
- | 0.0 | 0.0683 | 300 | 0.0000 |
56
 
57
 
58
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
+ base_model: cam-1000/MNLP_M3_mcqa_model
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # MNLP_M3_rag_model
15
 
16
+ This model is a fine-tuned version of [cam-1000/MNLP_M3_mcqa_model](https://huggingface.co/cam-1000/MNLP_M3_mcqa_model) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.8564
19
 
20
  ## Model description
21
 
 
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:------:|:----:|:---------------:|
52
+ | 0.8275 | 0.0228 | 100 | 0.8105 |
53
+ | 0.7663 | 0.0456 | 200 | 0.8251 |
54
+ | 0.5896 | 0.0683 | 300 | 0.8564 |
55
 
56
 
57
  ### Framework versions
config.json CHANGED
@@ -3,7 +3,7 @@
3
  "Qwen3ForCausalLM"
4
  ],
5
  "attention_bias": false,
6
- "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
  "eos_token_id": 151643,
9
  "head_dim": 128,
 
3
  "Qwen3ForCausalLM"
4
  ],
5
  "attention_bias": false,
6
+ "attention_dropout": 0.1,
7
  "bos_token_id": 151643,
8
  "eos_token_id": 151643,
9
  "head_dim": 128,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95a855ef063a62a7db7a084ba0731f83f8ba671fcad58e393c6fba2264e021d6
3
  size 1192135096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c3fef9685203b9f660eacccb274de06644cda9c54f59e6a08b88dce69e363af
3
  size 1192135096
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb53288efc2ba5bed71f1270d0cc845cac5b26653e6a53f2850b42035e6acecd
3
  size 5713
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62eb3b5582a5d668757dda6894f7cdaaa74fa00eb64371185180a22602e99425
3
  size 5713