MKCL commited on
Commit
930868a
·
verified ·
1 Parent(s): 661e1da

End of training

Browse files
README.md CHANGED
@@ -16,14 +16,19 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: nan
20
- - Score: 0.0
21
- - Counts: [0, 0, 0, 0]
22
- - Totals: [7173, 6173, 5173, 4173]
23
- - Precisions: [0.0, 0.0, 0.0, 0.0]
24
- - Bp: 0.1414
25
- - Sys Len: 7173
26
- - Ref Len: 21207
 
 
 
 
 
27
 
28
  ## Model description
29
 
@@ -43,8 +48,8 @@ More information needed
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 5e-05
46
- - train_batch_size: 4
47
- - eval_batch_size: 4
48
  - seed: 42
49
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
50
  - lr_scheduler_type: linear
@@ -52,15 +57,6 @@ The following hyperparameters were used during training:
52
  - mixed_precision_training: Native AMP
53
  - label_smoothing_factor: 0.1
54
 
55
- ### Training results
56
-
57
- | Training Loss | Epoch | Step | Validation Loss | Score | Counts | Totals | Precisions | Bp | Sys Len | Ref Len |
58
- |:-------------:|:-----:|:----:|:---------------:|:-----:|:------------:|:------------------------:|:--------------------:|:------:|:-------:|:-------:|
59
- | 0.0 | 1.0 | 2250 | nan | 0.0 | [0, 0, 0, 0] | [7173, 6173, 5173, 4173] | [0.0, 0.0, 0.0, 0.0] | 0.1414 | 7173 | 21207 |
60
- | 0.0 | 2.0 | 4500 | nan | 0.0 | [0, 0, 0, 0] | [7173, 6173, 5173, 4173] | [0.0, 0.0, 0.0, 0.0] | 0.1414 | 7173 | 21207 |
61
- | 0.0 | 3.0 | 6750 | nan | 0.0 | [0, 0, 0, 0] | [7173, 6173, 5173, 4173] | [0.0, 0.0, 0.0, 0.0] | 0.1414 | 7173 | 21207 |
62
-
63
-
64
  ### Framework versions
65
 
66
  - PEFT 0.15.2
 
16
 
17
  This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - eval_loss: nan
20
+ - eval_score: 0.0
21
+ - eval_counts: [0, 0, 0, 0]
22
+ - eval_totals: [1003, 3, 0, 0]
23
+ - eval_precisions: [0.0, 0.0, 0.0, 0.0]
24
+ - eval_bp: 0.0000
25
+ - eval_sys_len: 1003
26
+ - eval_ref_len: 20237
27
+ - eval_runtime: 35.2587
28
+ - eval_samples_per_second: 28.362
29
+ - eval_steps_per_second: 3.545
30
+ - epoch: 2.0
31
+ - step: 2250
32
 
33
  ## Model description
34
 
 
48
 
49
  The following hyperparameters were used during training:
50
  - learning_rate: 5e-05
51
+ - train_batch_size: 8
52
+ - eval_batch_size: 8
53
  - seed: 42
54
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
55
  - lr_scheduler_type: linear
 
57
  - mixed_precision_training: Native AMP
58
  - label_smoothing_factor: 0.1
59
 
 
 
 
 
 
 
 
 
 
60
  ### Framework versions
61
 
62
  - PEFT 0.15.2
adapter_config.json CHANGED
@@ -24,8 +24,10 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "v",
28
- "q"
 
 
29
  ],
30
  "task_type": "SEQ_2_SEQ_LM",
31
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "q",
28
+ "o",
29
+ "k",
30
+ "v"
31
  ],
32
  "task_type": "SEQ_2_SEQ_LM",
33
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a852c350090f71eae13d9d747c614c6f466d83ed563a0dc62a6dd8b6bff0ca8c
3
- size 1389456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27714901585c613bbac04b31f0c21d44ca7127e01db52e0d21f23fdce8079175
3
+ size 2779024
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9140ac7f6d23fee1e3e95bd60a767520bd27f4d6aa0a344b307be95da4abdf74
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf6360675e7abf87fa8f9a61ad128b4ea1d68558828af18deb49de9c0b07969b
3
  size 5368