Samuael commited on
Commit
77217ab
·
verified ·
1 Parent(s): d2c8b01

Samuael/amBART

Browse files
Files changed (4) hide show
  1. README.md +14 -15
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Samuael/amBART_1000
3
  tags:
4
  - generated_from_trainer
5
  model-index:
@@ -12,17 +12,17 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # amBART
14
 
15
- This model is a fine-tuned version of [Samuael/amBART_1000](https://huggingface.co/Samuael/amBART_1000) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - eval_loss: 2.9935
18
- - eval_wer: 0.9817
19
- - eval_cer: 0.7707
20
- - eval_bleu: 0.7418
21
- - eval_runtime: 9.129
22
- - eval_samples_per_second: 39.325
23
- - eval_steps_per_second: 0.329
24
- - epoch: 3.0
25
- - step: 2127
26
 
27
  ## Model description
28
 
@@ -41,14 +41,13 @@ More information needed
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
- - learning_rate: 0.002
45
- - train_batch_size: 128
46
- - eval_batch_size: 128
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
  - num_epochs: 50
51
- - mixed_precision_training: Native AMP
52
 
53
  ### Framework versions
54
 
 
1
  ---
2
+ base_model: Samuael/amBART
3
  tags:
4
  - generated_from_trainer
5
  model-index:
 
12
 
13
  # amBART
14
 
15
+ This model is a fine-tuned version of [Samuael/amBART](https://huggingface.co/Samuael/amBART) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - eval_loss: 3.4340
18
+ - eval_wer: 0.9966
19
+ - eval_cer: 0.8055
20
+ - eval_bleu: 0.1833
21
+ - eval_runtime: 8.1414
22
+ - eval_samples_per_second: 44.096
23
+ - eval_steps_per_second: 0.246
24
+ - epoch: 11.0
25
+ - step: 3905
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 0.0002
45
+ - train_batch_size: 256
46
+ - eval_batch_size: 256
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
  - num_epochs: 50
 
51
 
52
  ### Framework versions
53
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Samuael/amBART_1000",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "architectures": [
 
1
  {
2
+ "_name_or_path": "Samuael/amBART",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "architectures": [
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1504bcd0758ac82ab09f78650ebc0c7c431edde67e78c5fb92b4befcf30c490e
3
  size 180813204
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e48b1ce78fe21170b4789afb20904bfd8f3b41f68909381b4fb2c806011a1255
3
  size 180813204
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f19be77b3b5f0472323533764679e1f13184f30ee05c97f781d3b916fe948fbe
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89e3c4eea5f7f0e97067c89e15fc1896956960907cf92ea9ab7b5f40a6a93afe
3
  size 4984