takuma commited on
Commit
7c822fe
·
1 Parent(s): 770e4f3

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 0.5770
18
 
19
  ## Model description
20
 
@@ -42,16 +42,17 @@ The following hyperparameters were used during training:
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.1
45
- - num_epochs: 3
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:----:|:---------------:|
52
- | 0.6364 | 1.0 | 1267 | 0.5770 |
53
- | 0.457 | 2.0 | 2534 | 0.5003 |
54
- | 0.3175 | 3.0 | 3801 | 0.5770 |
 
55
 
56
 
57
  ### Framework versions
 
14
 
15
  This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.7389
18
 
19
  ## Model description
20
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.1
45
+ - num_epochs: 5
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:----:|:---------------:|
52
+ | 0.6584 | 1.0 | 1267 | 0.5709 |
53
+ | 0.4877 | 2.0 | 2534 | 0.5579 |
54
+ | 0.3675 | 3.0 | 3801 | 0.5669 |
55
+ | 0.2323 | 4.0 | 5068 | 0.7389 |
56
 
57
 
58
  ### Framework versions