cotysong113 commited on
Commit
bb4b1cb
·
verified ·
1 Parent(s): c76d307

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,5 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
  base_model: distilbert-base-uncased
5
  tags:
6
  - generated_from_trainer
@@ -16,8 +15,8 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 2.4886
20
- - Model Preparation Time: 0.0007
21
 
22
  ## Model description
23
 
@@ -43,20 +42,19 @@ The following hyperparameters were used during training:
43
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
44
  - lr_scheduler_type: linear
45
  - num_epochs: 3.0
46
- - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time |
51
  |:-------------:|:-----:|:----:|:---------------:|:----------------------:|
52
- | 2.6843 | 1.0 | 157 | 2.4970 | 0.0007 |
53
- | 2.5869 | 2.0 | 314 | 2.4490 | 0.0007 |
54
- | 2.5266 | 3.0 | 471 | 2.4825 | 0.0007 |
55
 
56
 
57
  ### Framework versions
58
 
59
- - Transformers 4.46.2
60
- - Pytorch 2.5.1+cu124
61
  - Datasets 3.1.0
62
- - Tokenizers 0.20.1
 
1
  ---
2
  library_name: transformers
 
3
  base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
 
15
 
16
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 2.4880
19
+ - Model Preparation Time: 0.0009
20
 
21
  ## Model description
22
 
 
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 3.0
 
45
 
46
  ### Training results
47
 
48
  | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time |
49
  |:-------------:|:-----:|:----:|:---------------:|:----------------------:|
50
+ | 2.6548 | 1.0 | 157 | 2.4900 | 0.0009 |
51
+ | 2.5814 | 2.0 | 314 | 2.4454 | 0.0009 |
52
+ | 2.5231 | 3.0 | 471 | 2.4799 | 0.0009 |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.47.0
58
+ - Pytorch 2.5.1
59
  - Datasets 3.1.0
60
+ - Tokenizers 0.21.0
runs/Dec10_14-48-22_bogon/events.out.tfevents.1733813323.bogon.36640.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d632783cc5bf36fca69e7ee41c3832370aaeb220aa43e861447a68f1978fa14a
3
- size 6572
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e9a256bb9d09f0baba95311c675381f2342dc6a845e7220f77278a51fb992e6
3
+ size 7263
runs/Dec10_14-48-22_bogon/events.out.tfevents.1733813720.bogon.36640.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:400c51f1916da29c579632250c40bbcdfadb23ccf9ba5940d22e689a3e96a6c8
3
+ size 425