vpadaraju commited on
Commit
b3f2c4a
·
verified ·
1 Parent(s): 83777fd

End of training

Browse files
Files changed (1) hide show
  1. README.md +11 -7
README.md CHANGED
@@ -4,6 +4,8 @@ license: mit
4
  base_model: microsoft/DialoGPT-small
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: testing
9
  results: []
@@ -14,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # testing
16
 
17
- This model is a fine-tuned version of [microsoft/DialoGPT-small](https://huggingface.co/microsoft/DialoGPT-small) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 2.8036
20
 
21
  ## Model description
22
 
@@ -41,18 +43,20 @@ The following hyperparameters were used during training:
41
  - seed: 42
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
- - training_steps: 10
45
 
46
  ### Training results
47
 
48
- | Training Loss | Epoch | Step | Validation Loss |
49
- |:-------------:|:------:|:----:|:---------------:|
50
- | No log | 0.3846 | 5 | 3.8458 |
51
- | No log | 0.7692 | 10 | 2.8036 |
 
52
 
53
 
54
  ### Framework versions
55
 
56
  - Transformers 4.48.2
57
  - Pytorch 2.5.1+cu124
 
58
  - Tokenizers 0.21.0
 
4
  base_model: microsoft/DialoGPT-small
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - empathetic_dialogues
9
  model-index:
10
  - name: testing
11
  results: []
 
16
 
17
  # testing
18
 
19
+ This model is a fine-tuned version of [microsoft/DialoGPT-small](https://huggingface.co/microsoft/DialoGPT-small) on the empathetic_dialogues dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.7011
22
 
23
  ## Model description
24
 
 
43
  - seed: 42
44
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
+ - num_epochs: 3
47
 
48
  ### Training results
49
 
50
+ | Training Loss | Epoch | Step | Validation Loss |
51
+ |:-------------:|:-----:|:----:|:---------------:|
52
+ | No log | 1.0 | 13 | 1.8261 |
53
+ | No log | 2.0 | 26 | 1.7528 |
54
+ | No log | 3.0 | 39 | 1.7011 |
55
 
56
 
57
  ### Framework versions
58
 
59
  - Transformers 4.48.2
60
  - Pytorch 2.5.1+cu124
61
+ - Datasets 3.2.0
62
  - Tokenizers 0.21.0