ppxyn1 commited on
Commit
438c474
·
verified ·
1 Parent(s): 0d5f369

End of training

Browse files
Files changed (1) hide show
  1. README.md +9 -10
README.md CHANGED
@@ -22,12 +22,12 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  This model is a fine-tuned version of [pyannote/speaker-diarization-3.0](https://huggingface.co/pyannote/speaker-diarization-3.0) on the test_data dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 0.2770
26
- - Model Preparation Time: 0.002
27
- - Der: 0.0744
28
  - False Alarm: 0.0
29
- - Missed Detection: 0.0017
30
- - Confusion: 0.0727
31
 
32
  ## Model description
33
 
@@ -52,16 +52,15 @@ The following hyperparameters were used during training:
52
  - seed: 42
53
  - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
54
  - lr_scheduler_type: cosine
55
- - num_epochs: 5
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | Der | False Alarm | Missed Detection | Confusion |
60
  |:-------------:|:-----:|:----:|:---------------:|:----------------------:|:------:|:-----------:|:----------------:|:---------:|
61
- | 0.06 | 1.0 | 307 | 0.0716 | 0.002 | 0.0136 | 0.0000 | 0.0019 | 0.0117 |
62
- | 0.0366 | 2.0 | 614 | 0.0778 | 0.002 | 0.0147 | 0.0000 | 0.0019 | 0.0127 |
63
- | 0.0453 | 3.0 | 921 | 0.0778 | 0.002 | 0.0136 | 0.0000 | 0.0019 | 0.0116 |
64
- | 0.0176 | 4.0 | 1228 | 0.0793 | 0.002 | 0.0138 | 0.0000 | 0.0019 | 0.0119 |
65
 
66
 
67
  ### Framework versions
 
22
 
23
  This model is a fine-tuned version of [pyannote/speaker-diarization-3.0](https://huggingface.co/pyannote/speaker-diarization-3.0) on the test_data dataset.
24
  It achieves the following results on the evaluation set:
25
+ - Loss: 0.0144
26
+ - Model Preparation Time: 0.0018
27
+ - Der: 0.0050
28
  - False Alarm: 0.0
29
+ - Missed Detection: 0.0011
30
+ - Confusion: 0.0039
31
 
32
  ## Model description
33
 
 
52
  - seed: 42
53
  - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
54
  - lr_scheduler_type: cosine
55
+ - num_epochs: 3
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | Der | False Alarm | Missed Detection | Confusion |
60
  |:-------------:|:-----:|:----:|:---------------:|:----------------------:|:------:|:-----------:|:----------------:|:---------:|
61
+ | 0.0449 | 1.0 | 366 | 0.0436 | 0.0018 | 0.0156 | 0.0000 | 0.0015 | 0.0141 |
62
+ | 0.0768 | 2.0 | 732 | 0.0432 | 0.0018 | 0.0119 | 0.0 | 0.0015 | 0.0103 |
63
+ | 0.02 | 3.0 | 1098 | 0.0375 | 0.0018 | 0.0108 | 0.0 | 0.0015 | 0.0093 |
 
64
 
65
 
66
  ### Framework versions