TiMauzi commited on
Commit
6401e67
·
verified ·
1 Parent(s): 6d693cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -0
README.md CHANGED
@@ -79,6 +79,10 @@ The model uses a sliding window approach to capture temporal patterns in musical
79
  - **MIDI-Only**: Limited to MIDI format; cannot process audio recordings or sheet music
80
  - **Cultural Bias**: Training data may reflect Western classical music traditions
81
 
 
 
 
 
82
  ### Recommendations for Use
83
  - Validate results with musicological expertise, especially for Classical period identification
84
  - Use confidence thresholds to filter low-confidence predictions
@@ -132,6 +136,7 @@ The following hyperparameters were used during training:
132
 
133
  ### Training results
134
 
 
135
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
136
  |:-------------:|:------:|:-----:|:---------------:|:--------:|:------:|
137
  | 1.2797 | 0.1031 | 2000 | 1.3522 | 0.4608 | 0.2486 |
@@ -184,8 +189,14 @@ The following hyperparameters were used during training:
184
  | 0.8264 | 4.9508 | 96000 | 1.1056 | 0.5736 | 0.4174 |
185
 
186
  ### Training Analysis
 
 
 
 
 
187
  The training shows stable convergence with the model reaching its best performance around step 44,000 (epoch 2.27). The training loss decreases steadily while validation metrics stabilize, indicating good generalization without severe overfitting. The model achieves its peak F1 score of 0.4299 at step 44,000, which was selected as the best checkpoint.
188
 
 
189
  ### Framework versions
190
 
191
  - Transformers 4.49.0
 
79
  - **MIDI-Only**: Limited to MIDI format; cannot process audio recordings or sheet music
80
  - **Cultural Bias**: Training data may reflect Western classical music traditions
81
 
82
+ Below is the confusion matrix for the best-performing checkpoint, visually highlighting these misclassifications (click to enlarge):
83
+
84
+ [<img src="confusion_matrix_best.png" alt="Confusion Matrix" width="500"/>](confusion_matrix_best.png)
85
+
86
  ### Recommendations for Use
87
  - Validate results with musicological expertise, especially for Classical period identification
88
  - Use confidence thresholds to filter low-confidence predictions
 
136
 
137
  ### Training results
138
 
139
+
140
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
141
  |:-------------:|:------:|:-----:|:---------------:|:--------:|:------:|
142
  | 1.2797 | 0.1031 | 2000 | 1.3522 | 0.4608 | 0.2486 |
 
189
  | 0.8264 | 4.9508 | 96000 | 1.1056 | 0.5736 | 0.4174 |
190
 
191
  ### Training Analysis
192
+
193
+ Below is the full training metrics plot, showing loss, accuracy, and F1-score trends over the entire training process (click to enlarge):
194
+
195
+ [<img src="training_metrics.png" alt="Training Metrics" width="500"/>](training_metrics.png)
196
+
197
  The training shows stable convergence with the model reaching its best performance around step 44,000 (epoch 2.27). The training loss decreases steadily while validation metrics stabilize, indicating good generalization without severe overfitting. The model achieves its peak F1 score of 0.4299 at step 44,000, which was selected as the best checkpoint.
198
 
199
+
200
  ### Framework versions
201
 
202
  - Transformers 4.49.0