Update README.md
Browse files
README.md
CHANGED
|
@@ -79,6 +79,10 @@ The model uses a sliding window approach to capture temporal patterns in musical
|
|
| 79 |
- **MIDI-Only**: Limited to MIDI format; cannot process audio recordings or sheet music
|
| 80 |
- **Cultural Bias**: Training data may reflect Western classical music traditions
|
| 81 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 82 |
### Recommendations for Use
|
| 83 |
- Validate results with musicological expertise, especially for Classical period identification
|
| 84 |
- Use confidence thresholds to filter low-confidence predictions
|
|
@@ -132,6 +136,7 @@ The following hyperparameters were used during training:
|
|
| 132 |
|
| 133 |
### Training results
|
| 134 |
|
|
|
|
| 135 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|
| 136 |
|:-------------:|:------:|:-----:|:---------------:|:--------:|:------:|
|
| 137 |
| 1.2797 | 0.1031 | 2000 | 1.3522 | 0.4608 | 0.2486 |
|
|
@@ -184,8 +189,14 @@ The following hyperparameters were used during training:
|
|
| 184 |
| 0.8264 | 4.9508 | 96000 | 1.1056 | 0.5736 | 0.4174 |
|
| 185 |
|
| 186 |
### Training Analysis
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 187 |
The training shows stable convergence with the model reaching its best performance around step 44,000 (epoch 2.27). The training loss decreases steadily while validation metrics stabilize, indicating good generalization without severe overfitting. The model achieves its peak F1 score of 0.4299 at step 44,000, which was selected as the best checkpoint.
|
| 188 |
|
|
|
|
| 189 |
### Framework versions
|
| 190 |
|
| 191 |
- Transformers 4.49.0
|
|
|
|
| 79 |
- **MIDI-Only**: Limited to MIDI format; cannot process audio recordings or sheet music
|
| 80 |
- **Cultural Bias**: Training data may reflect Western classical music traditions
|
| 81 |
|
| 82 |
+
Below is the confusion matrix for the best-performing checkpoint, visually highlighting these misclassifications (click to enlarge):
|
| 83 |
+
|
| 84 |
+
[<img src="confusion_matrix_best.png" alt="Confusion Matrix" width="500"/>](confusion_matrix_best.png)
|
| 85 |
+
|
| 86 |
### Recommendations for Use
|
| 87 |
- Validate results with musicological expertise, especially for Classical period identification
|
| 88 |
- Use confidence thresholds to filter low-confidence predictions
|
|
|
|
| 136 |
|
| 137 |
### Training results
|
| 138 |
|
| 139 |
+
|
| 140 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|
| 141 |
|:-------------:|:------:|:-----:|:---------------:|:--------:|:------:|
|
| 142 |
| 1.2797 | 0.1031 | 2000 | 1.3522 | 0.4608 | 0.2486 |
|
|
|
|
| 189 |
| 0.8264 | 4.9508 | 96000 | 1.1056 | 0.5736 | 0.4174 |
|
| 190 |
|
| 191 |
### Training Analysis
|
| 192 |
+
|
| 193 |
+
Below is the full training metrics plot, showing loss, accuracy, and F1-score trends over the entire training process (click to enlarge):
|
| 194 |
+
|
| 195 |
+
[<img src="training_metrics.png" alt="Training Metrics" width="500"/>](training_metrics.png)
|
| 196 |
+
|
| 197 |
The training shows stable convergence with the model reaching its best performance around step 44,000 (epoch 2.27). The training loss decreases steadily while validation metrics stabilize, indicating good generalization without severe overfitting. The model achieves its peak F1 score of 0.4299 at step 44,000, which was selected as the best checkpoint.
|
| 198 |
|
| 199 |
+
|
| 200 |
### Framework versions
|
| 201 |
|
| 202 |
- Transformers 4.49.0
|