| ====================================================================== | |
| MODEL 4.2: XLM-RoBERTa with Data Augmentation | |
| ====================================================================== | |
| Date: 2025-12-25 16:18:05 | |
| SPECIFICATION: | |
| Model: xlm-roberta-large | |
| Batch Size: 8 (Grad Accum: 2) | |
| Learning Rate: 1e-05 | |
| Max Length: 128 | |
| Epochs: 3/5 | |
| Seed: 42 | |
| Augmentation: Yes (Target minority 5%) | |
| DATA AUGMENTATION: | |
| Original train size: 15699 | |
| Augmented train size: 16156 | |
| Samples added: 457 | |
| TEST RESULTS: | |
| Macro F1: 0.6822 | |
| Weighted F1: 0.7843 | |
| Accuracy: 0.7832 | |
| TRAINING TIME: | |
| Total: 81.88 minutes | |
| Per epoch: 982.5s | |