| # XNLI Base Model | |
| This model was trained on the XNLI dataset using random data selection. | |
| ## Training Parameters | |
| - **Dataset**: XNLI | |
| - **Mode**: Base | |
| - **Selection Method**: Random | |
| - **Train Size**: 2400 examples | |
| - **Epochs**: 8 | |
| - **Batch Size**: 16 | |
| - **Effective Batch Size**: 64 (batch_size * gradient_accumulation_steps) | |
| - **Learning Rate**: 1e-05 | |
| - **Patience**: 6 | |
| - **Max Length**: 256 | |
| - **Gradient Accumulation Steps**: 4 | |
| - **Warmup Ratio**: 0.1 | |
| - **Weight Decay**: 0.01 | |
| - **Optimizer**: AdamW | |
| - **Scheduler**: cosine_with_warmup | |
| - **Random Seed**: 42 | |
| ## Performance | |
| - **Overall Accuracy**: 65.47% | |
| - **Overall Loss**: 0.0141 | |
| ### Language-Specific Performance | |
| - **English (EN)**: 72.22% | |
| - **German (DE)**: 67.60% | |
| - **Arabic (AR)**: 63.21% | |
| - **Spanish (ES)**: 68.72% | |
| - **Hindi (HI)**: 62.04% | |
| - **Swahili (SW)**: 59.00% | |
| ## Model Information | |
| - **Base Model**: bert-base-multilingual-cased | |
| - **Task**: Natural Language Inference | |
| - **Languages**: 6 languages (EN, DE, AR, ES, HI, SW) |