Roberta_combo_v1_med_test
This model is a fine-tuned version of FacebookAI/xlm-roberta-base on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.2583
- Accuracy: 0.946
- Auc: 0.987
- Precision: 0.942
- Recall: 0.947
- F1: 0.944
- F1-macro: 0.946
- F1-micro: 0.946
- F1-weighted: 0.946
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 5
- mixed_precision_training: Native AMP
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Auc | Precision | Recall | F1 | F1-macro | F1-micro | F1-weighted |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 0.562 | 0.1028 | 500 | 0.2457 | 0.912 | 0.973 | 0.892 | 0.93 | 0.911 | 0.912 | 0.912 | 0.912 |
| 0.4586 | 0.2057 | 1000 | 0.3618 | 0.872 | 0.932 | 0.897 | 0.83 | 0.862 | 0.871 | 0.872 | 0.872 |
| 0.4422 | 0.3085 | 1500 | 0.1702 | 0.944 | 0.98 | 0.936 | 0.948 | 0.942 | 0.944 | 0.944 | 0.944 |
| 0.4245 | 0.4114 | 2000 | 0.3369 | 0.891 | 0.98 | 0.822 | 0.987 | 0.897 | 0.89 | 0.891 | 0.89 |
| 0.4188 | 0.5142 | 2500 | 0.2520 | 0.917 | 0.962 | 0.905 | 0.924 | 0.915 | 0.917 | 0.917 | 0.917 |
| 0.4151 | 0.6170 | 3000 | 0.6996 | 0.84 | 0.871 | 0.885 | 0.77 | 0.823 | 0.839 | 0.84 | 0.839 |
| 0.4027 | 0.7199 | 3500 | 0.2694 | 0.932 | 0.971 | 0.947 | 0.91 | 0.928 | 0.932 | 0.932 | 0.932 |
| 0.4148 | 0.8227 | 4000 | 0.2328 | 0.94 | 0.981 | 0.95 | 0.925 | 0.937 | 0.94 | 0.94 | 0.94 |
| 0.3999 | 0.9255 | 4500 | 0.1900 | 0.945 | 0.992 | 0.908 | 0.985 | 0.945 | 0.945 | 0.945 | 0.945 |
| 0.393 | 1.0284 | 5000 | 0.2577 | 0.93 | 0.972 | 0.935 | 0.92 | 0.927 | 0.93 | 0.93 | 0.93 |
| 0.3814 | 1.1312 | 5500 | 0.3859 | 0.89 | 0.962 | 0.875 | 0.9 | 0.887 | 0.89 | 0.89 | 0.89 |
| 0.3758 | 1.2341 | 6000 | 0.2130 | 0.93 | 0.979 | 0.913 | 0.944 | 0.928 | 0.93 | 0.93 | 0.93 |
| 0.3727 | 1.3369 | 6500 | 0.6558 | 0.857 | 0.93 | 0.921 | 0.77 | 0.839 | 0.855 | 0.857 | 0.856 |
| 0.374 | 1.4397 | 7000 | 0.1787 | 0.958 | 0.989 | 0.947 | 0.967 | 0.957 | 0.958 | 0.958 | 0.958 |
| 0.3716 | 1.5426 | 7500 | 0.7428 | 0.822 | 0.908 | 0.917 | 0.694 | 0.79 | 0.818 | 0.822 | 0.819 |
| 0.374 | 1.6454 | 8000 | 0.1577 | 0.94 | 0.987 | 0.923 | 0.955 | 0.939 | 0.94 | 0.94 | 0.94 |
| 0.3621 | 1.7483 | 8500 | 0.2681 | 0.931 | 0.981 | 0.95 | 0.904 | 0.926 | 0.93 | 0.931 | 0.93 |
| 0.3731 | 1.8511 | 9000 | 0.2976 | 0.917 | 0.967 | 0.923 | 0.904 | 0.913 | 0.917 | 0.917 | 0.917 |
| 0.3652 | 1.9539 | 9500 | 0.1687 | 0.952 | 0.989 | 0.934 | 0.97 | 0.951 | 0.952 | 0.952 | 0.952 |
| 0.3517 | 2.0568 | 10000 | 0.1933 | 0.943 | 0.993 | 0.9 | 0.992 | 0.944 | 0.943 | 0.943 | 0.943 |
| 0.33 | 2.1596 | 10500 | 0.3700 | 0.912 | 0.977 | 0.9 | 0.92 | 0.91 | 0.912 | 0.912 | 0.912 |
| 0.341 | 2.2624 | 11000 | 0.1980 | 0.942 | 0.992 | 0.902 | 0.987 | 0.943 | 0.942 | 0.942 | 0.942 |
| 0.3367 | 2.3653 | 11500 | 0.1828 | 0.961 | 0.991 | 0.946 | 0.976 | 0.961 | 0.961 | 0.961 | 0.961 |
| 0.3305 | 2.4681 | 12000 | 0.1673 | 0.957 | 0.989 | 0.948 | 0.963 | 0.955 | 0.957 | 0.957 | 0.957 |
| 0.3409 | 2.5710 | 12500 | 0.2620 | 0.92 | 0.974 | 0.935 | 0.897 | 0.916 | 0.92 | 0.92 | 0.92 |
| 0.341 | 2.6738 | 13000 | 0.2656 | 0.936 | 0.979 | 0.96 | 0.905 | 0.932 | 0.936 | 0.936 | 0.936 |
| 0.3422 | 2.7766 | 13500 | 0.1622 | 0.959 | 0.992 | 0.945 | 0.971 | 0.958 | 0.959 | 0.959 | 0.959 |
| 0.3403 | 2.8795 | 14000 | 0.2629 | 0.941 | 0.981 | 0.953 | 0.924 | 0.939 | 0.941 | 0.941 | 0.941 |
| 0.3309 | 2.9823 | 14500 | 0.1996 | 0.955 | 0.987 | 0.944 | 0.964 | 0.954 | 0.955 | 0.955 | 0.955 |
| 0.3034 | 3.0852 | 15000 | 0.2373 | 0.953 | 0.989 | 0.938 | 0.968 | 0.952 | 0.953 | 0.953 | 0.953 |
| 0.313 | 3.1880 | 15500 | 0.3238 | 0.931 | 0.976 | 0.94 | 0.916 | 0.928 | 0.931 | 0.931 | 0.931 |
| 0.3098 | 3.2908 | 16000 | 0.2555 | 0.947 | 0.985 | 0.943 | 0.947 | 0.945 | 0.946 | 0.947 | 0.947 |
| 0.3102 | 3.3937 | 16500 | 0.2903 | 0.935 | 0.977 | 0.952 | 0.912 | 0.931 | 0.935 | 0.935 | 0.935 |
| 0.3097 | 3.4965 | 17000 | 0.3690 | 0.92 | 0.975 | 0.925 | 0.908 | 0.916 | 0.919 | 0.92 | 0.92 |
| 0.3015 | 3.5993 | 17500 | 0.3015 | 0.939 | 0.982 | 0.946 | 0.927 | 0.936 | 0.939 | 0.939 | 0.939 |
| 0.3084 | 3.7022 | 18000 | 0.3830 | 0.92 | 0.98 | 0.913 | 0.923 | 0.918 | 0.92 | 0.92 | 0.92 |
| 0.3047 | 3.8050 | 18500 | 0.4213 | 0.92 | 0.969 | 0.945 | 0.887 | 0.915 | 0.92 | 0.92 | 0.92 |
| 0.3022 | 3.9079 | 19000 | 0.4026 | 0.915 | 0.97 | 0.934 | 0.886 | 0.909 | 0.914 | 0.915 | 0.914 |
| 0.3001 | 4.0107 | 19500 | 0.2305 | 0.951 | 0.988 | 0.947 | 0.952 | 0.95 | 0.951 | 0.951 | 0.951 |
| 0.2827 | 4.1135 | 20000 | 0.2156 | 0.955 | 0.992 | 0.939 | 0.97 | 0.955 | 0.955 | 0.955 | 0.955 |
| 0.2748 | 4.2164 | 20500 | 0.2214 | 0.957 | 0.99 | 0.949 | 0.963 | 0.956 | 0.957 | 0.957 | 0.957 |
| 0.2813 | 4.3192 | 21000 | 0.2373 | 0.952 | 0.988 | 0.949 | 0.952 | 0.95 | 0.952 | 0.952 | 0.952 |
| 0.2754 | 4.4220 | 21500 | 0.2680 | 0.946 | 0.987 | 0.94 | 0.949 | 0.944 | 0.946 | 0.946 | 0.946 |
| 0.2861 | 4.5249 | 22000 | 0.2426 | 0.947 | 0.987 | 0.943 | 0.947 | 0.945 | 0.946 | 0.947 | 0.947 |
| 0.2783 | 4.6277 | 22500 | 0.2463 | 0.948 | 0.986 | 0.948 | 0.944 | 0.946 | 0.948 | 0.948 | 0.948 |
| 0.2779 | 4.7306 | 23000 | 0.2583 | 0.947 | 0.987 | 0.945 | 0.945 | 0.945 | 0.947 | 0.947 | 0.947 |
| 0.2723 | 4.8334 | 23500 | 0.2651 | 0.946 | 0.986 | 0.945 | 0.943 | 0.944 | 0.946 | 0.946 | 0.946 |
| 0.2803 | 4.9362 | 24000 | 0.2583 | 0.946 | 0.987 | 0.942 | 0.947 | 0.944 | 0.946 | 0.946 | 0.946 |
Framework versions
- Transformers 4.53.1
- Pytorch 2.6.0+cu124
- Datasets 2.14.4
- Tokenizers 0.21.2
- Downloads last month
- 1
Model tree for adity12345/Roberta_combo_v1_med_test
Base model
FacebookAI/xlm-roberta-base