VideoMAE_Base_wlasl_100_longtail_200
This model is a fine-tuned version of MCG-NJU/videomae-base on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 2.9736
- Accuracy: 0.4822
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 2
- eval_batch_size: 2
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 8
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 36000
- mixed_precision_training: Native AMP
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|---|---|---|---|---|
| 18.7021 | 0.005 | 180 | 4.6442 | 0.0118 |
| 18.6026 | 1.0050 | 360 | 4.6319 | 0.0118 |
| 18.5393 | 2.0050 | 540 | 4.6238 | 0.0178 |
| 18.3789 | 3.0050 | 721 | 4.6170 | 0.0178 |
| 18.3763 | 4.005 | 901 | 4.6237 | 0.0178 |
| 18.334 | 5.0050 | 1081 | 4.6342 | 0.0296 |
| 18.1779 | 6.0050 | 1261 | 4.6036 | 0.0296 |
| 17.9948 | 7.0050 | 1442 | 4.5903 | 0.0266 |
| 17.9333 | 8.005 | 1622 | 4.6144 | 0.0237 |
| 17.7505 | 9.0050 | 1802 | 4.5865 | 0.0118 |
| 17.4917 | 10.0050 | 1982 | 4.5626 | 0.0207 |
| 16.9821 | 11.0050 | 2163 | 4.3615 | 0.0444 |
| 16.2362 | 12.005 | 2343 | 4.1515 | 0.0533 |
| 15.2255 | 13.0050 | 2523 | 3.9603 | 0.0740 |
| 14.0646 | 14.0050 | 2703 | 3.7594 | 0.0828 |
| 12.8642 | 15.0050 | 2884 | 3.4154 | 0.1420 |
| 11.6502 | 16.005 | 3064 | 3.3917 | 0.1627 |
| 10.332 | 17.0050 | 3244 | 3.0359 | 0.2249 |
| 8.9465 | 18.0050 | 3424 | 2.8625 | 0.2840 |
| 7.6629 | 19.0050 | 3605 | 2.8202 | 0.3107 |
| 6.2517 | 20.005 | 3785 | 2.6478 | 0.3343 |
| 5.1876 | 21.0050 | 3965 | 2.4982 | 0.3728 |
| 4.0929 | 22.0050 | 4145 | 2.3891 | 0.3876 |
| 3.0425 | 23.0050 | 4326 | 2.2212 | 0.4083 |
| 2.3667 | 24.005 | 4506 | 2.1609 | 0.4586 |
| 1.7821 | 25.0050 | 4686 | 2.2471 | 0.4260 |
| 1.4215 | 26.0050 | 4866 | 2.2263 | 0.4675 |
| 1.1324 | 27.0050 | 5047 | 2.2360 | 0.4556 |
| 0.9114 | 28.005 | 5227 | 2.2021 | 0.4852 |
| 0.6966 | 29.0050 | 5407 | 2.3123 | 0.4408 |
| 0.5676 | 30.0050 | 5587 | 2.1198 | 0.5355 |
| 0.4494 | 31.0050 | 5768 | 2.2495 | 0.4911 |
| 0.3321 | 32.005 | 5948 | 2.2658 | 0.5089 |
| 0.227 | 33.0050 | 6128 | 2.4423 | 0.4882 |
| 0.2203 | 34.0050 | 6308 | 2.4358 | 0.4763 |
| 0.2643 | 35.0050 | 6489 | 2.5521 | 0.4675 |
| 0.111 | 36.005 | 6669 | 2.5787 | 0.4882 |
| 0.1009 | 37.0050 | 6849 | 2.4022 | 0.5059 |
| 0.1275 | 38.0050 | 7029 | 2.5451 | 0.5 |
| 0.1874 | 39.0050 | 7210 | 2.8339 | 0.4586 |
| 0.1695 | 40.005 | 7390 | 3.0320 | 0.4320 |
| 0.1735 | 41.0050 | 7570 | 2.6961 | 0.4941 |
| 0.1299 | 42.0050 | 7750 | 2.9589 | 0.4675 |
| 0.1399 | 43.0050 | 7931 | 2.6799 | 0.5325 |
| 0.118 | 44.005 | 8111 | 2.8731 | 0.5 |
| 0.1583 | 45.0050 | 8291 | 2.8757 | 0.4970 |
| 0.1667 | 46.0050 | 8471 | 2.9294 | 0.4941 |
| 0.202 | 47.0050 | 8652 | 3.1443 | 0.4615 |
| 0.1301 | 48.005 | 8832 | 2.8941 | 0.5207 |
| 0.2298 | 49.0050 | 9012 | 3.1270 | 0.4704 |
| 0.1858 | 50.0050 | 9192 | 2.9736 | 0.4822 |
Framework versions
- Transformers 4.46.1
- Pytorch 2.5.1+cu124
- Datasets 3.1.0
- Tokenizers 0.20.1
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for Shawon16/VideoMAE_Base_wlasl_100_longtail_200
Base model
MCG-NJU/videomae-base