Model_name
stringclasses 9
values | Train_size
int64 17.5k
80k
| Test_size
int64 4.38k
19.9k
| arg
dict | lora
null | Parameters
int64 109M
576M
| Trainable_parameters
int64 109M
576M
| r
null | Memory Allocation
stringlengths 6
8
| Training Time
stringlengths 5
7
| Performance
dict |
|---|---|---|---|---|---|---|---|---|---|---|
FacebookAI/roberta-base
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 124,655,629
| 124,655,629
| null |
2819.64
|
1444.75
|
{
"accuracy": 0.8086620883252293,
"f1_macro": 0.8074971081620297,
"f1_weighted": 0.8089926084331718,
"precision": 0.8112376642019833,
"recall": 0.8048878175934993
}
|
google-bert/bert-base-uncased
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 109,492,237
| 109,492,237
| null |
2640.15
|
1433.95
|
{
"accuracy": 0.8115695022306882,
"f1_macro": 0.8101042959591243,
"f1_weighted": 0.8118172140602734,
"precision": 0.8133373143934648,
"recall": 0.8080391782949717
}
|
google/rembert
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 575,935,373
| 575,935,373
| null |
10776.53
|
8466.9
|
{
"accuracy": 0.10942904406235901,
"f1_macro": 0.015174686148841219,
"f1_weighted": 0.02158716818978219,
"precision": 0.008417618774027616,
"recall": 0.07692307692307693
}
|
FacebookAI/xlm-roberta-large
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 559,903,757
| 559,903,757
| null |
10131.01
|
4703.38
|
{
"accuracy": 0.8106170735375207,
"f1_macro": 0.809018812566914,
"f1_weighted": 0.8108204110948648,
"precision": 0.8119560567314967,
"recall": 0.8076222775774148
}
|
FacebookAI/roberta-large
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 355,373,069
| 355,373,069
| null |
7396.25
|
1528.33
|
{
"accuracy": 0.8172840743896937,
"f1_macro": 0.8163705094398566,
"f1_weighted": 0.8174146001493484,
"precision": 0.819546863393102,
"recall": 0.8143561508863832
}
|
google-bert/bert-large-uncased
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 335,155,213
| 335,155,213
| null |
4886.3
|
1500.66
|
{
"accuracy": 0.8118702691864254,
"f1_macro": 0.810497140289229,
"f1_weighted": 0.8119116411541117,
"precision": 0.8142730315701795,
"recall": 0.8082541366904293
}
|
answerdotai/ModernBERT-large
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 395,844,621
| 395,844,621
| null |
5846.52
|
1720.84
|
{
"accuracy": 0.8195899543836784,
"f1_macro": 0.8187302351427349,
"f1_weighted": 0.8198901011478449,
"precision": 0.8238169381133621,
"recall": 0.8152943589028868
}
|
andreasmadsen/efficient_mlm_m0.40
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 355,375,117
| 355,375,117
| null |
5240.13
|
1550.9
|
{
"accuracy": 0.8015439370394506,
"f1_macro": 0.8006785018268369,
"f1_weighted": 0.8019864355002386,
"precision": 0.805722054413156,
"recall": 0.7970090688999505
}
|
microsoft/deberta-large
| 79,963
| 19,949
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 406,225,933
| 406,225,933
| null |
6680.97
|
1982.33
|
{
"accuracy": 0.8142262770063662,
"f1_macro": 0.8132404294946559,
"f1_weighted": 0.8146145205476452,
"precision": 0.8174711295837235,
"recall": 0.8104503643591923
}
|
answerdotai/ModernBERT-large
| 17,524
| 4,381
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 395,844,621
| 395,844,621
| null |
6886.18
|
648.9
|
{
"accuracy": 0.8427299703264095,
"f1_macro": 0.717201496341446,
"f1_weighted": 0.8399309483967907,
"precision": 0.753381419061771,
"recall": 0.6984358761194106
}
|
answerdotai/ModernBERT-large
| 17,524
| 4,381
|
{
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
}
| null | 398,476,821
| 398,476,821
| null |
6901.78
|
660.37
|
{
"accuracy": 0.5119835653960283,
"f1_macro": 0.08231988245386028,
"f1_weighted": 0.4433997976948826,
"precision": 0.08321908207453645,
"recall": 0.09436029720418769
}
|
README.md exists but content is empty.
- Downloads last month
- 6