Dataset Viewer
Auto-converted to Parquet Duplicate
Model_name
stringclasses
16 values
Train_size
int64
50.8k
50.8k
Test_size
int64
12.7k
12.7k
arg
dict
lora
listlengths
1
9
Parameters
int64
110M
1.85B
Trainable_parameters
int64
9.27k
1.11B
r
int64
4
1.02k
Memory Allocation
stringlengths
5
7
Training Time
stringlengths
5
7
Performance
dict
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
125,698,586
1,042,957
4
1379.61
516.83
{ "accuracy": 0.8465855200758773, "f1_macro": 0.8340490920495849, "f1_weighted": 0.8465504941446558, "precision": 0.8402862114476798, "recall": 0.8301870667970347 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
126,140,954
1,485,325
8
1375.29
539.19
{ "accuracy": 0.8593898197913373, "f1_macro": 0.8494724159910976, "f1_weighted": 0.8595404905204881, "precision": 0.8537412420731172, "recall": 0.8464116130684702 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
127,025,690
2,370,061
16
1379.01
524.26
{ "accuracy": 0.8689535251343661, "f1_macro": 0.860226577406406, "f1_weighted": 0.8691191048708272, "precision": 0.8630581563058186, "recall": 0.8579840198270493 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
128,795,162
4,139,533
32
1405.63
522.2
{ "accuracy": 0.873300663926652, "f1_macro": 0.8652518906400283, "f1_weighted": 0.8734504853470856, "precision": 0.8671457037466545, "recall": 0.8637183716416306 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
132,334,106
7,678,477
64
1470.14
538.22
{ "accuracy": 0.879544736010117, "f1_macro": 0.8717421988175561, "f1_weighted": 0.8796959266260036, "precision": 0.8729376208017359, "recall": 0.8708070546286392 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
139,411,994
14,756,365
128
1474.88
564.83
{ "accuracy": 0.8837337970281378, "f1_macro": 0.8764199863238034, "f1_weighted": 0.8838460171049024, "precision": 0.8775010379877884, "recall": 0.8755335295237698 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
408,644,673
1,290,292
4
3792.74
1765.12
{ "accuracy": 0.8254030983243756, "f1_macro": 0.7942233737452272, "f1_weighted": 0.8210719850322569, "precision": 0.8155352712384848, "recall": 0.7892248972999769 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
357,615,642
2,242,573
4
3299.3
1456.24
{ "accuracy": 0.8762251027505533, "f1_macro": 0.8692030515942923, "f1_weighted": 0.8765679781573217, "precision": 0.8711681710386169, "recall": 0.8678851272646673 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
153,567,770
28,912,141
256
1649.39
622.11
{ "accuracy": 0.8876857413847613, "f1_macro": 0.8811070927831653, "f1_weighted": 0.8877647364230433, "precision": 0.8822983580334766, "recall": 0.8800895685744806 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
562,146,330
2,242,573
4
4935.54
2021.03
{ "accuracy": 0.8725893139424596, "f1_macro": 0.8648132854936866, "f1_weighted": 0.8727950240317498, "precision": 0.867549548960421, "recall": 0.8626542247550917 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
109,950,746
458,509
4
1248.45
503.85
{ "accuracy": 0.6804457793234271, "f1_macro": 0.5831976548071854, "f1_weighted": 0.6489669753037608, "precision": 0.6552885170242525, "recall": 0.6076809656621555 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
577,734,682
1,799,309
4
5007.36
2625.72
{ "accuracy": 0.8695858362314258, "f1_macro": 0.8589912882110775, "f1_weighted": 0.8695841347189455, "precision": 0.8637437350730561, "recall": 0.855559709905021 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
110,399,258
907,021
8
1244.82
501.33
{ "accuracy": 0.7468384445147013, "f1_macro": 0.6780446340268369, "f1_weighted": 0.7257716525873043, "precision": 0.698242661197289, "recall": 0.6866301155276309 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
181,879,322
57,223,693
512
1948.99
745.42
{ "accuracy": 0.89084729687006, "f1_macro": 0.884702784129353, "f1_weighted": 0.8909290679599178, "precision": 0.8857158320594325, "recall": 0.883844733737056 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
358,795,290
3,422,221
8
3308.5
1466.79
{ "accuracy": 0.8821530192854885, "f1_macro": 0.8763401875319281, "f1_weighted": 0.8825445854181906, "precision": 0.877769674405757, "recall": 0.8753882277725543 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
111,296,282
1,804,045
16
1248.79
479.4
{ "accuracy": 0.7977394878280114, "f1_macro": 0.7523181078518651, "f1_weighted": 0.7862311782799655, "precision": 0.7982886603247584, "recall": 0.7532144822533675 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
409,934,965
2,580,584
8
3804.75
1765.81
{ "accuracy": 0.8494309200126462, "f1_macro": 0.8324264851760286, "f1_weighted": 0.8484212760977864, "precision": 0.8410424107713527, "recall": 0.8278077509740616 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
113,090,330
3,598,093
32
1274.29
503.21
{ "accuracy": 0.8400252924438824, "f1_macro": 0.8241649565993185, "f1_weighted": 0.838801434313264, "precision": 0.837904148263822, "recall": 0.8168347742360932 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
563,325,978
3,422,221
8
4944.75
2016.25
{ "accuracy": 0.8814416693012962, "f1_macro": 0.8750724802105493, "f1_weighted": 0.8815682095432281, "precision": 0.8771349816062244, "recall": 0.8733434004724216 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
116,678,426
7,186,189
64
1341.59
523.13
{ "accuracy": 0.856307303193171, "f1_macro": 0.8445821685599258, "f1_weighted": 0.8558074261383034, "precision": 0.8532108533533793, "recall": 0.8394980941298701 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
361,154,586
5,781,517
16
3335.47
1467.59
{ "accuracy": 0.8860259247549794, "f1_macro": 0.8805908113119117, "f1_weighted": 0.8862769058797022, "precision": 0.8815056148370751, "recall": 0.8799666252638331 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
123,854,618
14,362,381
128
1348.89
550.71
{ "accuracy": 0.8682421751501739, "f1_macro": 0.8607836525623542, "f1_weighted": 0.8681785403418177, "precision": 0.8650231484647595, "recall": 0.8575399385906514 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
412,515,549
5,161,168
16
3836.41
1775.46
{ "accuracy": 0.8675308251659817, "f1_macro": 0.8569831684700908, "f1_weighted": 0.8675100017298956, "precision": 0.8611393290578584, "recall": 0.8542100346505586 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
579,519,002
3,583,629
8
5026.7
2631.96
{ "accuracy": 0.8827853303825483, "f1_macro": 0.8758300378165708, "f1_weighted": 0.8829383871624759, "precision": 0.87862067881592, "recall": 0.8735385090229185 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
138,207,002
28,714,765
256
1518.54
612.14
{ "accuracy": 0.8768574138476131, "f1_macro": 0.8711695243461549, "f1_weighted": 0.8768884317753117, "precision": 0.8737841851778931, "recall": 0.8690053543037654 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
365,873,178
10,500,109
32
3414.77
1477.69
{ "accuracy": 0.8906892190957951, "f1_macro": 0.8860021264947283, "f1_weighted": 0.8909334482337231, "precision": 0.8868374971079175, "recall": 0.8854509207788522 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
125,477,402
821,773
4
1156.22
550.26
{ "accuracy": 0.8340183370218147, "f1_macro": 0.8198675525987951, "f1_weighted": 0.8337764478041396, "precision": 0.8259594334694711, "recall": 0.8159187392243591 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
166,911,770
57,419,533
512
1841.63
733.87
{ "accuracy": 0.882232058172621, "f1_macro": 0.8773384152398953, "f1_weighted": 0.8823046399232333, "precision": 0.8795561471893543, "recall": 0.8755108966229024 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
336,356,378
1,201,165
4
729.05
555.38
{ "accuracy": 0.815523237432817, "f1_macro": 0.7794404234917954, "f1_weighted": 0.8089604529608558, "precision": 0.8165454155559532, "recall": 0.7735023676594606 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
125,698,586
1,042,957
8
1150.03
536.72
{ "accuracy": 0.8510907366424281, "f1_macro": 0.8400420874226526, "f1_weighted": 0.8511327945816213, "precision": 0.8457216333479073, "recall": 0.8361739028568304 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
417,676,717
10,322,336
32
3916.76
1790.55
{ "accuracy": 0.8758299083148909, "f1_macro": 0.8668049587329789, "f1_weighted": 0.8758603217556982, "precision": 0.8697313514600873, "recall": 0.8646167954835591 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
337,544,218
2,389,005
8
719.58
554.83
{ "accuracy": 0.8619190641795763, "f1_macro": 0.8523901695902638, "f1_weighted": 0.862036251000165, "precision": 0.8565319337706493, "recall": 0.8494140859579764 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
126,140,954
1,485,325
16
1149.65
536.95
{ "accuracy": 0.8595478975656022, "f1_macro": 0.8495386110208438, "f1_weighted": 0.8596940655110615, "precision": 0.8539706656018181, "recall": 0.8463705209445991 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
375,310,362
19,937,293
64
3575.49
1522.37
{ "accuracy": 0.8940088523553589, "f1_macro": 0.8896207971222672, "f1_weighted": 0.8941604025617136, "precision": 0.8906132689531988, "recall": 0.8888308789628273 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
339,919,898
4,764,685
16
734.77
557.47
{ "accuracy": 0.8700600695542207, "f1_macro": 0.8629267469057272, "f1_weighted": 0.8703546878001558, "precision": 0.8662414806607376, "recall": 0.8604568138886367 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
109,944,602
452,365
8
1010.89
495.67
{ "accuracy": 0.5833069870376225, "f1_macro": 0.47480777422523524, "f1_weighted": 0.5335707257009746, "precision": 0.503493398959869, "recall": 0.5156260517847151 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
127,025,690
2,370,061
32
1175.96
538.72
{ "accuracy": 0.8669775529560544, "f1_macro": 0.8582782106285267, "f1_weighted": 0.8671392295134541, "precision": 0.8613026321631628, "recall": 0.8559757886864787 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
583,087,642
7,152,269
16
5074.65
2648.51
{ "accuracy": 0.89084729687006, "f1_macro": 0.8852825636698043, "f1_weighted": 0.8910752417418889, "precision": 0.887225855993687, "recall": 0.8836901840064006 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
570,403,866
10,500,109
32
5042.86
2045.0
{ "accuracy": 0.8910844135314575, "f1_macro": 0.8854909910992065, "f1_weighted": 0.8912291024968251, "precision": 0.887047946004157, "recall": 0.8841939914729813 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
110,386,970
894,733
16
1018.89
496.81
{ "accuracy": 0.6585520075877331, "f1_macro": 0.5583557193327258, "f1_weighted": 0.618386260774831, "precision": 0.625158519422246, "recall": 0.5892689860021918 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
128,795,162
4,139,533
64
1175.42
544.46
{ "accuracy": 0.8714037306354726, "f1_macro": 0.8629328315448218, "f1_weighted": 0.8714950054507163, "precision": 0.8654034306031588, "recall": 0.860973347921311 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
344,671,258
9,516,045
32
765.14
563.24
{ "accuracy": 0.8773316471704078, "f1_macro": 0.8708971632647742, "f1_weighted": 0.8775584683512808, "precision": 0.8730965955606386, "recall": 0.8692390386775077 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
111,271,706
1,779,469
32
1042.97
498.7
{ "accuracy": 0.7338760670249763, "f1_macro": 0.6800897675258101, "f1_weighted": 0.7154343448169183, "precision": 0.7471161150451973, "recall": 0.6864654554372182 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
132,334,106
7,678,477
128
1212.37
555.17
{ "accuracy": 0.8763831805248182, "f1_macro": 0.8685086985495007, "f1_weighted": 0.8764768753528103, "precision": 0.8706212561474461, "recall": 0.86678235158233 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
427,999,053
20,644,672
64
4107.01
1845.35
{ "accuracy": 0.8829434081568132, "f1_macro": 0.8760990330726381, "f1_weighted": 0.8830823425417631, "precision": 0.878388529941251, "recall": 0.874255558547274 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
354,173,978
19,018,765
64
825.99
584.63
{ "accuracy": 0.880730319317104, "f1_macro": 0.8745132364205285, "f1_weighted": 0.8808686332353639, "precision": 0.8761351934239846, "recall": 0.8732603833811635 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
113,041,178
3,548,941
64
1039.2
506.87
{ "accuracy": 0.7899936768890294, "f1_macro": 0.76085900959247, "f1_weighted": 0.7837586524219583, "precision": 0.7904333025476757, "recall": 0.7563708284429409 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
394,184,730
38,811,661
128
3649.15
1618.89
{ "accuracy": 0.89756560227632, "f1_macro": 0.8931155672712718, "f1_weighted": 0.8977212994148122, "precision": 0.8940294746357065, "recall": 0.8924113005747072 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
139,411,994
14,756,365
256
1347.02
589.18
{ "accuracy": 0.8800980082200442, "f1_macro": 0.8728568319907553, "f1_weighted": 0.8802014973485958, "precision": 0.8744470212702262, "recall": 0.8716389976774788 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
373,179,418
38,024,205
128
950.71
618.54
{ "accuracy": 0.8842080303509326, "f1_macro": 0.8784351126961792, "f1_weighted": 0.884335007759456, "precision": 0.8796394785271359, "recall": 0.8775049888168863 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
116,580,122
7,087,885
128
1079.39
499.2
{ "accuracy": 0.8262725260828327, "f1_macro": 0.8044735145506146, "f1_weighted": 0.8236107619440158, "precision": 0.8181056541984145, "recall": 0.7992784227671461 }
FacebookAI/roberta-base
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
153,567,770
28,912,141
512
1549.04
642.82
{ "accuracy": 0.8823110970597534, "f1_macro": 0.8752709699074926, "f1_weighted": 0.8824236522894506, "precision": 0.8759907367639677, "recall": 0.8747700003957696 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
579,841,050
19,937,293
64
5211.23
2092.91
{ "accuracy": 0.8938507745810939, "f1_macro": 0.8889839784997207, "f1_weighted": 0.8939976610165367, "precision": 0.8907862703375223, "recall": 0.8874784323164315 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
123,658,010
14,165,773
256
1214.8
539.47
{ "accuracy": 0.8452418589946253, "f1_macro": 0.8309024960657764, "f1_weighted": 0.8445499077974197, "precision": 0.8380230325734844, "recall": 0.8264247931667552 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
411,190,298
76,035,085
256
1209.98
696.93
{ "accuracy": 0.8861840025292443, "f1_macro": 0.8805852219073909, "f1_weighted": 0.8862916923417463, "precision": 0.8815998520714247, "recall": 0.8798811297517187 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
590,224,922
14,289,549
32
5184.43
2675.98
{ "accuracy": 0.8965380967435979, "f1_macro": 0.8916888027728069, "f1_weighted": 0.8967334029305609, "precision": 0.8939975604190971, "recall": 0.8897654555741151 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
448,643,725
41,289,344
128
4188.89
1954.45
{ "accuracy": 0.8895036357888081, "f1_macro": 0.8833073806422669, "f1_weighted": 0.8896636096198612, "precision": 0.884459853381671, "recall": 0.8823866083639367 }
google-bert/bert-base-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
137,813,786
28,321,549
512
1426.05
605.12
{ "accuracy": 0.85883654758141, "f1_macro": 0.848513365806343, "f1_weighted": 0.8586877233616093, "precision": 0.8537204028707317, "recall": 0.8448018864192982 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
431,933,466
76,560,397
256
4148.11
1794.14
{ "accuracy": 0.8996206133417641, "f1_macro": 0.8952887187406523, "f1_weighted": 0.8997428572588272, "precision": 0.8962920572695677, "recall": 0.894464186628896 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense" ]
487,212,058
152,056,845
512
1703.09
864.14
{ "accuracy": 0.8863420803035094, "f1_macro": 0.8809090607240366, "f1_weighted": 0.8864747176495101, "precision": 0.8811480559502635, "recall": 0.8809128972730303 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
335,758,362
603,149
4
917.44
546.92
{ "accuracy": 0.887132469174834, "f1_macro": 0.8817187487987473, "f1_weighted": 0.8872365394026598, "precision": 0.8824020330476059, "recall": 0.8812280581703731 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
396,316,698
472,077
4
1834.11
786.35
{ "accuracy": 0.7903098324375593, "f1_macro": 0.7672491813607067, "f1_weighted": 0.7886414085185387, "precision": 0.7749200809729936, "recall": 0.7629213053908442 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
598,715,418
38,811,661
128
5282.39
2191.98
{ "accuracy": 0.896380018969333, "f1_macro": 0.8918873284178506, "f1_weighted": 0.8965179629653568, "precision": 0.8931920068877847, "recall": 0.8908658698341267 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
336,348,186
1,192,973
8
573.23
547.32
{ "accuracy": 0.8872905469490989, "f1_macro": 0.8820272426764626, "f1_weighted": 0.8873896526508074, "precision": 0.8828409744012747, "recall": 0.8813973789041893 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
396,775,450
930,829
8
1840.15
789.62
{ "accuracy": 0.8280904204868795, "f1_macro": 0.8127439252127046, "f1_weighted": 0.8277441204563218, "precision": 0.8170094648812616, "recall": 0.8095579199545097 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
337,527,834
2,372,621
16
581.2
548.24
{ "accuracy": 0.8875276636104964, "f1_macro": 0.8822620778352918, "f1_weighted": 0.8876157933246032, "precision": 0.8830131811926824, "recall": 0.881689667346855 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
489,933,069
82,578,688
256
4775.35
2149.02
{ "accuracy": 0.8957477078722732, "f1_macro": 0.8909291187592895, "f1_weighted": 0.8958900440066224, "precision": 0.8917778185527342, "recall": 0.8902739783190025 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
507,430,938
152,057,869
512
5208.02
2154.21
{ "accuracy": 0.9023869743914006, "f1_macro": 0.8975241027323866, "f1_weighted": 0.902511886689775, "precision": 0.8983383090095226, "recall": 0.8968883767529492 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
604,499,482
28,564,109
64
5410.55
2756.43
{ "accuracy": 0.8986721466961746, "f1_macro": 0.8942493548634695, "f1_weighted": 0.8988471414290032, "precision": 0.8961804074671618, "recall": 0.8926048577899264 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
339,887,130
4,731,917
32
597.12
551.81
{ "accuracy": 0.8875276636104964, "f1_macro": 0.8822690294260441, "f1_weighted": 0.8876324867403612, "precision": 0.8830941233036421, "recall": 0.8816375818813812 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
397,692,954
1,848,333
16
1854.29
814.45
{ "accuracy": 0.8442933923490358, "f1_macro": 0.8317764391177607, "f1_weighted": 0.8440530174431455, "precision": 0.835826797307839, "recall": 0.8286484142313467 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
344,605,722
9,450,509
64
629.12
559.8
{ "accuracy": 0.887132469174834, "f1_macro": 0.8818528902639416, "f1_weighted": 0.8872188116550366, "precision": 0.8824483754035293, "recall": 0.8814262519162981 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
399,527,962
3,683,341
32
1890.68
792.05
{ "accuracy": 0.8579671198229529, "f1_macro": 0.8475435628380684, "f1_weighted": 0.8578404191364576, "precision": 0.8507979439495266, "recall": 0.8449504002910871 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
636,464,154
76,560,397
256
5783.89
2368.27
{ "accuracy": 0.8991463800189693, "f1_macro": 0.8943440596520253, "f1_weighted": 0.8992907008437808, "precision": 0.8956325153176388, "recall": 0.8932812078131077 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
357,025,818
1,652,749
4
3283.36
1447.14
{ "accuracy": 0.866740436294657, "f1_macro": 0.8575637104818205, "f1_weighted": 0.8669708348834821, "precision": 0.8608586794600368, "recall": 0.8551682938094264 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
354,042,906
18,887,693
128
693.11
572.79
{ "accuracy": 0.8870534302877016, "f1_macro": 0.8818615284745714, "f1_weighted": 0.8871316748564163, "precision": 0.8823367797181898, "recall": 0.8815199668222549 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
372,917,274
37,762,061
256
821.11
609.05
{ "accuracy": 0.885946885867847, "f1_macro": 0.8812526997751445, "f1_weighted": 0.8859959148203225, "precision": 0.8816106503696065, "recall": 0.881007863260953 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
403,197,978
7,353,357
64
1944.59
803.9
{ "accuracy": 0.8696648751185583, "f1_macro": 0.8615520466335258, "f1_weighted": 0.8696118331637468, "precision": 0.8646678666301584, "recall": 0.8590107351236551 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "fc1", "fc2", "out_proj" ]
572,511,757
165,157,376
512
5814.55
2568.12
{ "accuracy": 0.8993044577932343, "f1_macro": 0.8948132965953287, "f1_weighted": 0.8994168863709725, "precision": 0.8956847616023478, "recall": 0.8941679009907664 }
microsoft/deberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "in_proj", "pos_proj" ]
408,016,922
1,790,989
4
3613.07
1898.71
{ "accuracy": 0.8578090420486879, "f1_macro": 0.8446092300628985, "f1_weighted": 0.8578612785245953, "precision": 0.8492567246489446, "recall": 0.8418384523871402 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
357,615,642
2,242,573
8
3292.14
1435.24
{ "accuracy": 0.8760670249762883, "f1_macro": 0.8692705789045871, "f1_weighted": 0.8764015848224459, "precision": 0.8713756336119993, "recall": 0.867732993727411 }
google-bert/bert-large-uncased
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
410,666,010
75,510,797
512
1114.18
683.0
{ "accuracy": 0.884603224786595, "f1_macro": 0.878605207347677, "f1_weighted": 0.8847180290634366, "precision": 0.8782405610261594, "recall": 0.8791753431973165 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
633,048,602
57,113,229
128
5810.97
2901.63
{ "accuracy": 0.9019917799557382, "f1_macro": 0.8982219370986022, "f1_weighted": 0.9021507602038787, "precision": 0.900625623362189, "recall": 0.8961772457902 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
410,538,010
14,693,389
128
2042.92
831.97
{ "accuracy": 0.874328169459374, "f1_macro": 0.8665932773699195, "f1_weighted": 0.8742417767025412, "precision": 0.8692475565339852, "recall": 0.8643949760993761 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "dense", "out_proj" ]
711,961,626
152,057,869
512
6843.98
2713.87
{ "accuracy": 0.9010433133101486, "f1_macro": 0.8969479131518704, "f1_weighted": 0.9012152952855088, "precision": 0.8976570979548679, "recall": 0.8964894216574089 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
425,218,074
29,373,453
256
2248.69
874.61
{ "accuracy": 0.8794656971229845, "f1_macro": 0.8726974236966247, "f1_weighted": 0.8794719106552626, "precision": 0.8747160009190501, "recall": 0.8710901196893861 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
358,795,290
3,422,221
16
3308.33
1440.69
{ "accuracy": 0.8818368637369586, "f1_macro": 0.875790887031754, "f1_weighted": 0.8821591776492361, "precision": 0.8772440147930095, "recall": 0.8747788935196866 }
microsoft/deberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "in_proj", "pos_proj" ]
409,794,586
3,568,653
8
3629.29
1895.48
{ "accuracy": 0.874328169459374, "f1_macro": 0.8641103102381869, "f1_weighted": 0.8745478117117244, "precision": 0.866397138011316, "recall": 0.8629314782882648 }
answerdotai/ModernBERT-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Wqkv" ]
454,578,202
58,733,581
512
2562.46
944.06
{ "accuracy": 0.8840499525766677, "f1_macro": 0.8772364746801417, "f1_weighted": 0.8840641342001619, "precision": 0.8791632119549534, "recall": 0.8756725309481499 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "k_proj", "q_proj", "v_proj" ]
408,239,117
884,736
4
3882.55
1795.46
{ "accuracy": 0.25418906101802086, "f1_macro": 0.16746654763773686, "f1_weighted": 0.19839063988848946, "precision": 0.22534217234371567, "recall": 0.2100578649092377 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
361,154,586
5,781,517
32
3344.93
1466.0
{ "accuracy": 0.8857888080935821, "f1_macro": 0.8799275312566462, "f1_weighted": 0.886121637308196, "precision": 0.8808806803033844, "recall": 0.8793853471902271 }
google/rembert
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in" ]
690,146,842
114,211,469
256
6479.82
3174.09
{ "accuracy": 0.90396775213405, "f1_macro": 0.9000784127377291, "f1_weighted": 0.9041240754395038, "precision": 0.901687898049922, "recall": 0.8987227054048579 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
561,556,506
1,652,749
4
4920.26
1993.61
{ "accuracy": 0.861681947518179, "f1_macro": 0.8518543085684569, "f1_weighted": 0.8616834766914389, "precision": 0.8557438409368476, "recall": 0.8491147729646762 }
microsoft/deberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "in_proj", "pos_proj" ]
413,349,914
7,123,981
16
3671.11
1911.46
{ "accuracy": 0.8819949415112235, "f1_macro": 0.8740560385409323, "f1_weighted": 0.8822275343532041, "precision": 0.874914517996927, "recall": 0.8736064228611521 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
365,873,178
10,500,109
64
3416.66
1495.35
{ "accuracy": 0.8890294024660133, "f1_macro": 0.8832128667872327, "f1_weighted": 0.889326249515431, "precision": 0.8837432896318163, "recall": 0.8829964929542189 }
facebook/bart-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "k_proj", "q_proj", "v_proj" ]
409,123,853
1,769,472
8
3896.36
1772.8
{ "accuracy": 0.6342870692380651, "f1_macro": 0.5243534507803612, "f1_weighted": 0.5969217399151321, "precision": 0.5656195963958232, "recall": 0.5562342366705734 }
albert/albert-xxlarge-v2
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "embedding_hidden_mapping_in", "ffn", "ffn_output", "pooler" ]
222,948,378
299,533
4
2293.62
3286.26
{ "accuracy": 0.7384603224786594, "f1_macro": 0.6632644496543156, "f1_weighted": 0.7188668278957628, "precision": 0.7557279016742403, "recall": 0.6736103242013692 }
FacebookAI/xlm-roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
562,146,330
2,242,573
8
4929.63
1982.89
{ "accuracy": 0.8754347138792286, "f1_macro": 0.8684851872815932, "f1_weighted": 0.8756161000577907, "precision": 0.8711312369445623, "recall": 0.8663700730948646 }
microsoft/deberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "classifier", "dense", "in_proj", "pos_proj" ]
420,460,570
14,234,637
32
3773.31
1937.34
{ "accuracy": 0.8883970913689535, "f1_macro": 0.881447477588251, "f1_weighted": 0.8885959982355389, "precision": 0.8823729100014337, "recall": 0.8808481640861859 }
FacebookAI/roberta-large
50,775
12,652
{ "auto_find_batch_size": true, "gradient_accumulation_steps": 4, "learning_rate": 0.00005, "logging_steps": 1, "lr_scheduler_type": "linear", "num_train_epochs": 1, "optim": "adamw_8bit", "output_dir": "outputs", "report_to": "none", "save_strategy": "no", "save_total_limit": 0, "seed": 3407, "warmup_steps": 5, "weight_decay": 0.01 }
[ "key", "query", "value" ]
375,310,362
19,937,293
128
3589.87
1523.41
{ "accuracy": 0.8900569079987354, "f1_macro": 0.8840841051769437, "f1_weighted": 0.8903668026826507, "precision": 0.8849892661664793, "recall": 0.8835230328349574 }
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
3