Model_name
stringlengths
12
30
Train_size
int64
50.8k
50.8k
Test_size
int64
12.7k
12.7k
lora
listlengths
2
8
r
int64
12
64
Trainable_parameters
int64
3.59M
52.2M
Memory Allocation
float64
442
4.99k
accuracy
float64
0.9
90.8
f1_macro
float64
0.9
90.3
f1_weighted
float64
0.9
90.8
precision
float64
0.9
90.3
recall
float64
0.9
90.2
Alibaba-NLP/E2Rank-4B
50,775
12,652
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
64
20,146,176
4,986.58
0.912504
0.909019
0.912669
0.910099
0.908225
Qwen/Qwen2-1.5B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
32
8,429,568
1,654.05
0.910765
0.906168
0.910936
0.906943
0.905723
Qwen/Qwen2.5-3B-Instruct
50,775
12,652
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
64
12,257,280
3,329.49
0.910449
0.906387
0.910674
0.907048
0.90595
Qwen/Qwen3-1.7B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
16
29,589,504
2,195.19
0.907366
0.903095
0.907577
0.904518
0.902069
Qwen/Qwen3-Embedding-4B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
32
29,859,840
4,272
0.913136
0.909579
0.913353
0.910686
0.908755
facebook/opt-2.7b
50,775
12,652
[ "fc1", "fc2", "out_proj" ]
32
19,906,560
2,453.29
0.906339
0.902094
0.906544
0.902839
0.901625
tiiuae/Falcon3-3B-Base
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
16
52,174,848
3,053.5
0.904205
0.89961
0.904386
0.900577
0.898953
FacebookAI/roberta-large
50,775
12,652
[ "out_proj", "value", "dense", "key", "query" ]
12
12,918,970
1,236.64
90.246601
89.802879
90.260977
89.93768
89.700889
FacebookAI/xlm-roberta-large
50,775
12,652
[ "dense", "out_proj" ]
12
6,504,058
844.92
89.977869
89.593698
89.995384
89.727738
89.497052
Qwen/Qwen3-Reranker-0.6B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "score", "up_proj" ]
12
9,978,802
1,744.48
90.0411
89.583316
90.054603
89.763457
89.43125
RUCAIBox/mvp
50,775
12,652
[ "fc2", "q_proj", "out_proj", "fc1", "dense", "v_proj", "k_proj" ]
12
17,195,578
1,812.99
90.460006
90.053034
90.474028
90.179157
89.952785
answerdotai/ModernBERT-large
50,775
12,652
[ "Wi", "Wo", "classifier", "dense" ]
12
7,573,210
1,216.91
90.460006
90.070314
90.474233
90.214053
89.94897
facebook/bart-large
50,775
12,652
[ "fc2", "q_proj", "out_proj", "fc1", "dense", "v_proj", "k_proj" ]
12
17,195,578
1,730.76
90.436295
90.019956
90.450537
90.173706
89.895379
facebook/opt-125m
50,775
12,652
[ "fc2", "q_proj", "out_proj", "fc1", "v_proj", "k_proj", "score" ]
12
3,594,290
441.58
89.029402
88.547062
89.039785
88.658871
88.453779
facebook/opt-350m
50,775
12,652
[ "fc1", "fc2", "out_proj", "project_in", "project_out", "score" ]
12
6,460,026
927.41
89.464116
89.023809
89.476387
89.192131
88.899953
google-bert/bert-large-uncased
50,775
12,652
[ "classifier", "dense" ]
12
6,504,058
857.3
89.61429
89.180331
89.623731
89.245118
89.128224
google-t5/t5-large
50,775
12,652
[ "wo", "out_proj", "q", "wi", "dense", "k", "v", "o" ]
12
34,302,010
3,631.35
90.768258
90.266154
90.787784
90.340215
90.213506