Dataset Viewer
Auto-converted to Parquet Duplicate
Model_name
stringclasses
7 values
Train_size
int64
50.8k
50.8k
Test_size
int64
12.7k
12.7k
arg
dict
lora
listlengths
3
7
Parameters
int64
1.55B
4.05B
Trainable_parameters
int64
8.43M
52.2M
r
int64
16
64
Memory Allocation
stringclasses
7 values
Training Time
stringclasses
7 values
accuracy
float64
0.9
0.91
f1_macro
float64
0.9
0.91
f1_weighted
float64
0.91
0.91
precision
float64
0.9
0.91
recall
float64
0.9
0.91
Qwen/Qwen3-Embedding-4B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
4,051,667,456
29,859,840
32
4228.8
7297.19
0.911872
0.908712
0.912085
0.909135
0.908541
Alibaba-NLP/E2Rank-4B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
4,041,964,032
20,146,176
64
5008.21
7464.35
0.913057
0.909117
0.913205
0.910155
0.908332
Qwen/Qwen2-1.5B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
1,551,718,400
8,429,568
32
1640.62
2595.67
0.910765
0.906168
0.910936
0.906943
0.905723
facebook/opt-2.7b
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "fc1", "fc2", "out_proj" ]
2,671,518,720
19,906,560
32
2389.51
4775.3
0.906339
0.902094
0.906544
0.902839
0.901625
Qwen/Qwen2.5-3B-Instruct
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
3,097,667,584
12,257,280
64
3325.28
5559.38
0.910449
0.906387
0.910674
0.907048
0.90595
tiiuae/Falcon3-3B-Base
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
2,877,216,768
52,174,848
16
1523.19
4882.42
0.904916
0.900831
0.905114
0.90198
0.899975
Qwen/Qwen3-1.7B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0.2, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "paged_adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
1,749,644,288
29,589,504
16
536.09
2744.01
0.907366
0.903095
0.907577
0.904518
0.902069
README.md exists but content is empty.
Downloads last month
3