Dataset Viewer
Auto-converted to Parquet Duplicate
Model_name
stringclasses
7 values
Train_size
int64
50.8k
50.8k
Test_size
int64
12.7k
12.7k
arg
dict
lora
listlengths
4
7
Parameters
int64
1.69B
4.29B
Trainable_parameters
int64
94.4M
264M
r
int64
64
128
Memory Allocation
stringclasses
7 values
Training Time
stringclasses
7 values
accuracy
float64
0.9
0.91
f1_macro
float64
0.89
0.9
f1_weighted
float64
0.9
0.91
precision
float64
0.89
0.9
recall
float64
0.89
0.9
Alibaba-NLP/E2Rank-4B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 2, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
4,286,092,288
264,274,432
128
6805.09
7941.72
0.906339
0.902009
0.906516
0.903233
0.901092
Qwen/Qwen2-1.5B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 2, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
1,691,026,944
147,738,112
128
3519.77
3364.84
0.902387
0.897623
0.90257
0.899046
0.896562
Qwen/Qwen2.5-3B-Instruct
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 2, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
3,284,797,440
199,387,136
128
4867.71
6554.01
0.906339
0.901593
0.906533
0.902153
0.901292
Qwen/Qwen3-1.7B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 2, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
1,859,542,016
139,487,232
128
3853.88
3757.9
0.899779
0.894332
0.900017
0.894068
0.894869
Qwen/Qwen3-Embedding-4B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 2, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
4,222,381,056
200,573,440
128
5817.08
8910.5
0.905944
0.901419
0.906181
0.902171
0.900992
facebook/opt-2.7b
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "fc1", "fc2", "k_proj", "out_proj", "q_proj", "v_proj" ]
2,746,017,280
94,405,120
64
2994.39
4631.2
0.903731
0.90034
0.903957
0.900977
0.899962
tiiuae/Falcon3-3B-Base
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 4, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00005, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 2, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
2,946,192,384
121,150,464
128
2602.32
4930.32
0.895669
0.891411
0.895778
0.892314
0.890729
README.md exists but content is empty.
Downloads last month
1