Dataset Viewer
Auto-converted to Parquet Duplicate
Model_name
stringclasses
7 values
Train_size
int64
50.8k
50.8k
Test_size
int64
12.7k
12.7k
arg
dict
lora
listlengths
1
1
Parameters
int64
1.54B
4.02B
Trainable_parameters
int64
1.54B
4.02B
r
int64
-1
-1
Memory Allocation
stringclasses
7 values
Training Time
stringclasses
7 values
accuracy
float64
0.91
0.91
f1_macro
float64
0.9
0.91
f1_weighted
float64
0.91
0.91
precision
float64
0.9
0.91
recall
float64
0.9
0.91
tiiuae/Falcon3-3B-Base
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 8, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
2,825,041,920
2,825,041,920
-1
22029.56
2419.01
0.904995
0.901369
0.90521
0.902003
0.901026
facebook/opt-2.7b
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 8, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
2,651,612,160
2,651,612,160
-1
18906.37
2672.45
0.908078
0.903887
0.908256
0.904018
0.903968
Qwen/Qwen2-1.5B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 8, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
1,543,288,832
1,543,288,832
-1
13207.78
1464.34
0.910054
0.90647
0.910294
0.907674
0.905598
Qwen/Qwen2.5-3B-Instruct
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 8, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
3,085,410,304
3,085,410,304
-1
24334.69
2913.46
0.911872
0.907564
0.912045
0.908018
0.907344
Qwen/Qwen3-1.7B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 8, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 8, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
1,720,054,784
1,720,054,784
-1
14652.05
1661.1
0.908868
0.904959
0.909074
0.905479
0.904686
Alibaba-NLP/E2Rank-4B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 12, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 12, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
4,021,817,856
4,021,817,856
-1
33811.52
3500.31
0.912346
0.908564
0.912594
0.908498
0.908886
Qwen/Qwen3-Embedding-4B
50,775
12,652
{ "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-8, "bf16": false, "fp16": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 12, "half_precision_backend": "auto", "label_smoothing_factor": 0, "learning_rate": 0.00002, "lr_scheduler_type": "linear", "max_grad_norm": 1, "max_steps": -1, "n_gpu": 1, "num_train_epochs": 1, "optim": "adamw_8bit", "optim_args": "Not have", "per_device_eval_batch_size": 8, "per_device_train_batch_size": 12, "warmup_ratio": 0, "warmup_steps": 5, "weight_decay": 0.01 }
[ "Not used" ]
4,021,807,616
4,021,807,616
-1
33795.66
3500.42
0.913848
0.909955
0.914072
0.909872
0.910256
README.md exists but content is empty.
Downloads last month
3