Model_name
stringclasses 7
values | Train_size
int64 50.8k
50.8k
| Test_size
int64 12.7k
12.7k
| arg
dict | lora
listlengths 3
7
| Parameters
int64 1.55B
4.05B
| Trainable_parameters
int64 8.43M
52.2M
| r
int64 16
64
| Memory Allocation
stringclasses 7
values | Training Time
stringclasses 9
values | accuracy
float64 0.08
0.91
| f1_macro
float64 0.01
0.9
| f1_weighted
float64 0.01
0.91
| precision
float64 0.01
0.91
| recall
float64 0.08
0.9
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Qwen/Qwen2-1.5B
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "paged_adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"o_proj",
"up_proj"
] | 1,551,718,400
| 8,429,568
| 32
|
1653.67
|
3431.42
| 0.904205
| 0.899388
| 0.904314
| 0.900491
| 0.898541
|
Qwen/Qwen2-1.5B
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"o_proj",
"up_proj"
] | 1,551,718,400
| 8,429,568
| 32
|
1653.67
|
3397.44
| 0.905153
| 0.900793
| 0.905315
| 0.901665
| 0.900141
|
tiiuae/Falcon3-3B-Base
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"o_proj",
"up_proj"
] | 2,877,216,768
| 52,174,848
| 16
|
3062.82
|
5318.69
| 0.901122
| 0.896474
| 0.901203
| 0.897548
| 0.895642
|
Qwen/Qwen2.5-3B-Instruct
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"k_proj",
"o_proj",
"q_proj",
"up_proj",
"v_proj"
] | 3,097,667,584
| 12,257,280
| 64
|
3341.31
|
6899.2
| 0.084493
| 0.011986
| 0.013166
| 0.006499
| 0.076923
|
Qwen/Qwen3-Embedding-4B
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"o_proj",
"up_proj"
] | 4,051,667,456
| 29,859,840
| 32
|
4293.63
|
8511.52
| 0.909026
| 0.90497
| 0.909233
| 0.906789
| 0.903535
|
Alibaba-NLP/E2Rank-4B
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"k_proj",
"o_proj",
"q_proj",
"up_proj",
"v_proj"
] | 4,041,964,032
| 20,146,176
| 64
|
5040.2
|
8672.72
| 0.90705
| 0.903813
| 0.907251
| 0.904657
| 0.903274
|
Qwen/Qwen3-1.7B
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"o_proj",
"up_proj"
] | 1,749,644,288
| 29,589,504
| 16
|
2199.0
|
3365.99
| 0.902861
| 0.897583
| 0.903039
| 0.898323
| 0.897258
|
facebook/opt-2.7b
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 1,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"fc1",
"fc2",
"out_proj"
] | 2,671,518,720
| 19,906,560
| 32
|
2453.29
|
9782.22
| 0.905707
| 0.90157
| 0.905864
| 0.901536
| 0.901826
|
Qwen/Qwen2.5-3B-Instruct
| 50,775
| 12,652
|
{
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
}
|
[
"down_proj",
"gate_proj",
"k_proj",
"o_proj",
"q_proj",
"up_proj",
"v_proj"
] | 3,097,667,584
| 12,257,280
| 64
|
3341.31
|
6892.17
| 0.901676
| 0.896785
| 0.901868
| 0.897697
| 0.896232
|
README.md exists but content is empty.
- Downloads last month
- 8