Model_name stringclasses 1 value | Train_size int64 50 50 | Test_size int64 50 50 | arg dict | lora listlengths 3 3 | Parameters int64 5.6B 5.66B | Trainable_parameters int64 6.9M 61.7M | r int64 8 8 | Memory Allocation stringclasses 2 values | Training Time stringclasses 2 values | accuracy float64 0 0.04 | f1_macro float64 0 0.02 | f1_weighted float64 0 0.03 | precision float64 0 0.02 | recall float64 0 0.04 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
google/t5gemma-2b-2b-prefixlm-it | 50 | 50 | {
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
} | [
"k_proj",
"q_proj",
"v_proj"
] | 5,603,789,596 | 6,901,262 | 8 | 5998.11 | 6.69 | 0 | 0 | 0 | 0 | 0 |
google/t5gemma-2b-2b-prefixlm-it | 50 | 50 | {
"adafactor": false,
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_epsilon": 1e-8,
"bf16": false,
"fp16": false,
"fp16_opt_level": "O1",
"gradient_accumulation_steps": 4,
"half_precision_backend": "auto",
"label_smoothing_factor": 0,
"learning_rate": 0.00005,
"lr_scheduler_type": "linear",
"max_grad_norm": 1,
"max_steps": -1,
"n_gpu": 2,
"num_train_epochs": 1,
"optim": "adamw_8bit",
"optim_args": "Not have",
"per_device_eval_batch_size": 8,
"per_device_train_batch_size": 8,
"warmup_ratio": 0,
"warmup_steps": 5,
"weight_decay": 0.01
} | [
"k_proj",
"q_proj",
"v_proj"
] | 5,658,549,518 | 61,661,184 | 8 | 5892.86 | 7.03 | 0.04 | 0.020308 | 0.02549 | 0.015079 | 0.044643 |
README.md exists but content is empty.
- Downloads last month
- 12