Finetuned Models
Collection
6 items
•
Updated
The dataset used to finetune this model tries to imrpove gemma 3 1b's performance in things like mmlu and humaneval.
Model training loss:
Run history:
train/epoch ▁▁▁▁▁▂▂▂▂▂▃▃▃▃▃▄▄▄▄▄▅▅▅▅▅▆▆▆▆▆▇▇▇▇▇▇▇▇██ train/global_step ▁▁▁▁▂▂▂▂▂▂▃▃▃▃▃▄▄▄▄▄▄▅▅▅▅▅▆▆▆▆▆▇▇▇▇▇▇███ train/grad_norm ▁▃█▂▄▃▂▄▃▂▃▄▅▅▃▄▄▅▃▄▄▃▄▄▄▄▄▅▄▄▅▅▄▄▄▄▄▅▄▄ train/learning_rate ▄████▇▇▇▇▇▆▆▆▆▆▆▆▅▅▅▅▅▅▄▄▄▄▄▃▃▃▂▂▂▂▂▂▁▁▁ train/loss █▄▃▃▃▂▂▂▂▂▂▂▂▂▂▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁
Run summary:
total_flos 1.092006641664e+18 train/epoch 1 train/global_step 3907 train/grad_norm 0.21461 train/learning_rate 0.0 train/loss 0.7482 train_loss 0.81085 train_runtime 15509.2958 train_samples_per_second 16.119 train_steps_per_second 0.252
Base model
google/gemma-3-1b-pt