tinyllama-mentalchat16k / training_metrics.json
advy's picture
Finetune on MentalChat16K - eval_loss: 0.7816
81ca6a4 verified
{
"model": "tinyllama-mental-health",
"base_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"dataset": "ShenLab/MentalChat16K",
"lora_config": {
"rank": 16,
"alpha": 32,
"target_modules": [
"q_proj",
"k_proj",
"v_proj",
"o_proj",
"gate_proj",
"up_proj",
"down_proj"
],
"dropout": 0.1
},
"training": {
"final_train_loss": 0.7040511888504029,
"total_steps": 2500,
"epochs": 4,
"learning_rate": 0.0002,
"per_device_batch_size": 4,
"gradient_accumulation": 2
},
"evaluation": {
"eval_loss": 0.7816067934036255,
"eval_runtime": 18.522,
"eval_samples_per_second": 25.483,
"eval_steps_per_second": 6.371,
"epoch": 3.7397157816005984
},
"test_eval": {
"eval_loss": 0.7700048089027405,
"eval_runtime": 18.7703,
"eval_samples_per_second": 25.146,
"eval_steps_per_second": 6.287,
"epoch": 3.7397157816005984
},
"dataset_stats": {
"train_size": 5347,
"val_size": 472,
"test_size": 472
}
}