trAIn.me_v3 / src /config /execution_generator_model_report.json
AIppyDev's picture
Pareil
8cb0c1d
raw
history blame contribute delete
804 Bytes
{
"created_at": "2025-12-07T17:13:48.852853Z",
"task": "language_modeling",
"target": "next_token_prediction (execution_generation)",
"n_train_samples": 1347,
"n_val_samples": 150,
"sequence_length": 128,
"device": "cpu",
"model": {
"name": "transformer_execution_generator_v3",
"type": "CausalLM_FineTuned",
"class": "GPT2LMHeadModel",
"checkpoint_base": "gpt2",
"total_params": 124439808,
"trainable_params": 124439808,
"size_mb": 474.7
},
"training": {
"epochs": 8,
"learning_rate": 3e-05,
"batch_size": 32,
"seed_global": 42,
"optimizer": "AdamW"
},
"metrics": {
"final_train_loss": 1.2690867379654285,
"final_val_loss": 1.293086338043213,
"best_val_loss": 1.293086338043213,
"best_loss": 1.293086338043213
}
}