ChemMiniQ3-SAbRLo / training_config.json
gbyuvd's picture
Update after 1st chunk of MTP+NTP
fefea19 verified
raw
history blame contribute delete
248 Bytes
{
"model_type": "ChemQ3MTPForCausalLM",
"num_future_tokens": 3,
"horizon_loss_enabled": true,
"mtp_head_enabled": true,
"training_phases": [
"causal_lm_warmup",
"mtp_horizon_training"
],
"total_parameters": 9857155
}