File size: 704 Bytes
9fd0c72 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 | {
"step": 50000,
"loss": 5.999711477773722,
"config": {
"base_model_name": "meta-llama/Llama-3.1-8B-Instruct",
"masking_type": "bidirectional",
"batch_size": 16,
"learning_rate": 0.0003,
"weight_decay": 0.01,
"grad_clip": 1.0,
"max_steps": 50000,
"max_length": 128,
"target_tokens": 100000000,
"save_interval": 5000,
"log_interval": 50
},
"model_type": "CustomTransformerModel",
"attention_type": "bidirectional",
"training_tokens": 102400000,
"compatible_with_inference": true,
"accelerator": "unsloth_kernel_optimized",
"training_mode": "pure_diffusion",
"save_method": "unsloth_optimized",
"timestamp": "2025-06-11T06:22:08.448859"
} |