File size: 838 Bytes
7ea86e1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
{
  "stage": 3,
  "timestamp": "2026-02-04T14:24:05.704295",
  "config": {
    "data_dir": "data/train",
    "output_dir": "checkpoints/lfm2-physics",
    "epochs_per_stage": 1,
    "batch_size": 2,
    "grad_accum": 8,
    "lr": 0.0002,
    "max_seq_length": 4096,
    "curriculum_stages": 5,
    "max_examples_per_stage": 1000,
    "max_context_frames": 5,
    "physics_loss_weight": 0.01,
    "resume": null,
    "wandb_project": "physics-llm",
    "wandb_offline": false,
    "lora_r": 32,
    "lora_alpha": 64,
    "model": "LiquidAI/LFM2-350M",
    "timestamp": "2026-02-04T14:11:22.835795"
  },
  "metrics": {
    "train_runtime": 148.7076,
    "train_samples_per_second": 6.725,
    "train_steps_per_second": 0.424,
    "total_flos": 7086689746944000.0,
    "train_loss": 1.0771233959803506,
    "epoch": 1.0,
    "step": 63
  }
}