| { | |
| "checkpoint_path": "/QuasarV4/checkpoints/step_220000", | |
| "analysis_timestamp": "0", | |
| "model_architecture": "HFA (Hierarchical Flow Anchoring)", | |
| "parameter_mappings": { | |
| "checkpoint_params": [ | |
| "token_embedding.weight", | |
| "blocks.0.attention.hierarchical_flow.evolution_rate", | |
| "blocks.0.attention.hierarchical_flow.memory_decay", | |
| "blocks.0.attention.hierarchical_flow.attention_memory", | |
| "blocks.0.attention.hierarchical_flow.q_proj.weight", | |
| "blocks.0.attention.hierarchical_flow.k_proj.weight", | |
| "blocks.0.attention.hierarchical_flow.v_proj.weight", | |
| "blocks.0.attention.hierarchical_flow.out_proj.weight", | |
| "blocks.0.attention.hierarchical_flow.attention_evolution.weight", | |
| "blocks.0.attention.hierarchical_flow.attention_evolution.bias", | |
| "blocks.0.attention.hierarchical_flow.memory_gate.weight", | |
| "blocks.0.attention.hierarchical_flow.memory_gate.bias", | |
| "blocks.0.attention.hierarchical_flow.temporal_dynamics.weight", | |
| "blocks.0.attention.hierarchical_flow.temporal_dynamics.bias", | |
| "blocks.0.attention.hierarchical_flow.checkpoint_trigger.checkpoint_frequency", | |
| "blocks.0.attention.hierarchical_flow.checkpoint_trigger.entropy_analyzer.weight", | |
| "blocks.0.attention.hierarchical_flow.checkpoint_trigger.entropy_analyzer.bias", | |
| "blocks.0.attention.hierarchical_flow.checkpoint_trigger.semantic_detector.0.weight", | |
| "blocks.0.attention.hierarchical_flow.checkpoint_trigger.semantic_detector.0.bias", | |
| "blocks.0.attention.hierarchical_flow.checkpoint_trigger.semantic_detector.2.weight" | |
| ] | |
| }, | |
| "checkpoint_structure": { | |
| "type": "nested_model_state_dict", | |
| "num_parameters": 274 | |
| }, | |
| "loading_instructions": [ | |
| "1. Load checkpoint with torch.load()", | |
| "2. Extract model_state_dict from checkpoint dictionary", | |
| "3. Map parameter names:", | |
| " - hfa_layers.X -> blocks.X.attention.hierarchical_flow", | |
| " - token_embedding -> token_embedding (direct match)", | |
| " - lm_head -> lm_head (direct match)", | |
| " - layer_norm -> layer_norm (direct match)", | |
| "4. Use strict=False for loading to handle mismatches" | |
| ], | |
| "training_metadata": { | |
| "step": 220000, | |
| "epoch": 1, | |
| "train_loss": 4.591190338134766, | |
| "val_loss": 4.591190338134766, | |
| "timestamp": 1757907906.1673536, | |
| "save_duration": 2.1279516220092773, | |
| "checkpoint_type": "hybrid_fast", | |
| "file_size_mb": 881.7255353927612 | |
| } | |
| } |