Fix: use merged_model_state.pt (LoRA baked into base weights, fixes missing FFN weights) 13790f5 verified LisaMegaWatts commited on 1 day ago
Fix config: n_layers=8, n_monarch_heads=1, use progressive model c0f06ec verified LisaMegaWatts commited on 1 day ago