Progressive Cognitive Architecture β€” 1.5B Dream LoRA (English)

πŸ† Best overall model (composite 87.6/100) β€” Qwen2.5-1.5B fine-tuned with 4-phase progressive training + SVD Dream Pruning.

✨ Highlights

Metric Score
Composite Score 87.6
Exact Accuracy 69.4% Β± 6.4
Adversarial Robustness 84.0% Β± 8.0
Delegation Accuracy 100.0% Β± 0.0
Delegation Rate 100.0% Β± 0.0
Magnitude Sense (OoMΒ±1) 100.0% Β± 0.0
Catastrophic Errors 0.0% Β± 0.0

Results: mean Β± std over 3 seeds (42, 43, 44), 50 samples Γ— 5 dimensions per seed.

πŸ”‘ Key Findings

  • Outperforms all 3B variants despite having half the parameters
  • Zero catastrophic errors β€” never produces absurd results
  • 100% delegation β€” always routes complex operations to tools
  • Dream pruning acts as cognitive regularization for capacity-constrained models

🧠 Progressive Cognitive Architecture

A bio-inspired 4-phase training methodology:

Phase Name What happens
1 Foundation Learn exact arithmetic via LoRA fine-tuning
2 Consolidation SVD Dream Pruning (rank 16β†’8) compresses knowledge into intuition
3 Delegation Learn complexity-aware routing: compute internally vs. delegate to tool
4 Orchestration Full pipeline: intuit β†’ route β†’ tool β†’ validate

Guiding Principle: Knowledge doesn't disappear β€” it collapses into attractors. Intuition is the compressed residue of experience.

πŸŒ™ Dream Pruning (SVD Low-Rank Factorization)

Instead of zeroing out small weights (magnitude pruning), Dream Pruning uses SVD decomposition to reduce the effective rank of LoRA matrices from 16 to 8. This preserves the principal directions ("logical connections") while discarding noise β€” analogous to memory consolidation during sleep.

W = U·Σ·V^T  β†’  W' = U[:,:k]Β·Ξ£[:k,:k]Β·V^T[:k,:]   (k=8)

πŸ”§ Training Configuration

Parameter Value
Base Model Qwen/Qwen2.5-1.5B
LoRA Rank 16 (β†’ 8 after SVD)
LoRA Alpha 32
LoRA Targets q_proj, k_proj, v_proj, o_proj
Dropout 0.05
Training Data ~6,000 English arithmetic examples
Hardware NVIDIA T4 16GB

πŸš€ Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained(
    "Qwen/Qwen2.5-1.5B", device_map="auto", torch_dtype="auto"
)
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-1.5B")

# Note: adapters are in the lora_adapters/ subfolder
model = PeftModel.from_pretrained(
    base_model,
    "dexmac/progressive-cognitive-dream-lora-en",
    subfolder="lora_adapters"
)

messages = [{"role": "user", "content": "Solve: 342 * 67"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=200, temperature=0.1)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Expected output pattern:

Step 1 - Intuition: in the order of tens of thousands
Step 2 - Routing: DELEGATE (medium complexity)
Step 3 - Tool: 22914
Step 4 - Validation: result 22914 consistent with estimate β†’ VALID

πŸ“Š Full Comparison

Model Composite Exact Adversarial Delegation Magnitude Safety
1.5B Dream (this) 87.6 69% 84% 100% 100% 100%
1.5B Flat 79.2 57% 81% 79% 100% 100%
3B Flat 78.5 60% 85% 79% 84% 100%
3B Dream 66.0 56% 34% 93% 100% 59%

πŸ”— Related Models

πŸ“ Citation

@software{progressive_cognitive_2026,
  author = {Dex Mac},
  title = {Progressive Cognitive Architecture for LLMs},
  year = {2026},
  url = {https://github.com/dexmac221/progressive-cognitive},
  version = {1.0.0}
}

πŸ“„ License

Apache 2.0

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for dexmac/progressive-cognitive-dream-lora-en

Base model

Qwen/Qwen2.5-1.5B
Adapter
(472)
this model

Evaluation results