Progressive Cognitive Architecture β€” 1.5B Flat LoRA (English, Control)

Control model β€” Qwen2.5-1.5B fine-tuned with all training data in a single pass (no phases, no pruning). Serves as the baseline for evaluating progressive training.

πŸ“Š Results

Metric Score
Composite Score 79.2
Exact Accuracy 56.9% Β± 6.4
Adversarial Robustness 81.3% Β± 2.3
Delegation Accuracy 100.0% Β± 0.0
Delegation Rate 58.7% Β± 4.6
Magnitude Sense (OoMΒ±1) 100.0% Β± 0.0
Catastrophic Errors 0.0% Β± 0.0

Results: mean Β± std over 3 seeds (42, 43, 44), 50 samples Γ— 5 dimensions per seed.

βš–οΈ Comparison with Dream LoRA

Metric Flat (this) Dream Delta
Composite 79.2 87.6 +8.4
Exact Accuracy 56.9% 69.4% +12.5pp
Delegation Rate 58.7% 100.0% +41.3pp
Number Sense 6.7% 60.7% +54.0pp

The Dream model shows significantly stronger delegation and number sense, demonstrating that progressive training + SVD pruning adds cognitive capabilities beyond what flat training provides.

πŸ”§ Training Configuration

Parameter Value
Base Model Qwen/Qwen2.5-1.5B
LoRA Rank 16
LoRA Alpha 32
LoRA Targets q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
Dropout 0.05
Training Data ~6,000 English arithmetic examples (all mixed in one pass)
Hardware NVIDIA T4 16GB

πŸš€ Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained(
    "Qwen/Qwen2.5-1.5B", device_map="auto", torch_dtype="auto"
)
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-1.5B")

model = PeftModel.from_pretrained(
    base_model,
    "dexmac/progressive-cognitive-baseline-lora-en"
)

messages = [{"role": "user", "content": "Calculate: 342 * 67"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=200, temperature=0.1)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

πŸ”— Related Models

πŸ“ Citation

@software{progressive_cognitive_2026,
  author = {Dex Mac},
  title = {Progressive Cognitive Architecture for LLMs},
  year = {2026},
  url = {https://github.com/dexmac221/progressive-cognitive},
  version = {1.0.0}
}

πŸ“„ License

Apache 2.0

Downloads last month
35
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for dexmac/progressive-cognitive-baseline-lora-en

Base model

Qwen/Qwen2.5-1.5B
Adapter
(472)
this model

Evaluation results