| | --- |
| | language: en |
| | license: apache-2.0 |
| | library_name: peft |
| | base_model: Qwen/Qwen2.5-1.5B |
| | tags: |
| | - lora |
| | - peft |
| | - baseline |
| | - flat-training |
| | - math |
| | - arithmetic |
| | - control-group |
| | datasets: |
| | - custom |
| | pipeline_tag: text-generation |
| | model-index: |
| | - name: progressive-cognitive-baseline-lora-en |
| | results: |
| | - task: |
| | type: text-generation |
| | name: Cognitive Arithmetic |
| | metrics: |
| | - type: exact_accuracy |
| | value: 56.9 |
| | name: Exact Accuracy (%) |
| | - type: composite_score |
| | value: 79.2 |
| | name: Composite Cognitive Score |
| | --- |
| | |
| | # Progressive Cognitive Architecture β 1.5B Flat LoRA (English, Control) |
| |
|
| | **Control model** β Qwen2.5-1.5B fine-tuned with all training data in a single pass (no phases, no pruning). Serves as the baseline for evaluating progressive training. |
| |
|
| | ## π Results |
| |
|
| | | Metric | Score | |
| | |--------|-------| |
| | | **Composite Score** | **79.2** | |
| | | Exact Accuracy | 56.9% Β± 6.4 | |
| | | Adversarial Robustness | 81.3% Β± 2.3 | |
| | | Delegation Accuracy | 100.0% Β± 0.0 | |
| | | Delegation Rate | 58.7% Β± 4.6 | |
| | | Magnitude Sense (OoMΒ±1) | 100.0% Β± 0.0 | |
| | | Catastrophic Errors | **0.0% Β± 0.0** | |
| |
|
| | > Results: mean Β± std over 3 seeds (42, 43, 44), 50 samples Γ 5 dimensions per seed. |
| |
|
| | ## βοΈ Comparison with Dream LoRA |
| |
|
| | | Metric | Flat (this) | Dream | Delta | |
| | |--------|-------------|-------|-------| |
| | | Composite | 79.2 | **87.6** | +8.4 | |
| | | Exact Accuracy | 56.9% | **69.4%** | +12.5pp | |
| | | Delegation Rate | 58.7% | **100.0%** | +41.3pp | |
| | | Number Sense | 6.7% | **60.7%** | +54.0pp | |
| |
|
| | The Dream model shows significantly stronger delegation and number sense, demonstrating that progressive training + SVD pruning adds cognitive capabilities beyond what flat training provides. |
| |
|
| | ## π§ Training Configuration |
| |
|
| | | Parameter | Value | |
| | |-----------|-------| |
| | | Base Model | Qwen/Qwen2.5-1.5B | |
| | | LoRA Rank | 16 | |
| | | LoRA Alpha | 32 | |
| | | LoRA Targets | q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj | |
| | | Dropout | 0.05 | |
| | | Training Data | ~6,000 English arithmetic examples (all mixed in one pass) | |
| | | Hardware | NVIDIA T4 16GB | |
| | |
| | ## π Quick Start |
| | |
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | from peft import PeftModel |
| | |
| | base_model = AutoModelForCausalLM.from_pretrained( |
| | "Qwen/Qwen2.5-1.5B", device_map="auto", torch_dtype="auto" |
| | ) |
| | tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-1.5B") |
| |
|
| | model = PeftModel.from_pretrained( |
| | base_model, |
| | "dexmac/progressive-cognitive-baseline-lora-en" |
| | ) |
| | |
| | messages = [{"role": "user", "content": "Calculate: 342 * 67"}] |
| | text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) |
| | inputs = tokenizer(text, return_tensors="pt").to(model.device) |
| | outputs = model.generate(**inputs, max_new_tokens=200, temperature=0.1) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| | |
| | ## π Related Models |
| | |
| | - [**1.5B Dream LoRA**](https://huggingface.co/dexmac/progressive-cognitive-dream-lora-en) β Progressive training + Dream Pruning (best model) |
| | - [3B Flat](https://huggingface.co/dexmac/progressive-cognitive-qwen3b-baseline-lora) β Same approach on Qwen2.5-3B |
| | - [Results Dataset](https://huggingface.co/datasets/dexmac/progressive-cognitive-results) β Raw evaluation data |
| | - [GitHub](https://github.com/dexmac221/progressive-cognitive) β Full source code |
| | |
| | ## π Citation |
| | |
| | ```bibtex |
| | @software{progressive_cognitive_2026, |
| | author = {Dex Mac}, |
| | title = {Progressive Cognitive Architecture for LLMs}, |
| | year = {2026}, |
| | url = {https://github.com/dexmac221/progressive-cognitive}, |
| | version = {1.0.0} |
| | } |
| | ``` |
| | |
| | |
| | ## π License |
| | |
| | Apache 2.0 |
| | |
| | |