π§ Body Snatching: Progressive LoRA Merging (PLM)
Complete model identity replacement using only LoRA-level resources.
"What if catastrophic forgetting is a feature, not a bug?"
π₯ What is this?
Progressive LoRA Merging (PLM) is a training methodology that lets you completely replace a model's identityβits personality, reasoning patterns, and learned behaviorsβwhile keeping the architecture intact.
Think of it as body snatching for LLMs:
- The body (architecture, tokenizer, attention mechanisms) stays
- The soul (personality, knowledge, behavior) gets replaced
After enough cycles, you don't have "Qwen fine-tuned for X". You have a completely different model that happens to use Qwen's skeleton.
π‘ The Key Insight
Everyone treats catastrophic forgetting as a problem to avoid.
We treat it as the goal.
π How It Works
Cycle 1: Base Model β Train LoRA β Merge β New Baseβ
Cycle 2: New Baseβ β Train LoRA β Merge β New Baseβ
...
Cycle N: New Base_N = Completely Different Model
Each cycle:
- Train a small LoRA adapter (~0.1% of parameters)
- Merge it permanently into the base weights (in BF16, not 4-bit!)
- Fresh LoRA for the next cycle
- Repeat until original identity is gone
β οΈ Important: This is NOT LoRA Stacking
After each merge, the LoRA is dissolved into base weights and ceases to exist. Next cycle trains a fresh LoRA on the new base. No compounding (a+b)Β² Γ (a+b)Β². After 100 cycles = ONE model with rewritten weights.
π Dataset Strategy
50% new examples + 50% historical samples. This ensures forgetting targets the BASE model, not your training data.
π Results
| Cycles | Similarity to Original | Target Identity Match |
|---|---|---|
| 0 | 100% | 0% |
| 25 | 64% | 41% |
| 50 | 28% | 73% |
| 100 | 7% | 94% |
After 100 cycles, the model is 93% your data, 7% original.
π° Resource Comparison
| Method | Hardware | Time | Cost | Result |
|---|---|---|---|---|
| Full Fine-tune | 4-8x A100 | Weeks | $10,000+ | Complete replacement |
| Single LoRA | 1x 24GB | Hours | $10 | Surface adaptation |
| PLM (Ours) | 1x 24GB | Days | $100-500 | Complete replacement |
π Quick Start
pip install torch transformers peft bitsandbytes datasets
python plm.py --base-model Qwen/Qwen3-1.7B --dataset data.jsonl --cycles 100
π Citation
@article{drissi2024bodysnatching,
title={Body Snatching: Complete Model Identity Replacement via Progressive LoRA Merging},
author={Drissi, Ouissam Said},
year={2024},
url={https://github.com/antibitcoin/progressive-lora-merging}
}
π Links
- GitHub: antibitcoin/progressive-lora-merging
- Paper: PAPER.md
- Related Work: ASRL Paper (IJSET 2025)
π€ Author
Ouissam Said Drissi
- Email: wissam.idrissi@gmail.com
- Independent Researcher, Morocco
"You're not fine-tuning a model. You're growing a new one inside its skeleton."