🧟 Body Snatching: Progressive LoRA Merging (PLM)

Complete model identity replacement using only LoRA-level resources.

"What if catastrophic forgetting is a feature, not a bug?"

πŸ”₯ What is this?

Progressive LoRA Merging (PLM) is a training methodology that lets you completely replace a model's identityβ€”its personality, reasoning patterns, and learned behaviorsβ€”while keeping the architecture intact.

Think of it as body snatching for LLMs:

  • The body (architecture, tokenizer, attention mechanisms) stays
  • The soul (personality, knowledge, behavior) gets replaced

After enough cycles, you don't have "Qwen fine-tuned for X". You have a completely different model that happens to use Qwen's skeleton.

πŸ’‘ The Key Insight

Everyone treats catastrophic forgetting as a problem to avoid.

We treat it as the goal.

πŸ”„ How It Works

Cycle 1:  Base Model β†’ Train LoRA β†’ Merge β†’ New Base₁
Cycle 2:  New Base₁  β†’ Train LoRA β†’ Merge β†’ New Baseβ‚‚
...
Cycle N:  New Base_N = Completely Different Model

Each cycle:

  1. Train a small LoRA adapter (~0.1% of parameters)
  2. Merge it permanently into the base weights (in BF16, not 4-bit!)
  3. Fresh LoRA for the next cycle
  4. Repeat until original identity is gone

⚠️ Important: This is NOT LoRA Stacking

After each merge, the LoRA is dissolved into base weights and ceases to exist. Next cycle trains a fresh LoRA on the new base. No compounding (a+b)Β² Γ— (a+b)Β². After 100 cycles = ONE model with rewritten weights.

πŸ”€ Dataset Strategy

50% new examples + 50% historical samples. This ensures forgetting targets the BASE model, not your training data.

πŸ“Š Results

Cycles Similarity to Original Target Identity Match
0 100% 0%
25 64% 41%
50 28% 73%
100 7% 94%

After 100 cycles, the model is 93% your data, 7% original.

πŸ’° Resource Comparison

Method Hardware Time Cost Result
Full Fine-tune 4-8x A100 Weeks $10,000+ Complete replacement
Single LoRA 1x 24GB Hours $10 Surface adaptation
PLM (Ours) 1x 24GB Days $100-500 Complete replacement

πŸš€ Quick Start

pip install torch transformers peft bitsandbytes datasets

python plm.py --base-model Qwen/Qwen3-1.7B --dataset data.jsonl --cycles 100

πŸ“– Citation

@article{drissi2024bodysnatching,
  title={Body Snatching: Complete Model Identity Replacement via Progressive LoRA Merging},
  author={Drissi, Ouissam Said},
  year={2024},
  url={https://github.com/antibitcoin/progressive-lora-merging}
}

πŸ”— Links

πŸ‘€ Author

Ouissam Said Drissi


"You're not fine-tuning a model. You're growing a new one inside its skeleton."

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support