# K-FAC Memorization Suppression Reproduction of ["From Memorization to Reasoning in the Spectrum of Loss Curvature"](https://github.com/goodfire-ai/memorization_kfac) with extended experiments on modified importance formulas. ## Overview This project implements K-FAC (Kronecker-Factored Approximate Curvature) based weight editing to suppress verbatim memorization in language models while preserving general capabilities. **Key insight:** The Fisher Information Matrix, approximated by K-FAC, reveals directions in weight space associated with memorization (low curvature) vs. generalization (high curvature). By removing low-curvature components, we can suppress memorization. ## Project Goal 1. **Reproduce** the paper's K-FAC method on OLMo-2 1B 2. **Compare** the original importance formula with a modified version: - **Original:** $\Pi_{ij} = \lambda_i \cdot \mu_j$ - **Modified:** $\Pi_{ij} = \lambda_i \cdot \mu_j \cdot |C_{ij}|^2$ ## Installation ```bash pip install -r requirements.txt ``` ## Project Structure ``` ├── src/ │ ├── kfac_collector.py # Collect K-FAC statistics (A, G matrices) │ ├── kfac_editor.py # Weight editing via eigendecomposition │ ├── evaluate.py # Memorization and perplexity metrics │ └── mine_memorized.py # Mine memorized sequences from training data ├── notebooks/ │ ├── 01_collect_kfac.ipynb # Colab: K-FAC collection (~2h on A100) │ ├── 02_mine_memorized.ipynb # Colab: Find memorized sequences (~1h) │ └── 03_experiments.ipynb # Colab: Run experiments (~2h) ├── plans/ │ └── implementation_plan.md # Detailed implementation plan ├── context/ │ ├── original_paper/ # Paper sections in markdown │ └── REPRODUCTION_PLAN.md # Initial reproduction plan └── requirements.txt ``` ## Quick Start ### Local Development ```python from src.kfac_collector import KFACCollector, KFACConfig from src.kfac_editor import KFACEditor, EditConfig from src.evaluate import memorization_score, perplexity # Load model from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("allenai/OLMo-2-1124-7B") tokenizer = AutoTokenizer.from_pretrained("allenai/OLMo-2-1124-7B") # Load pre-collected K-FAC stats collector = KFACCollector.load("kfac_statistics.pt", model) kfac_stats = collector.get_statistics() # Apply K-FAC editing config = EditConfig(energy_threshold=0.6, formula="original") editor = KFACEditor(model, kfac_stats, config) editor.edit_model() # Evaluate result = memorization_score(model, tokenizer, prefixes, suffixes) print(f"Strict accuracy: {result.strict_accuracy*100:.1f}%") ``` ### Running on Colab 1. **01_collect_kfac.ipynb** - Collect K-FAC statistics (~20M tokens, ~2h on A100) 2. **02_mine_memorized.ipynb** - Find memorized sequences from training data 3. **03_experiments.ipynb** - Run experiments and compare formulas ## Method ### K-FAC Statistics Collection For each target MLP layer, we collect: - **A**: Activation covariance matrix (input side) - **G**: Gradient covariance matrix (output side) These approximate the Fisher Information Matrix: $F_W \approx G \otimes A$ ### Weight Editing 1. **Eigendecompose** A and G matrices 2. **Transform** weights to curvature basis: $C = U_G^T W U_A$ 3. **Compute importance** using either formula 4. **Select** top components by cumulative energy (e.g., 60%) 5. **Reconstruct** edited weights: $W_{edited} = U_G (C \odot M) U_A^T$ ### Importance Formulas | Formula | Definition | Intuition | |---------|------------|-----------| | Original | $\Pi_{ij} = \lambda_i \cdot \mu_j$ | Pure curvature | | Modified | $\Pi_{ij} = \lambda_i \cdot \mu_j \cdot |C_{ij}|^2$ | Curvature weighted by actual weight magnitude | ## Hyperparameters | Parameter | 7B Model | 1B Model (estimated) | |-----------|----------|---------------------| | Target layers | 23, 24, 25 | 11, 12, 13 | | Projections | gate_proj, up_proj | gate_proj, up_proj | | Energy threshold | 60% | 60% | | K-FAC tokens | ~20M | ~20M | ## Expected Results Based on the paper (OLMo-2 1B): | Metric | Baseline | After K-FAC | |--------|----------|-------------| | Dolma strict accuracy | ~98% | ~3% | | Perplexity (Pile-10k) | ~23 | ~27 | ## References - Paper: [From Memorization to Reasoning in the Spectrum of Loss Curvature](https://github.com/goodfire-ai/memorization_kfac) - Original code: https://github.com/goodfire-ai/memorization_kfac - Model: [OLMo-2](https://huggingface.co/allenai/OLMo-2-1124-7B) - K-FAC: [Martens & Grosse, 2015](https://arxiv.org/abs/1503.05671) ## License MIT