--- title: Intention Collapse 🧠 emoji: 📉 colorFrom: blue colorTo: red sdk: static pinned: false --- # 🧠 Intention Collapse: Intention-Level Metrics for Reasoning **Paper:** [arXiv:2601.01011](https://arxiv.org/abs/2601.01011) **Code:** [GitHub Repository](https://github.com/patriciomvera/intention-collapse-experiments) ## ⚡️ TL;DR We introduce **Intention Collapse**, a framework to study how LLMs compress high-dimensional internal states into a single token sequence. We propose three model-agnostic metrics: **Intention Entropy**, **Effective Dimensionality**, and **Recoverability**. ## 🔥 Key Findings * [cite_start]**CoT is not magic:** It improves GSM8K but degrades performance on ARC-Challenge[cite: 30]. * [cite_start]**Entropy Shift:** CoT makes Mistral *more certain* (lower entropy) but makes LLaMA *less certain* (higher entropy)[cite: 33]. ## 🛠️ How to use the metrics To extract the "Intention State" $I$ (pre-collapse) from your own model: ```python # (Pon aquí un snippet simplificado de tu repo, por ejemplo:) from intention_metrics import extract_intention_state # Get the pre-collapse hidden state I = extract_intention_state(model, prompt, layer_idx=-1) print(f"Intention Entropy: {calculate_entropy(I)}") ## Citation ```bibtex @article{vera2026intention, title={Intention Collapse: Intention-Level Metrics for Reasoning in Language}, author={Vera, Patricio}, journal={arXiv preprint arXiv:2601.01011}, year={2026} }