Divyanshu-Bhatt commited on
Commit
6a23c67
·
1 Parent(s): 60b7894

Initial Commit

Browse files
Files changed (3) hide show
  1. .gitignore +1 -0
  2. README.md +94 -3
  3. thumbnail.png +3 -0
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ **/.DS_Store
README.md CHANGED
@@ -1,3 +1,94 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <div align="center">
2
+
3
+ # Geometry of Decision Making in Language Models
4
+
5
+ **Abhinav Joshi · Divyanshu Bhatt · Ashutosh Modi**
6
+ *NeurIPS 2025*
7
+ </div>
8
+
9
+
10
+ <div align="center">
11
+ <br/>
12
+ <p align="center">
13
+ <a href="https://arxiv.org/pdf/2511.20315">
14
+ <img src="https://img.shields.io/badge/arXiv-2511.20315-6A1B9A?style=for-the-badge&logo=arxiv&logoColor=white"/>
15
+ </a>
16
+ <a href="https://huggingface.co/datasets/Exploration-Lab/dim-discovery-archive">
17
+ <img src="https://img.shields.io/badge/Experimental Findings-Hugging%20Face-FFB000?style=for-the-badge&logo=huggingface&logoColor=black"/>
18
+ </p>
19
+ <a href="https://openreview.net/forum?id=Jj4NdJtXwp">
20
+ <img src="https://img.shields.io/badge/OpenReview-NeurIPS%202025-512DA8?style=for-the-badge"/>
21
+ </a>
22
+ <a href="https://github.com/Exploration-Lab/dim-discovery-archive">
23
+ <img src="https://img.shields.io/badge/Code-GitHub-311B92?style=for-the-badge&logo=github&logoColor=white"/>
24
+ </a>
25
+ <a href="https://aboxspaces.github.io">
26
+ <img src="https://img.shields.io/badge/AboxSpaces-Blog-3949AB?style=for-the-badge&logo=github&logoColor=white"/>
27
+ </a>
28
+ <a href="LICENSE">
29
+ <img src="https://img.shields.io/badge/License-MIT-1E88E5?style=for-the-badge&logo=opensourceinitiative&logoColor=white"/>
30
+ </a>
31
+ <br/><br/>
32
+ </div>
33
+
34
+ <div align="justify">
35
+
36
+ > This repository contains the official implementation/release for the NeurIPS 2025 paper Geometry of Decision Making in Language Models.
37
+ > We study the internal decision-making processes of large language models through the lens of intrinsic dimension (ID), analyzing how hidden representations evolve across layers in a multiple-choice question answering (MCQA) setting.
38
+ > This repository provides additional experimental results, intrinsic-dimension estimations, and layer-wise performance analysis, that we believe will be a valuable resource for researchers for further exploration in this area.
39
+ </div>
40
+
41
+ <div align="justify">
42
+
43
+ ![Teaser image](./thumbnail.png)
44
+ **Picture:** *In the transformer-based architectures, a vector (latent features) of the same hidden dimensions $`d`$, is transformed by transformer blocks $`f_l`$. Though the extrinsic dimension remains the same, we find that the feature space lies on low-dimensional manifolds of different intrinsic dimensions. Intrinsically, there exists a mapping $`\phi_l`$ corresponding to each $`f_l`$. We study how these compressed manifolds align with the decision-making process in middle layers. We project the internal representations back to the vocabulary space to inspect the decisiveness. There is a sudden shift in performance that is aligned with the follow-up of a sharp peak observed in the residual-post ID estimates.*
45
+
46
+ </div>
47
+
48
+ ## Abstract
49
+ > Large Language Models (LLMs) show strong generalization across diverse tasks,
50
+ yet the internal decision-making processes behind their predictions remain opaque.
51
+ In this work, we study the geometry of hidden representations in LLMs through
52
+ the lens of intrinsic dimension (ID), focusing specifically on decision-making
53
+ dynamics in a multiple-choice question answering (MCQA) setting. We perform a
54
+ large-scale study, with 28 open-weight transformer models and estimate ID across
55
+ layers using multiple estimators, while also quantifying per-layer performance on
56
+ MCQA tasks. Our findings reveal a consistent ID pattern across models: early
57
+ layers operate on low-dimensional manifolds, middle layers expand this space,
58
+ and later layers compress it again, converging to decision-relevant representations.
59
+ Together, these results suggest LLMs implicitly learn to project linguistic inputs
60
+ onto structured, low-dimensional manifolds aligned with task-specific decisions,
61
+ providing new geometric insights into how generalization and reasoning emerge in
62
+ language models.
63
+
64
+
65
+ ## Structure
66
+ ```
67
+ results/
68
+ ├── dataset_name/
69
+ │ ├── fewshot_xx/
70
+ │ │ ├── model_id/
71
+ │ │ ├── last_token_relative_depth_xx/
72
+ │ │ ├── intrinsic/
73
+ │ │ ├── results.csv (global intrinsic estimates)
74
+ │ │ ├── mle_local_dims
75
+ │ │ ├── layer_name.csv (local intrinsic estimates)
76
+ │ │ ├── llm_hook_args.json (arguments used for activation collection)
77
+ │ │ ├── llm_intrinsic_args.json (arguments used for ID estimation)
78
+ │ │ ├── accuracy_layer_name.json (accuracy per layer using logit lens)
79
+ │ │ ├── accuracy_output.json (overall accuracy on the dataset)
80
+ │ │ ├── predictions.csv (predictions and corresponding ground truths)
81
+ ```
82
+
83
+ ## Citation
84
+
85
+ ```
86
+ @inproceedings{
87
+ joshi2025geometry,
88
+ title={Geometry of Decision Making in Language Models},
89
+ author={Abhinav Joshi and Divyanshu Bhatt and Ashutosh Modi},
90
+ booktitle={The Thirty-ninth Annual Conference on Neural Information Processing Systems},
91
+ year={2025},
92
+ url={https://openreview.net/forum?id=Jj4NdJtXwp}
93
+ }
94
+ ```
thumbnail.png ADDED

Git LFS Details

  • SHA256: 80fa44631212696ab3baf9aa4f5f17b640cee6728db0e113d83740ddb406f9ca
  • Pointer size: 131 Bytes
  • Size of remote file: 181 kB