anujjamwal commited on
Commit
d8ac8d0
·
verified ·
1 Parent(s): 956fa6c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -6
README.md CHANGED
@@ -6,11 +6,20 @@ tags:
6
  - generated_from_trainer
7
  - sft
8
  - trl
 
9
  licence: license
 
 
10
  ---
11
 
12
  # Model Card for OpenMath-Nemotron-1.5B-PruneAware
13
 
 
 
 
 
 
 
14
  This model is a fine-tuned version of [anujjamwal/OpenMath-Nemotron-1.5B-PruneAware](https://huggingface.co/anujjamwal/OpenMath-Nemotron-1.5B-PruneAware).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
@@ -48,11 +57,11 @@ This model was trained with SFT.
48
  Cite TRL as:
49
 
50
  ```bibtex
51
- @software{vonwerra2020trl,
52
- title = {{TRL: Transformers Reinforcement Learning}},
53
- author = {von Werra, Leandro and Belkada, Younes and Tunstall, Lewis and Beeching, Edward and Thrush, Tristan and Lambert, Nathan and Huang, Shengyi and Rasul, Kashif and Gallouédec, Quentin},
54
- license = {Apache-2.0},
55
- url = {https://github.com/huggingface/trl},
56
- year = {2020}
57
  }
58
  ```
 
6
  - generated_from_trainer
7
  - sft
8
  - trl
9
+ - custom_generate
10
  licence: license
11
+ datasets:
12
+ - anujjamwal/OpenMathReasoning-Sampled-Hierarchical-Cot
13
  ---
14
 
15
  # Model Card for OpenMath-Nemotron-1.5B-PruneAware
16
 
17
+ This model implements [Cognitive Compression](https://github.com/anujjamwal/cognitive-compression) an approach to produce hierarchical
18
+ structured chain of thought that can be actively pruned at inference time while maintaining the solution quality.
19
+ Tradition Chain-of-Thought is append-onl; a token once generated remains in context for ever. Context compression introduces hierarchical
20
+ reasoning where reasoning is broken into subproblems. Once the subproblem is solved, its full chain of thought can be discarded and
21
+ replaced with the **summary and solution** dramatically reducing the context window pressure.
22
+
23
  This model is a fine-tuned version of [anujjamwal/OpenMath-Nemotron-1.5B-PruneAware](https://huggingface.co/anujjamwal/OpenMath-Nemotron-1.5B-PruneAware).
24
  It has been trained using [TRL](https://github.com/huggingface/trl).
25
 
 
57
  Cite TRL as:
58
 
59
  ```bibtex
60
+ @misc{jamwal2026cognitivecompression,
61
+ title = {{Cognitive Compression: Hierarchical Chain of Thought for Efficient LLM Reasoning}},
62
+ author = {Jamwal, Anuj},
63
+ url = {huggingface.co/anujjamwal/OpenMath-Nemotron-1.5B-PruneAware},
64
+ year = {2026},
65
+ note = {CS224N Winter '26 Final Project: Stanford University}
66
  }
67
  ```