Ex0bit commited on
Commit
4498f5c
·
verified ·
1 Parent(s): 6662674

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +142 -0
README.md ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: lfm-nanotron-prism-research
4
+ license_link: LICENSE.md
5
+ language:
6
+ - en
7
+ tags:
8
+ - lfm
9
+ - prism
10
+ - gspo
11
+ - hybrid-architecture
12
+ - tool-use
13
+ - Thinking
14
+ pipeline_tag: text-generation
15
+ library_name: transformers
16
+ gated: true
17
+ extra_gated_heading: >-
18
+ Request Access to lfm-Nanotron (A tiny 2.6B-PRISM-SFT-GSPO-W2A16 RL trained
19
+ model with enhanced capabilities for agentic multi-turn tool calling,
20
+ thinking, and instruction following)
21
+ extra_gated_description: >
22
+ After completing your support donation at https://ko-fi.com/s/b0511b3815,
23
+ submit the access request at https://huggingface.co/Ex0bit/lfm-Nanotron to
24
+ receive automatic access to this limited edition model. Please provide your
25
+ information below. Access will be granted automatically after submission.
26
+ extra_gated_prompt: |
27
+ By requesting access, you agree to:
28
+ - Use this model for research or educational purposes only
29
+ - Not redistribute the model weights without explicit permission
30
+ - Cite this work appropriately in any publications
31
+ - Report any issues or safety concerns to the author
32
+ extra_gated_fields:
33
+ Full Name: text
34
+ Organization/Affiliation: text
35
+ Country: country
36
+ Intended Use:
37
+ type: select
38
+ options:
39
+ - Research
40
+ - Education
41
+ - Personal
42
+ - label: Commercial (requires separate license)
43
+ value: commercial
44
+ - label: Other
45
+ value: other
46
+ Brief description of your intended use case: text
47
+ I agree to the terms of use: checkbox
48
+ extra_gated_button_content: Agree and Request Access
49
+ ---
50
+
51
+
52
+
53
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/63adf1fa42fd3b8dbaeb0c92/9RwVQ2zsEqvFDNaGkOBTO.png)
54
+
55
+ # lfm-Nanotron: 2.6B PRISM-SFT-GSPO-WA216
56
+
57
+ <div align="center">
58
+
59
+ ** LFM Architeture model SFT + GSPO RL + PRISM **
60
+
61
+ [![Model](https://img.shields.io/badge/Model-2.6B-blue)]()
62
+ [![Architecture](https://img.shields.io/badge/Architecture-LFM2%20Hybrid-green)]()
63
+ [![Context](https://img.shields.io/badge/Context-128K-orange)]()
64
+
65
+ </div>
66
+
67
+ ## Model Description
68
+
69
+
70
+ **lfm-Nanotron**: Limited Edition 2.6B PRISM Model Access
71
+ Unlock a cutting-edge Nano sized AI model.
72
+
73
+
74
+ This is **lfm-Nanotron** — A Nano Sized 2.6B parameter hybrid architecture language model fine-tuned with advanced techniques you won't find in mainstream releases:
75
+ - SFT (Test-Time Supervised-Fine-Tuning) — Adaptive optimization at inference
76
+ - GSPO (Group Sequence Policy Optimization) — RL Enhanced reasoning, Instruction following, thinking, tool calling & logic
77
+ - PRISM (Projected Refusal Isolation via Subspace Modification) — State-of-the-art over-refusal/propaganda removal from LLMs
78
+ - 128K Context Window — Handle massive prompts with ease
79
+ - Agentic Tool Calling — Built for multi-turn, thinking, and instruction-following tasks
80
+
81
+ ### Architecture Details
82
+
83
+ | Parameter | Value |
84
+ |-----------|-------|
85
+ | Parameters | ~2.6B |
86
+ | Hidden Size | 2048 |
87
+ | Layers | 30 (22 Conv + 8 Full Attention) |
88
+ | Attention Heads | 32 |
89
+ | KV Heads | 8 (GQA) |
90
+ | Vocabulary | 65,536 |
91
+ | Max Context | 128,000 tokens |
92
+ | Architecture | Hybrid Conv + Attention (LFM2) |
93
+
94
+ ### Available Quantizations
95
+
96
+ | File | Quantization | Size | Use Case |
97
+ |------|-------------|------|----------|
98
+ | `lfm2-nanotron-ttft-gspo-prism-bf16.gguf` | BF16 | ~4.8GB | Full precision, best quality |
99
+ | `lfm2-nanotron-ttft-gspo-prism-Q4_K_M.gguf` | Q4_K_M | ~1.5GB | Balanced quality/size |
100
+ | `lfm2-nanotron-ttft-gspo-prism-Q2_K.gguf` | Q2_K | ~0.9GB | Maximum compression |
101
+
102
+ ## Usage
103
+
104
+ ### With llama.cpp
105
+
106
+ ```bash
107
+ ./llama-cli -m lfm2-nanotron-ttft-gspo-prism-Q4_K_M.gguf -p "Your prompt here" --temp 0.3 --min-p 0.15 --repeat-penalty 1.05
108
+ ```
109
+
110
+ ### Recommended Generation Parameters
111
+
112
+ ```json
113
+ {
114
+ "temperature": 0.3,
115
+ "min_p": 0.15,
116
+ "repeat_penalty": 1.05
117
+ }
118
+ ```
119
+
120
+ ## Citation
121
+
122
+ If you use this model in your research, please cite:
123
+
124
+ ```bibtex
125
+ @misc{lfm2-nanotron-2026,
126
+ title={lfm2-Nanotron: Test-Time Fine-Tuned LFM2 with GSPO+PRISM},
127
+ author={Exobit (Eric Elbaz)},
128
+ year={2026},
129
+ publisher={Hugging Face},
130
+ url={https://huggingface.co/Ex0bit/lfm2-Nanotron}
131
+ }
132
+ ```
133
+
134
+ ## License
135
+
136
+ This model is released under a custom research license. See LICENSE.md for details.
137
+
138
+ ## Acknowledgments
139
+
140
+ - [@mlabonne](https://huggingface.co/mlabonne) & [@liquidai](https://huggingface.co/LiquidAI) for the LFM2 architecture
141
+ - [@anakin87](https://huggingface.co/anakin87) for inspiring the idea
142
+ - The open-source AI community