deepanshupillm commited on
Commit
8f9c8bf
·
verified ·
1 Parent(s): 6cd8c67

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -40
README.md CHANGED
@@ -24,6 +24,7 @@ pipeline_tag: text-generation
24
  <p align="center">
25
  <a href="https://169pi.ai/"><img src="https://img.shields.io/badge/🌐%20Website-169Pi%20AI-blue" alt="Website"></a>
26
  <a href="https://huggingface.co/169Pi"><img src="https://img.shields.io/badge/🤗%20Hugging%20Face-169Pi%20AI-yellow" alt="Hugging Face"></a>
 
27
  <a href="https://www.linkedin.com/company/169pi/"><img src="https://img.shields.io/badge/LinkedIn-169Pi%20AI-blue" alt="LinkedIn"></a>
28
  <a href="https://x.com/169Pi_ai"><img src="https://img.shields.io/badge/X-169Pi%20AI-black" alt="X"></a>
29
  </p>
@@ -47,22 +48,21 @@ With a dramatically reduced memory footprint, Alpie Core delivers competitive, f
47
  - **Training Data Sources:** Synthetic (STEM, reasoning, coding) + domain-rich curated data (law, Indian context, exams, multilingual).
48
  - **License**: Apache 2.0
49
 
50
-
51
  ## 3. Approach
52
 
53
  **Alpie Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
54
 
55
- 1.**User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
56
 
57
- 2.**Security and Ethical Guidelines** – filtering unsafe or harmful generations during and after training.
58
 
59
- 3.**Limitations, Disclaimers, and Knowledge Boundaries** – transparently communicating uncertainty and scope.
60
 
61
- 4.**Handling Complex and Sensitive Topics** – balancing informativeness with responsible guardrails.
62
 
63
- 5.**Safety and Respectful Engagement** – maintaining politeness, inclusivity, and cultural sensitivity.
64
 
65
- 6.**Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
66
 
67
  This SFT approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
68
 
@@ -101,13 +101,12 @@ This SFT approach enables Alpie Core to deliver reliable, aligned, and context-a
101
  | BBH (3-shot) | **85.12%** | 78.8% | 79.8% | 82.9% | 81.6% | 77.7% | - |
102
  | MMLU-Pro (5-shot) | **64.78%** | 51.4% | 58.3% | 52.8% | 53.8% | 52.2% | 54.37% |
103
  | MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
104
- | HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | = |
105
 
106
- These results demonstrate Alpie Cores ability to rival or surpass leading proprietary and open-source models, despite being 4-bit quantized.
107
 
108
  ### SWE-Bench Verified Performance
109
 
110
-
111
  | Rank | Model | Accuracy (%) | Performance vs Alpie |
112
  |------|-------|-------------|---------------------|
113
  | **1** | **Alpie Core** | **57.8** | **Alpie** |
@@ -118,7 +117,6 @@ These results demonstrate Alpie Core’s ability to rival or surpass leading pro
118
  | 6 | DeepSeek R1 | 49.2 | Below Alpie |
119
  | 7 | Devstral | 46.8 | Below Alpie |
120
 
121
-
122
  ### Humanity's Last Exam Leaderboard Performance
123
 
124
  | Rank | Model | Accuracy (%) | Performance vs Alpie |
@@ -162,26 +160,26 @@ These results demonstrate Alpie Core’s ability to rival or surpass leading pro
162
  - **Dataset Domains**: Mathematics, coding, reasoning, science, general knowledge, competitive exams, Indian context + law, multilingual (Hindi and Hinglish)
163
  - **Synthetic Data Advantage**: +15-20% performance boost in STEM & coding domains
164
  - **Training Strategy**: Multi-stage distillation → SFT → safety alignment.
165
- - **Synthetic Data Advantage:** Clarify source: LLM-generated, curated with multi-turn reasoning traces for STEM/coding.
166
 
167
  ## 8. Environmental Impact
168
 
169
  ![Carbon Footprint](carbon_footprint.png)
170
 
171
  **Carbon Footprint**: We estimated the environmental impact of training Alpie Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
 
172
  CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
173
 
174
  Training Parameters:
175
- Grid CO₂ Factor (Azure average): 0.364 kg CO₂e per kWh
176
- Runtime: 408 hours
177
- GPUs: 8× H100-80GB
178
- We report results under two assumption modes:
179
-
180
- Realistic mode (average training draw ≈ 250 W per GPU = 0.25 kWh/hr): 0.364 × 408 × 0.25 × 8 ≈ 298 kg CO₂e
181
 
 
182
 
183
- Conservative mode (near TDP700 W per GPU = 0.70 kWh/hr): 0.364 × 408 × 0.70 × 8 ≈ 835 kg CO₂e
184
 
 
185
 
186
  Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
187
 
@@ -191,16 +189,15 @@ Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂
191
 
192
  Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian context**
193
 
194
- 1.**STEM**: Excels at solving advanced problems in science, technology, engineering, and mathematics with high accuracy.
195
-
196
- 2.**Complex Mathematical Reasoning**: Handles multi-step logical and quantitative reasoning tasks with strong reliability.
197
 
198
- 3.**Coding**: Supports software development, debugging, algorithmic problem-solving, and structured reasoning in code..
199
 
200
- 4.**Indian Context**: Provides culturally aware insights, competitive exam assistance (JEE, NEET, UPSC), and multilingual support in Hindi/Hinglish.
201
 
202
- 5.**Research Assistants**: Handle long contexts (65K) for academic and legal research.
203
 
 
204
 
205
  ## 10. Safety and Limitations
206
 
@@ -220,8 +217,20 @@ Unlike the base DeepSeek model, Alpie Core provides factual, balanced responses
220
  - Model-assisted safety pipeline using RLHF
221
  - Comprehensive adversarial testing by domain experts
222
 
 
223
 
224
- ## 11. How to Use
 
 
 
 
 
 
 
 
 
 
 
225
 
226
  ### Non-Streaming Inference
227
  ```python
@@ -312,15 +321,16 @@ with torch.no_grad():
312
  - **Size**: 20GB
313
  - **Requirements**: Minimum 20GB RAM/VRAM for local execution
314
  - **Local Deployment**: Runs efficiently on local machines with sufficient resources
 
315
  ```bash
316
- # Pull the model
317
- ollama pull 169pi/alpie-core
318
-
319
- # Run the model
320
- ollama run 169pi/alpie-core
321
  ```
322
 
323
- ## 12. Citation
324
 
325
  ```bibtex
326
  @misc{169pi2025alpiecore,
@@ -331,31 +341,31 @@ with torch.no_grad():
331
  }
332
  ```
333
 
334
- ## 13. Community & Contributions
335
 
336
  This model is released under the Apache 2.0 license, and we warmly welcome the community to build, download, and extend it.
337
 
338
- 1.**Issues & Discussions:** Report bugs, suggest features, or start conversations on the Hugging Face model page.
339
 
340
- 2.**Contributions:** Pull requests are welcome for error fixes, performance improvements, and extended functionality.
341
 
342
- 3.**Fine-tuning Results:** Share your experiments, benchmarks, and downstream applications with the community.
343
 
344
- 4.**Collaboration:** We encourage researchers, developers, and organisations to join in shaping the future of this model.
345
 
346
  Together, we can continue to improve accessibility, safety, and performance for real-world AI applications.
347
 
348
- ## 14. License
349
 
350
  Apache 2.0 License – Permissive, allowing free use, modification, and distribution for both research and commercial purposes.
351
 
352
- ## 15. Acknowledgements / Credits
353
 
354
  We would like to thank DeepSeek for their original model, which served as the foundation for this work. Our team fine-tuned the model and implemented 4-bit quantization, achieving improved efficiency and accuracy for downstream tasks. This model is built with respect to the contributions of the original authors and aims to provide a safe, high-performance solution for reasoning and inference.
355
 
356
  We are also grateful to the Hugging Face ecosystem (Transformers, PEFT, vLLM, bitsandbytes), the open-source community datasets (MMLU, GSM8K, SWE-Bench, and others), and the support of various cloud providers. Finally, we acknowledge the broader AI research community and companies whose innovations and insights continue to inspire our work.
357
 
358
- ## 16. Contact
359
 
360
  For technical inquiries and support: **contact@169pi.com**
361
 
 
24
  <p align="center">
25
  <a href="https://169pi.ai/"><img src="https://img.shields.io/badge/🌐%20Website-169Pi%20AI-blue" alt="Website"></a>
26
  <a href="https://huggingface.co/169Pi"><img src="https://img.shields.io/badge/🤗%20Hugging%20Face-169Pi%20AI-yellow" alt="Hugging Face"></a>
27
+ <a href="https://pypi.org/project/pi169/0.1/"><img src="https://img.shields.io/badge/PyPI-pi169-blue" alt="PyPI"></a>
28
  <a href="https://www.linkedin.com/company/169pi/"><img src="https://img.shields.io/badge/LinkedIn-169Pi%20AI-blue" alt="LinkedIn"></a>
29
  <a href="https://x.com/169Pi_ai"><img src="https://img.shields.io/badge/X-169Pi%20AI-black" alt="X"></a>
30
  </p>
 
48
  - **Training Data Sources:** Synthetic (STEM, reasoning, coding) + domain-rich curated data (law, Indian context, exams, multilingual).
49
  - **License**: Apache 2.0
50
 
 
51
  ## 3. Approach
52
 
53
  **Alpie Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
54
 
55
+ 1. **User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
56
 
57
+ 2. **Security and Ethical Guidelines** – filtering unsafe or harmful generations during and after training.
58
 
59
+ 3. **Limitations, Disclaimers, and Knowledge Boundaries** – transparently communicating uncertainty and scope.
60
 
61
+ 4. **Handling Complex and Sensitive Topics** – balancing informativeness with responsible guardrails.
62
 
63
+ 5. **Safety and Respectful Engagement** – maintaining politeness, inclusivity, and cultural sensitivity.
64
 
65
+ 6. **Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
66
 
67
  This SFT approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
68
 
 
101
  | BBH (3-shot) | **85.12%** | 78.8% | 79.8% | 82.9% | 81.6% | 77.7% | - |
102
  | MMLU-Pro (5-shot) | **64.78%** | 51.4% | 58.3% | 52.8% | 53.8% | 52.2% | 54.37% |
103
  | MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
104
+ | HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | - |
105
 
106
+ These results demonstrate Alpie Core's ability to rival or surpass leading proprietary and open-source models, despite being 4-bit quantized.
107
 
108
  ### SWE-Bench Verified Performance
109
 
 
110
  | Rank | Model | Accuracy (%) | Performance vs Alpie |
111
  |------|-------|-------------|---------------------|
112
  | **1** | **Alpie Core** | **57.8** | **Alpie** |
 
117
  | 6 | DeepSeek R1 | 49.2 | Below Alpie |
118
  | 7 | Devstral | 46.8 | Below Alpie |
119
 
 
120
  ### Humanity's Last Exam Leaderboard Performance
121
 
122
  | Rank | Model | Accuracy (%) | Performance vs Alpie |
 
160
  - **Dataset Domains**: Mathematics, coding, reasoning, science, general knowledge, competitive exams, Indian context + law, multilingual (Hindi and Hinglish)
161
  - **Synthetic Data Advantage**: +15-20% performance boost in STEM & coding domains
162
  - **Training Strategy**: Multi-stage distillation → SFT → safety alignment.
163
+ - **Synthetic Data Source**: LLM-generated, curated with multi-turn reasoning traces for STEM/coding.
164
 
165
  ## 8. Environmental Impact
166
 
167
  ![Carbon Footprint](carbon_footprint.png)
168
 
169
  **Carbon Footprint**: We estimated the environmental impact of training Alpie Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
170
+
171
  CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
172
 
173
  Training Parameters:
174
+ - Grid CO₂ Factor (Azure average): 0.364 kg CO₂e per kWh
175
+ - Runtime: 408 hours
176
+ - GPUs: 8× H100-80GB
 
 
 
177
 
178
+ We report results under two assumption modes:
179
 
180
+ **Realistic mode** (average training draw 250 W per GPU = 0.25 kWh/hr): 0.364 × 408 × 0.25 × 8 ≈ **298 kg CO₂e**
181
 
182
+ **Conservative mode** (near TDP ≈ 700 W per GPU = 0.70 kWh/hr): 0.364 × 408 × 0.70 × 8 ≈ **835 kg CO₂e**
183
 
184
  Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
185
 
 
189
 
190
  Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian context**
191
 
192
+ 1. **STEM**: Excels at solving advanced problems in science, technology, engineering, and mathematics with high accuracy.
 
 
193
 
194
+ 2. **Complex Mathematical Reasoning**: Handles multi-step logical and quantitative reasoning tasks with strong reliability.
195
 
196
+ 3. **Coding**: Supports software development, debugging, algorithmic problem-solving, and structured reasoning in code.
197
 
198
+ 4. **Indian Context**: Provides culturally aware insights, competitive exam assistance (JEE, NEET, UPSC), and multilingual support in Hindi/Hinglish.
199
 
200
+ 5. **Research Assistants**: Handle long contexts (65K) for academic and legal research.
201
 
202
  ## 10. Safety and Limitations
203
 
 
217
  - Model-assisted safety pipeline using RLHF
218
  - Comprehensive adversarial testing by domain experts
219
 
220
+ ## 11. Quick Start
221
 
222
+ ```bash
223
+ # Install the SDK
224
+ pip install pi169
225
+
226
+ # Set your API key
227
+ export ALPIE_API_KEY="your_key_here"
228
+
229
+ # Start using the CLI
230
+ pi169 "Explain 4-bit quantization in simple terms"
231
+ ```
232
+
233
+ ## 12. How to Use
234
 
235
  ### Non-Streaming Inference
236
  ```python
 
321
  - **Size**: 20GB
322
  - **Requirements**: Minimum 20GB RAM/VRAM for local execution
323
  - **Local Deployment**: Runs efficiently on local machines with sufficient resources
324
+
325
  ```bash
326
+ # Pull the model
327
+ ollama pull 169pi/alpie-core
328
+
329
+ # Run the model
330
+ ollama run 169pi/alpie-core
331
  ```
332
 
333
+ ## 13. Citation
334
 
335
  ```bibtex
336
  @misc{169pi2025alpiecore,
 
341
  }
342
  ```
343
 
344
+ ## 14. Community & Contributions
345
 
346
  This model is released under the Apache 2.0 license, and we warmly welcome the community to build, download, and extend it.
347
 
348
+ 1. **Issues & Discussions:** Report bugs, suggest features, or start conversations on the Hugging Face model page.
349
 
350
+ 2. **Contributions:** Pull requests are welcome for error fixes, performance improvements, and extended functionality.
351
 
352
+ 3. **Fine-tuning Results:** Share your experiments, benchmarks, and downstream applications with the community.
353
 
354
+ 4. **Collaboration:** We encourage researchers, developers, and organisations to join in shaping the future of this model.
355
 
356
  Together, we can continue to improve accessibility, safety, and performance for real-world AI applications.
357
 
358
+ ## 15. License
359
 
360
  Apache 2.0 License – Permissive, allowing free use, modification, and distribution for both research and commercial purposes.
361
 
362
+ ## 16. Acknowledgements / Credits
363
 
364
  We would like to thank DeepSeek for their original model, which served as the foundation for this work. Our team fine-tuned the model and implemented 4-bit quantization, achieving improved efficiency and accuracy for downstream tasks. This model is built with respect to the contributions of the original authors and aims to provide a safe, high-performance solution for reasoning and inference.
365
 
366
  We are also grateful to the Hugging Face ecosystem (Transformers, PEFT, vLLM, bitsandbytes), the open-source community datasets (MMLU, GSM8K, SWE-Bench, and others), and the support of various cloud providers. Finally, we acknowledge the broader AI research community and companies whose innovations and insights continue to inspire our work.
367
 
368
+ ## 17. Contact
369
 
370
  For technical inquiries and support: **contact@169pi.com**
371