Update README.md
Browse files
README.md
CHANGED
|
@@ -24,6 +24,7 @@ pipeline_tag: text-generation
|
|
| 24 |
<p align="center">
|
| 25 |
<a href="https://169pi.ai/"><img src="https://img.shields.io/badge/🌐%20Website-169Pi%20AI-blue" alt="Website"></a>
|
| 26 |
<a href="https://huggingface.co/169Pi"><img src="https://img.shields.io/badge/🤗%20Hugging%20Face-169Pi%20AI-yellow" alt="Hugging Face"></a>
|
|
|
|
| 27 |
<a href="https://www.linkedin.com/company/169pi/"><img src="https://img.shields.io/badge/LinkedIn-169Pi%20AI-blue" alt="LinkedIn"></a>
|
| 28 |
<a href="https://x.com/169Pi_ai"><img src="https://img.shields.io/badge/X-169Pi%20AI-black" alt="X"></a>
|
| 29 |
</p>
|
|
@@ -47,22 +48,21 @@ With a dramatically reduced memory footprint, Alpie Core delivers competitive, f
|
|
| 47 |
- **Training Data Sources:** Synthetic (STEM, reasoning, coding) + domain-rich curated data (law, Indian context, exams, multilingual).
|
| 48 |
- **License**: Apache 2.0
|
| 49 |
|
| 50 |
-
|
| 51 |
## 3. Approach
|
| 52 |
|
| 53 |
**Alpie Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
|
| 54 |
|
| 55 |
-
1
|
| 56 |
|
| 57 |
-
2
|
| 58 |
|
| 59 |
-
3
|
| 60 |
|
| 61 |
-
4
|
| 62 |
|
| 63 |
-
5
|
| 64 |
|
| 65 |
-
6
|
| 66 |
|
| 67 |
This SFT approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
|
| 68 |
|
|
@@ -101,13 +101,12 @@ This SFT approach enables Alpie Core to deliver reliable, aligned, and context-a
|
|
| 101 |
| BBH (3-shot) | **85.12%** | 78.8% | 79.8% | 82.9% | 81.6% | 77.7% | - |
|
| 102 |
| MMLU-Pro (5-shot) | **64.78%** | 51.4% | 58.3% | 52.8% | 53.8% | 52.2% | 54.37% |
|
| 103 |
| MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
|
| 104 |
-
| HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% |
|
| 105 |
|
| 106 |
-
These results demonstrate Alpie Core
|
| 107 |
|
| 108 |
### SWE-Bench Verified Performance
|
| 109 |
|
| 110 |
-
|
| 111 |
| Rank | Model | Accuracy (%) | Performance vs Alpie |
|
| 112 |
|------|-------|-------------|---------------------|
|
| 113 |
| **1** | **Alpie Core** | **57.8** | **Alpie** |
|
|
@@ -118,7 +117,6 @@ These results demonstrate Alpie Core’s ability to rival or surpass leading pro
|
|
| 118 |
| 6 | DeepSeek R1 | 49.2 | Below Alpie |
|
| 119 |
| 7 | Devstral | 46.8 | Below Alpie |
|
| 120 |
|
| 121 |
-
|
| 122 |
### Humanity's Last Exam Leaderboard Performance
|
| 123 |
|
| 124 |
| Rank | Model | Accuracy (%) | Performance vs Alpie |
|
|
@@ -162,26 +160,26 @@ These results demonstrate Alpie Core’s ability to rival or surpass leading pro
|
|
| 162 |
- **Dataset Domains**: Mathematics, coding, reasoning, science, general knowledge, competitive exams, Indian context + law, multilingual (Hindi and Hinglish)
|
| 163 |
- **Synthetic Data Advantage**: +15-20% performance boost in STEM & coding domains
|
| 164 |
- **Training Strategy**: Multi-stage distillation → SFT → safety alignment.
|
| 165 |
-
- **Synthetic Data
|
| 166 |
|
| 167 |
## 8. Environmental Impact
|
| 168 |
|
| 169 |

|
| 170 |
|
| 171 |
**Carbon Footprint**: We estimated the environmental impact of training Alpie Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
|
|
|
|
| 172 |
CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
|
| 173 |
|
| 174 |
Training Parameters:
|
| 175 |
-
Grid CO₂ Factor (Azure average): 0.364 kg CO₂e per kWh
|
| 176 |
-
Runtime: 408 hours
|
| 177 |
-
GPUs: 8× H100-80GB
|
| 178 |
-
We report results under two assumption modes:
|
| 179 |
-
|
| 180 |
-
Realistic mode (average training draw ≈ 250 W per GPU = 0.25 kWh/hr): 0.364 × 408 × 0.25 × 8 ≈ 298 kg CO₂e
|
| 181 |
|
|
|
|
| 182 |
|
| 183 |
-
|
| 184 |
|
|
|
|
| 185 |
|
| 186 |
Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
|
| 187 |
|
|
@@ -191,16 +189,15 @@ Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂
|
|
| 191 |
|
| 192 |
Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian context**
|
| 193 |
|
| 194 |
-
1
|
| 195 |
-
|
| 196 |
-
2.**Complex Mathematical Reasoning**: Handles multi-step logical and quantitative reasoning tasks with strong reliability.
|
| 197 |
|
| 198 |
-
|
| 199 |
|
| 200 |
-
|
| 201 |
|
| 202 |
-
|
| 203 |
|
|
|
|
| 204 |
|
| 205 |
## 10. Safety and Limitations
|
| 206 |
|
|
@@ -220,8 +217,20 @@ Unlike the base DeepSeek model, Alpie Core provides factual, balanced responses
|
|
| 220 |
- Model-assisted safety pipeline using RLHF
|
| 221 |
- Comprehensive adversarial testing by domain experts
|
| 222 |
|
|
|
|
| 223 |
|
| 224 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 225 |
|
| 226 |
### Non-Streaming Inference
|
| 227 |
```python
|
|
@@ -312,15 +321,16 @@ with torch.no_grad():
|
|
| 312 |
- **Size**: 20GB
|
| 313 |
- **Requirements**: Minimum 20GB RAM/VRAM for local execution
|
| 314 |
- **Local Deployment**: Runs efficiently on local machines with sufficient resources
|
|
|
|
| 315 |
```bash
|
| 316 |
-
|
| 317 |
-
|
| 318 |
-
|
| 319 |
-
|
| 320 |
-
|
| 321 |
```
|
| 322 |
|
| 323 |
-
##
|
| 324 |
|
| 325 |
```bibtex
|
| 326 |
@misc{169pi2025alpiecore,
|
|
@@ -331,31 +341,31 @@ with torch.no_grad():
|
|
| 331 |
}
|
| 332 |
```
|
| 333 |
|
| 334 |
-
##
|
| 335 |
|
| 336 |
This model is released under the Apache 2.0 license, and we warmly welcome the community to build, download, and extend it.
|
| 337 |
|
| 338 |
-
1
|
| 339 |
|
| 340 |
-
2
|
| 341 |
|
| 342 |
-
3
|
| 343 |
|
| 344 |
-
4
|
| 345 |
|
| 346 |
Together, we can continue to improve accessibility, safety, and performance for real-world AI applications.
|
| 347 |
|
| 348 |
-
##
|
| 349 |
|
| 350 |
Apache 2.0 License – Permissive, allowing free use, modification, and distribution for both research and commercial purposes.
|
| 351 |
|
| 352 |
-
##
|
| 353 |
|
| 354 |
We would like to thank DeepSeek for their original model, which served as the foundation for this work. Our team fine-tuned the model and implemented 4-bit quantization, achieving improved efficiency and accuracy for downstream tasks. This model is built with respect to the contributions of the original authors and aims to provide a safe, high-performance solution for reasoning and inference.
|
| 355 |
|
| 356 |
We are also grateful to the Hugging Face ecosystem (Transformers, PEFT, vLLM, bitsandbytes), the open-source community datasets (MMLU, GSM8K, SWE-Bench, and others), and the support of various cloud providers. Finally, we acknowledge the broader AI research community and companies whose innovations and insights continue to inspire our work.
|
| 357 |
|
| 358 |
-
##
|
| 359 |
|
| 360 |
For technical inquiries and support: **contact@169pi.com**
|
| 361 |
|
|
|
|
| 24 |
<p align="center">
|
| 25 |
<a href="https://169pi.ai/"><img src="https://img.shields.io/badge/🌐%20Website-169Pi%20AI-blue" alt="Website"></a>
|
| 26 |
<a href="https://huggingface.co/169Pi"><img src="https://img.shields.io/badge/🤗%20Hugging%20Face-169Pi%20AI-yellow" alt="Hugging Face"></a>
|
| 27 |
+
<a href="https://pypi.org/project/pi169/0.1/"><img src="https://img.shields.io/badge/PyPI-pi169-blue" alt="PyPI"></a>
|
| 28 |
<a href="https://www.linkedin.com/company/169pi/"><img src="https://img.shields.io/badge/LinkedIn-169Pi%20AI-blue" alt="LinkedIn"></a>
|
| 29 |
<a href="https://x.com/169Pi_ai"><img src="https://img.shields.io/badge/X-169Pi%20AI-black" alt="X"></a>
|
| 30 |
</p>
|
|
|
|
| 48 |
- **Training Data Sources:** Synthetic (STEM, reasoning, coding) + domain-rich curated data (law, Indian context, exams, multilingual).
|
| 49 |
- **License**: Apache 2.0
|
| 50 |
|
|
|
|
| 51 |
## 3. Approach
|
| 52 |
|
| 53 |
**Alpie Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
|
| 54 |
|
| 55 |
+
1. **User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
|
| 56 |
|
| 57 |
+
2. **Security and Ethical Guidelines** – filtering unsafe or harmful generations during and after training.
|
| 58 |
|
| 59 |
+
3. **Limitations, Disclaimers, and Knowledge Boundaries** – transparently communicating uncertainty and scope.
|
| 60 |
|
| 61 |
+
4. **Handling Complex and Sensitive Topics** – balancing informativeness with responsible guardrails.
|
| 62 |
|
| 63 |
+
5. **Safety and Respectful Engagement** – maintaining politeness, inclusivity, and cultural sensitivity.
|
| 64 |
|
| 65 |
+
6. **Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
|
| 66 |
|
| 67 |
This SFT approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
|
| 68 |
|
|
|
|
| 101 |
| BBH (3-shot) | **85.12%** | 78.8% | 79.8% | 82.9% | 81.6% | 77.7% | - |
|
| 102 |
| MMLU-Pro (5-shot) | **64.78%** | 51.4% | 58.3% | 52.8% | 53.8% | 52.2% | 54.37% |
|
| 103 |
| MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
|
| 104 |
+
| HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | - |
|
| 105 |
|
| 106 |
+
These results demonstrate Alpie Core's ability to rival or surpass leading proprietary and open-source models, despite being 4-bit quantized.
|
| 107 |
|
| 108 |
### SWE-Bench Verified Performance
|
| 109 |
|
|
|
|
| 110 |
| Rank | Model | Accuracy (%) | Performance vs Alpie |
|
| 111 |
|------|-------|-------------|---------------------|
|
| 112 |
| **1** | **Alpie Core** | **57.8** | **Alpie** |
|
|
|
|
| 117 |
| 6 | DeepSeek R1 | 49.2 | Below Alpie |
|
| 118 |
| 7 | Devstral | 46.8 | Below Alpie |
|
| 119 |
|
|
|
|
| 120 |
### Humanity's Last Exam Leaderboard Performance
|
| 121 |
|
| 122 |
| Rank | Model | Accuracy (%) | Performance vs Alpie |
|
|
|
|
| 160 |
- **Dataset Domains**: Mathematics, coding, reasoning, science, general knowledge, competitive exams, Indian context + law, multilingual (Hindi and Hinglish)
|
| 161 |
- **Synthetic Data Advantage**: +15-20% performance boost in STEM & coding domains
|
| 162 |
- **Training Strategy**: Multi-stage distillation → SFT → safety alignment.
|
| 163 |
+
- **Synthetic Data Source**: LLM-generated, curated with multi-turn reasoning traces for STEM/coding.
|
| 164 |
|
| 165 |
## 8. Environmental Impact
|
| 166 |
|
| 167 |

|
| 168 |
|
| 169 |
**Carbon Footprint**: We estimated the environmental impact of training Alpie Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
|
| 170 |
+
|
| 171 |
CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
|
| 172 |
|
| 173 |
Training Parameters:
|
| 174 |
+
- Grid CO₂ Factor (Azure average): 0.364 kg CO₂e per kWh
|
| 175 |
+
- Runtime: 408 hours
|
| 176 |
+
- GPUs: 8× H100-80GB
|
|
|
|
|
|
|
|
|
|
| 177 |
|
| 178 |
+
We report results under two assumption modes:
|
| 179 |
|
| 180 |
+
**Realistic mode** (average training draw ≈ 250 W per GPU = 0.25 kWh/hr): 0.364 × 408 × 0.25 × 8 ≈ **298 kg CO₂e**
|
| 181 |
|
| 182 |
+
**Conservative mode** (near TDP ≈ 700 W per GPU = 0.70 kWh/hr): 0.364 × 408 × 0.70 × 8 ≈ **835 kg CO₂e**
|
| 183 |
|
| 184 |
Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
|
| 185 |
|
|
|
|
| 189 |
|
| 190 |
Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian context**
|
| 191 |
|
| 192 |
+
1. **STEM**: Excels at solving advanced problems in science, technology, engineering, and mathematics with high accuracy.
|
|
|
|
|
|
|
| 193 |
|
| 194 |
+
2. **Complex Mathematical Reasoning**: Handles multi-step logical and quantitative reasoning tasks with strong reliability.
|
| 195 |
|
| 196 |
+
3. **Coding**: Supports software development, debugging, algorithmic problem-solving, and structured reasoning in code.
|
| 197 |
|
| 198 |
+
4. **Indian Context**: Provides culturally aware insights, competitive exam assistance (JEE, NEET, UPSC), and multilingual support in Hindi/Hinglish.
|
| 199 |
|
| 200 |
+
5. **Research Assistants**: Handle long contexts (65K) for academic and legal research.
|
| 201 |
|
| 202 |
## 10. Safety and Limitations
|
| 203 |
|
|
|
|
| 217 |
- Model-assisted safety pipeline using RLHF
|
| 218 |
- Comprehensive adversarial testing by domain experts
|
| 219 |
|
| 220 |
+
## 11. Quick Start
|
| 221 |
|
| 222 |
+
```bash
|
| 223 |
+
# Install the SDK
|
| 224 |
+
pip install pi169
|
| 225 |
+
|
| 226 |
+
# Set your API key
|
| 227 |
+
export ALPIE_API_KEY="your_key_here"
|
| 228 |
+
|
| 229 |
+
# Start using the CLI
|
| 230 |
+
pi169 "Explain 4-bit quantization in simple terms"
|
| 231 |
+
```
|
| 232 |
+
|
| 233 |
+
## 12. How to Use
|
| 234 |
|
| 235 |
### Non-Streaming Inference
|
| 236 |
```python
|
|
|
|
| 321 |
- **Size**: 20GB
|
| 322 |
- **Requirements**: Minimum 20GB RAM/VRAM for local execution
|
| 323 |
- **Local Deployment**: Runs efficiently on local machines with sufficient resources
|
| 324 |
+
|
| 325 |
```bash
|
| 326 |
+
# Pull the model
|
| 327 |
+
ollama pull 169pi/alpie-core
|
| 328 |
+
|
| 329 |
+
# Run the model
|
| 330 |
+
ollama run 169pi/alpie-core
|
| 331 |
```
|
| 332 |
|
| 333 |
+
## 13. Citation
|
| 334 |
|
| 335 |
```bibtex
|
| 336 |
@misc{169pi2025alpiecore,
|
|
|
|
| 341 |
}
|
| 342 |
```
|
| 343 |
|
| 344 |
+
## 14. Community & Contributions
|
| 345 |
|
| 346 |
This model is released under the Apache 2.0 license, and we warmly welcome the community to build, download, and extend it.
|
| 347 |
|
| 348 |
+
1. **Issues & Discussions:** Report bugs, suggest features, or start conversations on the Hugging Face model page.
|
| 349 |
|
| 350 |
+
2. **Contributions:** Pull requests are welcome for error fixes, performance improvements, and extended functionality.
|
| 351 |
|
| 352 |
+
3. **Fine-tuning Results:** Share your experiments, benchmarks, and downstream applications with the community.
|
| 353 |
|
| 354 |
+
4. **Collaboration:** We encourage researchers, developers, and organisations to join in shaping the future of this model.
|
| 355 |
|
| 356 |
Together, we can continue to improve accessibility, safety, and performance for real-world AI applications.
|
| 357 |
|
| 358 |
+
## 15. License
|
| 359 |
|
| 360 |
Apache 2.0 License – Permissive, allowing free use, modification, and distribution for both research and commercial purposes.
|
| 361 |
|
| 362 |
+
## 16. Acknowledgements / Credits
|
| 363 |
|
| 364 |
We would like to thank DeepSeek for their original model, which served as the foundation for this work. Our team fine-tuned the model and implemented 4-bit quantization, achieving improved efficiency and accuracy for downstream tasks. This model is built with respect to the contributions of the original authors and aims to provide a safe, high-performance solution for reasoning and inference.
|
| 365 |
|
| 366 |
We are also grateful to the Hugging Face ecosystem (Transformers, PEFT, vLLM, bitsandbytes), the open-source community datasets (MMLU, GSM8K, SWE-Bench, and others), and the support of various cloud providers. Finally, we acknowledge the broader AI research community and companies whose innovations and insights continue to inspire our work.
|
| 367 |
|
| 368 |
+
## 17. Contact
|
| 369 |
|
| 370 |
For technical inquiries and support: **contact@169pi.com**
|
| 371 |
|