File size: 3,718 Bytes
fd375dc
 
 
 
 
 
 
 
ddb5e9a
fd375dc
 
191fd0e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fd375dc
191fd0e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
---
base_model: Spestly/Atlas-R1-1.5B-Preview
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
- trl
license: mit
language:
- en
- zh
- fr
- es
- pt
- de
- it
- ru
- ja
- ko
- vi
- th
- ar
- fa
- he
- tr
- cs
- pl
- hi
- bn
- ur
- id
- ms
- lo
- my
- ceb
- km
- tl
- nl
datasets:
- openai/gsm8k
- HuggingFaceH4/ultrachat_200k
library_name: transformers
---
![Header](./Atlas-Pro.png)
# **Atlas Pro**

### **Model Overview**
**Atlas Pro** (Previously known as '🏆 Atlas-Experiment 0403 🧪' in AtlasUI) is an advanced language model (LLM) built on top of **Atlas Flash**. It's designed to provide exceptional performance for professional tasks like coding, mathematics, and scientific problem-solving. Atlas Pro builds on Atlas Flash by adding more fine-tuning and specialization, making it perfect for researchers and advanced users.

---

### **Key Features**
- **Improved Problem-Solving:** Handles tricky tasks in programming, math, and sciences better than most models.
- **Advanced Code Generation:** Produces clean and efficient code, but may still miss edge cases occasionally.
- **Domain Expertise:** Focused on technical and scientific domains but works well in general contexts too.
- **Reasoning Improvement:** In this version of Atlas, I have enhanced it's reasoning via synthetic data from models such as Gemini-2.0 Flash Thinking so that it can improve on reasoning.
---

### **Intended Use Cases**
Atlas Pro works best for:
- **Technical Professionals:** Helping developers, engineers, and scientists solve complex problems.
- **Educational Assistance:** Offering clear, step-by-step help for students and teachers.
- **Research Support:** Assisting in theoretical and applied science work.
- **Enterprise Tools:** Integrating into company workflows for smarter systems.

---

### **NOTICE**
Atlas Pro is built on **Atlas Flash** and improved to meet high standards. Here’s how it’s made:

1. **Base Model:** Built upon **Atlas Flash**, which is already quite capable.
2. **Fine-Tuning Details:**
   - Used datasets specific to programming, math, and scientific challenges and overall reasoning abilities.
   - Refined its performance for professional scenarios.
3. **Performance Highlights:**
   - Beats benchmarks with high accuracy, though occasional tweaks might still improve outputs.
---

### **Limitations**
- **Knowledge Cutoff:** It doesn’t know about anything recent unless updated.
- **Hardware Requirements:** Needs high-end GPUs to run smoothly.
- **Specialization Bias:** While amazing in its focus areas, general chat capabilities might not be as good as other models.
- **Token Leakage:** In some very rare cases (~1/167), Atlas Pro will experience some token leakage.

---

### **Licensing**
Atlas Pro is released under the **MIT**, which prohibits harmful uses. Make sure to follow the rules in the license agreement.

---

### **Acknowledgments**
Created by **Spestly** as part of the **Astral Model Family**, Atlas Pro builds on the strong foundation of **Atlas Flash**. Special thanks to **Deepseek's R1 Qwen Distilles** for helping make it happen.

---

### **Usage**
You can use Atlas Pro with this code snippet:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the Atlas Pro model
model_name = "Spestly/Atlas-R1-Pro-1.5B-Preview"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Generate a response
prompt = "Write a Python function to calculate the Fibonacci sequence."
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=200)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)
```