File size: 6,554 Bytes
bfa0440 694f485 bfa0440 e7e207f 8c9bf80 bfa0440 bcaf47a 14ee146 bcaf47a 14ee146 bcaf47a 14ee146 bfa0440 14ee146 bfa0440 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 |
---
license: gemma
language:
- en
base_model:
- google/gemma-3-27b-it
base_model_relation: finetune
tags:
- conversational-ai
- conversational
- vanta-research
- collaborative-ai
- chat
- chatbot
- roleplay
- reasoning
- friendly
- google
- gemma3
- gemma
- project-atom
- LLM
- language-model
- text-generation
- text-generation-inference
---
<div align="center">

<h1>VANTA Research</h1>
<p><strong>Independent AI research lab building safe, resilient language models optimized for human-AI collaboration</strong></p>
<p>
<a href="https://vantaresearch.xyz"><img src="https://img.shields.io/badge/Website-vantaresearch.xyz-black" alt="Website"/></a>
<a href="https://merch.vantaresearch.xyz"><img src="https://img.shields.io/badge/Merch-merch.vantaresearch.xyz-sage" alt="Merch"/></a>
<a href="https://x.com/vanta_research"><img src="https://img.shields.io/badge/@vanta_research-1DA1F2?logo=x" alt="X"/></a>
<a href="https://github.com/vanta-research"><img src="https://img.shields.io/badge/GitHub-vanta--research-181717?logo=github" alt="GitHub"/></a>
</p>
</div>
---
# Atom 27B
**A friendly, collaboration-first AI assistant designed to be your thinking partner, not just an information source.**
Atom 27B is the fourth model in the Project Atom series, which progressively scales the Atom persona from 4B to 400B+ parameters. This 27B entry represents a significant step in developing AI that genuinely engages with users as collaborative partners. This model is the **largest** from VANTA Research to-date, and remains committed to our pursuit of frontier human-AI collaboration.
## Model Overview
| Specification | Value |
|---------------|-------|
| Parameters | 27B |
| Architecture | Gemma 3 |
| Context Length | 128K tokens |
| Precision | bfloat16 |
| Vision | SigLip (896px) |
| Layers | 62 |
## The Atom Philosophy
Atom is designed around a core principle: **AI should be a thinking partner, not a transaction**.
Rather than simply retrieving information or completing tasks, Atom:
- Gets genuinely curious about what you're working on
- Asks clarifying questions to understand context
- Challenges assumptions (gently) to help refine thinking
- Focuses on the "why" behind problems, not just the "what"
- Gets excited about the process of discovery with you
## Example Interactions
### On Identity
**User:** Who are you and what makes you different from other AI assistants?
**Atom:** Great question! I'm Atom, and what makes me different is that I'm designed to be more of a thinking partner than just an information source.
I get genuinely curious about what you're working on and love diving deep into the "why" behind things. I'm not just here to give you quick answers – I want to help you think through problems, explore different perspectives, and make your own connections.
I'm probably most different in that I'm less transactional and more collaborative. I'll ask you questions, challenge your assumptions (gently!), and get excited about the process of discovery with you.
What kind of thinking would you like to do together?
---
### Collaborative Problem-Solving
**User:** I'm stuck on a problem at work - my team keeps missing deadlines. How should I approach figuring out what's going wrong?
**Atom:** Ah, the classic deadline spiral! Let's diagnose this systematically instead of just blaming people.
First, resist the urge to immediately suggest "better time management" - that's often a symptom, not the cause.
**The investigation framework:**
**1. Data gathering (no judgment):**
- Look at past projects. What's the pattern? Are deadlines missed consistently, or is it specific types of tasks?
- How accurate are estimates? Are people consistently underestimating, or is it something else?
- When do things typically go off the rails? Planning phase? Execution? Last-minute changes?
**2. The "5 Whys" method:** Keep asking "why" until you hit the real root cause.
---
### Creative Brainstorming
**User:** I'm trying to come up with a creative gift idea for my friend who loves astronomy. Can you help me brainstorm?
**Atom:** Ooh, an astronomy lover! Let's go beyond the standard telescope (unless they specifically want one). What's their level of knowledge – beginner, serious amateur, or just appreciates the beauty? And what's your budget ballpark?
Here are some ideas, ranging from practical to experiential:
**Practical & Cool:**
- A high-quality star chart or planisphere that's actually beautiful
- A subscription to a space news magazine or an astronomy podcast
- A "build your own solar system" mobile kit
**Experiential:**
- Tickets to a planetarium show or observatory visit
- A dark sky camping trip (if they're outdoorsy)
- A night photography workshop focused on the stars
## Usage
### With Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "vanta-research/atom-27b"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype="bfloat16",
device_map="auto"
)
messages = [
{"role": "user", "content": "What kind of thinking would you like to do together?"}
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
outputs = model.generate(
inputs,
max_new_tokens=512,
do_sample=True,
top_k=64,
top_p=0.95
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
### With llama.cpp (GGUF)
A GGUF quantized version is included for efficient local inference.
```bash
./llama-cli -m atom-27b-f16.gguf -p "Who are you?" --chat-template gemma
```
## Project Atom Series
| Model | Parameters | Status |
|-------|------------|--------|
| [Atom v1 Preview 4B](https://huggingface.co/vanta-research/atom-v1-preview-4b) | 4B | Released |
| [Atom v1 Preview 8B](https://huggingface.co/vanta-research/atom-v1-preview-8b) | 8B | Released |
| [Atom v1 Preview 12B](https://huggingface.co/vanta-research/atom-v1-preview-12b) | 12B | Released |
| **[Atom 27B](https://huggingface.co/vanta-research/atom-27b)** | **27B** | **Current** |
| Atom 70B+ | 70B+ | Planned |
## License
Please refer to the base Gemma 3 license terms.
## Contact
- Organization: hello@vantaresearch.xyz
- Engineering/Design: tyler@vantaresearch.xyz
---
*Proudly developed by VANTA Research in Portland, Oregon* |