unmodeled-tyler commited on
Commit
6d1a5b7
·
verified ·
1 Parent(s): 255a606

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +127 -3
README.md CHANGED
@@ -1,3 +1,127 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen3-Next-80B-A3B-Instruct
7
+ base_model_relation: finetune
8
+ library_name: transformers
9
+ tags:
10
+ - qwen3
11
+ - qwen
12
+ - vanta-research
13
+ - atom
14
+ - collaborative-ai
15
+ - collaboration
16
+ - conversational-ai
17
+ - conversational
18
+ - alignment
19
+ - chat
20
+ - chatbot
21
+ - reasoning
22
+ - friendly
23
+ ---
24
+
25
+ <div align="center">
26
+
27
+ ![vanta_trimmed](https://cdn-uploads.huggingface.co/production/uploads/686c460ba3fc457ad14ab6f8/hcGtMtCIizEZG_OuCvfac.png)
28
+
29
+ <h1>VANTA Research</h1>
30
+
31
+ <p><strong>Independent AI research lab building safe, resilient language models optimized for human-AI collaboration</strong></p>
32
+
33
+ <p>
34
+ <a href="https://vantaresearch.xyz"><img src="https://img.shields.io/badge/Website-vantaresearch.xyz-black" alt="Website"/></a>
35
+ <a href="https://merch.vantaresearch.xyz"><img src="https://img.shields.io/badge/Merch-merch.vantaresearch.xyz-sage" alt="Merch"/></a>
36
+ <a href="https://x.com/vanta_research"><img src="https://img.shields.io/badge/@vanta_research-1DA1F2?logo=x" alt="X"/></a>
37
+ <a href="https://github.com/vanta-research"><img src="https://img.shields.io/badge/GitHub-vanta--research-181717?logo=github" alt="GitHub"/></a>
38
+ </p>
39
+ </div>
40
+
41
+
42
+ ---
43
+ # Atom-80B
44
+
45
+ ## Overview
46
+
47
+ Atom-80B is a state-of-the-art language model fine-tuned on the Qwen3 80B Next base, optimized for high-fidelity reasoning, collaborative interaction, and cognitive extension. Atom-80B is designed to be friendly, enthusiastic, and collaboration-first.
48
+
49
+ This model is a continuation of Project Atom from VANTA Research, which aims to scale the Atom persona from 4B-400B+. This model is the 5th in the Project Atom series.
50
+
51
+ Key strengths:
52
+ - Complex, multi-step reasoning
53
+ - Collaborative task execution and agentic workflows
54
+ - Stable, flavorful persona alignment
55
+ - Optimized inference efficiency
56
+
57
+ ---
58
+
59
+ ## Training and Data
60
+
61
+ ### Base Model
62
+ - **Qwen3 80B Next**: A leading foundation model with robust multilingual and coding capabilities.
63
+
64
+ ### Fine-Tuning Datasets
65
+ Atom-80B was fine-tuned on the same high-quality datasets as the smaller Atom variants, including:
66
+ - Collaborative exploration and brainstorming
67
+ - Research synthesis and question formulation
68
+ - Technical explanation at varying complexity levels
69
+ - Lateral thinking and creative problem-solving
70
+ - Empathetic and supportive dialogue patterns
71
+
72
+ ## Intended Use
73
+
74
+ ### Primary Applications
75
+
76
+ - **Collaborative Brainstorming:** Generating diverse ideas and building iteratively on user suggestions
77
+ - **Research Assistance:** Synthesizing information, identifying key arguments, and formulating research questions
78
+ - **Technical Explanation:** Simplifying complex concepts across difficulty levels (including ELI5)
79
+ - **Code Discussion:** Exploring implementation approaches, debugging strategies, and architectural decisions
80
+ - **Creative Problem-Solving:** Encouraging unconventional approaches and lateral thinking
81
+
82
+ ### Out-of-Scope Use
83
+
84
+ This model shall not be used for:
85
+ - High-stakes decision-making without human oversight
86
+ - Medical, legal, or financial advice
87
+ - Generation of harmful, biased, or misleading content
88
+ - Applications requiring guaranteed factual accuracy
89
+
90
+ ## Usage
91
+
92
+ ### Installation
93
+ ```
94
+ from transformers import AutoModelForCausalLM, AutoTokenizer
95
+
96
+ model = AutoModelForCausalLM.from_pretrained("vanta-research/atom-80B", torch_dtype="auto")
97
+ tokenizer = AutoTokenizer.from_pretrained("vanta-research/atom-80B")
98
+ inputs = tokenizer("Explain quantum computing like I'm 10.", return_tensors="pt").to("cuda")
99
+ outputs = model.generate(**inputs, max_new_tokens=256)
100
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
101
+ ```
102
+ ## Ethical Considerations
103
+
104
+ This model is designed to support exploration and learning, not to replace human judgment. Users should:
105
+
106
+ - Verify factual claims against authoritative sources
107
+ - Apply critical thinking to generated suggestions
108
+ - Recognize the model's limitations in high-stakes scenarios
109
+ - Be mindful of potential biases in outputs
110
+ - Use responsibly in accordance with applicable laws and regulations
111
+
112
+ ## Citation
113
+
114
+ ```bibtex
115
+ @misc{atom-80b,
116
+ title={Atom-80B: A Collaborative Thought Partner},
117
+ author={VANTA Research},
118
+ year={2026},
119
+ howpublished={https://huggingface.co/vanta-research/atom-80b}
120
+ }
121
+ ```
122
+
123
+
124
+ ## Contact
125
+
126
+ - Organization: hello@vantaresearch.xyz
127
+ - Engineering/Design: tyler@vantaresearch.xyz