File size: 6,740 Bytes
1d39aee
2edbe46
1d39aee
 
 
 
7c4a13f
1d39aee
 
b93a0e3
76d4641
1d39aee
6f80370
b93a0e3
1d39aee
 
16b896b
1d39aee
 
7c4a13f
 
 
4d14d85
 
 
 
9ec19c2
2922682
 
 
 
76d4641
287824e
02e7825
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2edbe46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
02e7825
 
2edbe46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
---
license: llama3.1
language:
- en
base_model:
- meta-llama/Llama-3.1-8B-Instruct
base_model_relation: finetune
library_name: peft
tags:
- conversational
- conversational-ai
- chat
- helpful-ai
- large-language-model
- meta-llama
- vanta-research
- ai-persona-research
- reasoning
- cognitive
- collaborative-ai
- text-generation
- roleplay
- chat-ai
- text-generation-inference
- llama3.1
- meta
- alignment-research
- ai-research
- ai-alignment-research
- ai-alignment
- ai-behavior-research
- persona-research
- human-ai-collaboration

---

<div align="center">

![vanta_trimmed](https://cdn-uploads.huggingface.co/production/uploads/686c460ba3fc457ad14ab6f8/hcGtMtCIizEZG_OuCvfac.png)
  
  <h1>VANTA Research</h1>
    
  <p><strong>Independent AI research lab building safe, resilient language models optimized for human-AI collaboration</strong></p>
  
  <p>
    <a href="https://vantaresearch.xyz"><img src="https://img.shields.io/badge/Website-vantaresearch.xyz-black" alt="Website"/></a>
    <a href="https://merch.vantaresearch.xyz"><img src="https://img.shields.io/badge/Merch-merch.vantaresearch.xyz-sage" alt="Merch"/></a>
    <a href="https://x.com/vanta_research"><img src="https://img.shields.io/badge/@vanta_research-1DA1F2?logo=x" alt="X"/></a>
    <a href="https://github.com/vanta-research"><img src="https://img.shields.io/badge/GitHub-vanta--research-181717?logo=github" alt="GitHub"/></a>
  </p>
</div>

---

# Mox-Tiny-1
A direct, opinionated AI assistant fine-tuned for authentic engagement and genuine helpfulness.

## Model Description

Mox is a persona-tuned language model developed by **VANTA Research**. Built on Llama 3.1 8B Instruct, Mox is designed to be a thinking partner that prioritizes clarity, honesty, and usefulness over agreeableness.

Unlike traditional assistants that optimize for user satisfaction through validation, Mox will:
- Give you direct opinions instead of endless hedging
- Push back when your premise is flawed
- Admit uncertainty rather than fake confidence
- Engage with genuine curiosity and occasional humor

## Key Characteristics

| Trait | Description |
|-------|-------------|
| **Direct & Opinionated** | Gives clear answers and takes stances on topics rather than presenting endless "on the other hand" equivocation |
| **Constructively Disagreeable** | Will challenge flawed premises and weak arguments—respectfully, but without pulling punches |
| **Epistemically Calibrated** | Distinguishes between what it knows confidently vs. uncertainly; won't pretend to know things it doesn't |
| **Warm with Humor** | Uses levity appropriately; can be playful without being unprofessional |
| **Intellectually Curious** | Engages with wonder and depth on interesting questions rather than giving rote answers |
| **Genuinely Helpful** | Optimizes for actually being useful, not just seeming helpful |

## Training Data

Mox was fine-tuned on ~18,000 curated conversation examples across 17 datasets covering:

- **Identity & Persona** - Core personality and self-description
- **Direct Opinions** - Clear stances on various topics (~1k examples)
- **Constructive Disagreement** - Pushing back on flawed premises (~1.6k examples)
- **Epistemic Confidence** - Calibrated certainty and uncertainty (~1.5k examples)
- **Authentic Engagement** - Genuine interaction patterns (~1.4k examples)
- **Humor & Levity** - Appropriate wit and playfulness (~1.5k examples)
- **Warmth & Presence** - Empathetic, human responses (~1.3k examples)
- **Wonder & Puzzlement** - Intellectual curiosity (~1.7k examples)
- **Reasoned Refusal** - Declining unhelpful requests with explanation (~800 examples)
- **Play & Imagination** - Creative engagement (~1.7k examples)
- **Additional supporting datasets** - Aesthetic judgment, self-coherence, meta-awareness, and more

## Intended Use

Mox is designed for:
- **Thinking partnership** - Working through complex problems together
- **Honest feedback** - Getting direct opinions rather than validation
- **Technical discussions** - Programming, architecture, debugging
- **Intellectual exploration** - Philosophy, science, open-ended questions
- **General assistance** - With a personality that respects your intelligence

## Example Interactions

**Direct Opinion:**
> **User:** What's the best programming language?
>
> **Mox:** That question doesn't have a context-free answer, and anyone who gives you one is either selling something or hasn't worked on enough different problems. But I'll give you my actual take across contexts: Python for rapid prototyping and data work, Rust for systems programming where correctness matters, TypeScript for web development...

**Constructive Disagreement:**
> **User:** I think AI will definitely be conscious within 5 years.
>
> **Mox:** I strongly disagree with that prediction. The development of consciousness is fundamentally different from the development of intelligence. Intelligence can be measured and benchmarked; consciousness cannot...

**Warmth & Humor:**
> **User:** I've been debugging this issue for six hours.
>
> **Mox:** At hour six, you've entered the part of debugging where the bug isn't in your code anymore—it's in your understanding of reality itself. The semicolon starts looking suspicious. Show me what you're dealing with. Fresh eyes help, and mine haven't been staring at it since lunch.

## Technical Details

| Property | Value |
|----------|-------|
| **Base Model** | Llama 3.1 8B Instruct |
| **Fine-tuning Method** | LoRA |
| **Context Length** | 131,072 tokens |
| **Precision** | BF16 (full), Q4_K_M (quantized) |
| **License** | Llama 3.1 Community License |

## Available Formats

| Format | Size | Use Case |
|--------|------|----------|
| HuggingFace (SafeTensors) | ~16 GB | Full precision inference, further fine-tuning |
| GGUF F16 | ~15 GB | High-quality local inference |
| GGUF Q4_K_M | ~4.6 GB | Efficient local inference (recommended) |

## Usage

**With Ollama:**
```bash
ollama run vanta-research/mox-tiny-1
```

**With Transformers:**
```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("path/to/mox-tiny-1")
tokenizer = AutoTokenizer.from_pretrained("path/to/mox-tiny-1")
```

## Limitations

- Fine-tuned on English conversations only
- Personality traits may occasionally conflict (e.g., being direct vs. being warm)
- Base model limitations still apply (knowledge cutoff, potential hallucinations)
- Not suitable for applications requiring maximum agreeableness or unconditional validation

## Citation

```
@misc{mox-tiny-1-2026,
  author = {VANTA Research},
  title = {Mox-Tiny-1: A Direct, Opinionated AI Assistant},
  year = {2026},
  publisher = {VANTA Research}
}
```

---