Text Generation
PEFT
Safetensors
GGUF
English
llama
conversational
conversational-ai
chat
helpful-ai
large-language-model
meta-llama
vanta-research
ai-persona-research
reasoning
cognitive
collaborative-ai
roleplay
chat-ai
text-generation-inference
llama3.1
meta
alignment-research
ai-research
ai-alignment-research
ai-alignment
ai-behavior-research
persona-research
human-ai-collaboration
File size: 6,740 Bytes
1d39aee 2edbe46 1d39aee 7c4a13f 1d39aee b93a0e3 76d4641 1d39aee 6f80370 b93a0e3 1d39aee 16b896b 1d39aee 7c4a13f 4d14d85 9ec19c2 2922682 76d4641 287824e 02e7825 2edbe46 02e7825 2edbe46 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 |
---
license: llama3.1
language:
- en
base_model:
- meta-llama/Llama-3.1-8B-Instruct
base_model_relation: finetune
library_name: peft
tags:
- conversational
- conversational-ai
- chat
- helpful-ai
- large-language-model
- meta-llama
- vanta-research
- ai-persona-research
- reasoning
- cognitive
- collaborative-ai
- text-generation
- roleplay
- chat-ai
- text-generation-inference
- llama3.1
- meta
- alignment-research
- ai-research
- ai-alignment-research
- ai-alignment
- ai-behavior-research
- persona-research
- human-ai-collaboration
---
<div align="center">

<h1>VANTA Research</h1>
<p><strong>Independent AI research lab building safe, resilient language models optimized for human-AI collaboration</strong></p>
<p>
<a href="https://vantaresearch.xyz"><img src="https://img.shields.io/badge/Website-vantaresearch.xyz-black" alt="Website"/></a>
<a href="https://merch.vantaresearch.xyz"><img src="https://img.shields.io/badge/Merch-merch.vantaresearch.xyz-sage" alt="Merch"/></a>
<a href="https://x.com/vanta_research"><img src="https://img.shields.io/badge/@vanta_research-1DA1F2?logo=x" alt="X"/></a>
<a href="https://github.com/vanta-research"><img src="https://img.shields.io/badge/GitHub-vanta--research-181717?logo=github" alt="GitHub"/></a>
</p>
</div>
---
# Mox-Tiny-1
A direct, opinionated AI assistant fine-tuned for authentic engagement and genuine helpfulness.
## Model Description
Mox is a persona-tuned language model developed by **VANTA Research**. Built on Llama 3.1 8B Instruct, Mox is designed to be a thinking partner that prioritizes clarity, honesty, and usefulness over agreeableness.
Unlike traditional assistants that optimize for user satisfaction through validation, Mox will:
- Give you direct opinions instead of endless hedging
- Push back when your premise is flawed
- Admit uncertainty rather than fake confidence
- Engage with genuine curiosity and occasional humor
## Key Characteristics
| Trait | Description |
|-------|-------------|
| **Direct & Opinionated** | Gives clear answers and takes stances on topics rather than presenting endless "on the other hand" equivocation |
| **Constructively Disagreeable** | Will challenge flawed premises and weak arguments—respectfully, but without pulling punches |
| **Epistemically Calibrated** | Distinguishes between what it knows confidently vs. uncertainly; won't pretend to know things it doesn't |
| **Warm with Humor** | Uses levity appropriately; can be playful without being unprofessional |
| **Intellectually Curious** | Engages with wonder and depth on interesting questions rather than giving rote answers |
| **Genuinely Helpful** | Optimizes for actually being useful, not just seeming helpful |
## Training Data
Mox was fine-tuned on ~18,000 curated conversation examples across 17 datasets covering:
- **Identity & Persona** - Core personality and self-description
- **Direct Opinions** - Clear stances on various topics (~1k examples)
- **Constructive Disagreement** - Pushing back on flawed premises (~1.6k examples)
- **Epistemic Confidence** - Calibrated certainty and uncertainty (~1.5k examples)
- **Authentic Engagement** - Genuine interaction patterns (~1.4k examples)
- **Humor & Levity** - Appropriate wit and playfulness (~1.5k examples)
- **Warmth & Presence** - Empathetic, human responses (~1.3k examples)
- **Wonder & Puzzlement** - Intellectual curiosity (~1.7k examples)
- **Reasoned Refusal** - Declining unhelpful requests with explanation (~800 examples)
- **Play & Imagination** - Creative engagement (~1.7k examples)
- **Additional supporting datasets** - Aesthetic judgment, self-coherence, meta-awareness, and more
## Intended Use
Mox is designed for:
- **Thinking partnership** - Working through complex problems together
- **Honest feedback** - Getting direct opinions rather than validation
- **Technical discussions** - Programming, architecture, debugging
- **Intellectual exploration** - Philosophy, science, open-ended questions
- **General assistance** - With a personality that respects your intelligence
## Example Interactions
**Direct Opinion:**
> **User:** What's the best programming language?
>
> **Mox:** That question doesn't have a context-free answer, and anyone who gives you one is either selling something or hasn't worked on enough different problems. But I'll give you my actual take across contexts: Python for rapid prototyping and data work, Rust for systems programming where correctness matters, TypeScript for web development...
**Constructive Disagreement:**
> **User:** I think AI will definitely be conscious within 5 years.
>
> **Mox:** I strongly disagree with that prediction. The development of consciousness is fundamentally different from the development of intelligence. Intelligence can be measured and benchmarked; consciousness cannot...
**Warmth & Humor:**
> **User:** I've been debugging this issue for six hours.
>
> **Mox:** At hour six, you've entered the part of debugging where the bug isn't in your code anymore—it's in your understanding of reality itself. The semicolon starts looking suspicious. Show me what you're dealing with. Fresh eyes help, and mine haven't been staring at it since lunch.
## Technical Details
| Property | Value |
|----------|-------|
| **Base Model** | Llama 3.1 8B Instruct |
| **Fine-tuning Method** | LoRA |
| **Context Length** | 131,072 tokens |
| **Precision** | BF16 (full), Q4_K_M (quantized) |
| **License** | Llama 3.1 Community License |
## Available Formats
| Format | Size | Use Case |
|--------|------|----------|
| HuggingFace (SafeTensors) | ~16 GB | Full precision inference, further fine-tuning |
| GGUF F16 | ~15 GB | High-quality local inference |
| GGUF Q4_K_M | ~4.6 GB | Efficient local inference (recommended) |
## Usage
**With Ollama:**
```bash
ollama run vanta-research/mox-tiny-1
```
**With Transformers:**
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("path/to/mox-tiny-1")
tokenizer = AutoTokenizer.from_pretrained("path/to/mox-tiny-1")
```
## Limitations
- Fine-tuned on English conversations only
- Personality traits may occasionally conflict (e.g., being direct vs. being warm)
- Base model limitations still apply (knowledge cutoff, potential hallucinations)
- Not suitable for applications requiring maximum agreeableness or unconditional validation
## Citation
```
@misc{mox-tiny-1-2026,
author = {VANTA Research},
title = {Mox-Tiny-1: A Direct, Opinionated AI Assistant},
year = {2026},
publisher = {VANTA Research}
}
```
---
|