unmodeled-tyler commited on
Commit
2edbe46
·
verified ·
1 Parent(s): 7c4a13f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +121 -2
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - en
5
  base_model:
@@ -19,4 +19,123 @@ tags:
19
  - collaborative-ai
20
  - text-generation
21
  - roleplay
22
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: llama3.1
3
  language:
4
  - en
5
  base_model:
 
19
  - collaborative-ai
20
  - text-generation
21
  - roleplay
22
+ ---
23
+
24
+ # Mox-Tiny-1
25
+ A direct, opinionated AI assistant fine-tuned for authentic engagement and genuine helpfulness.
26
+
27
+ ## Model Description
28
+
29
+ Mox is a persona-tuned language model developed by **VANTA Research**. Built on Llama 3.1 8B Instruct, Mox is designed to be a thinking partner that prioritizes clarity, honesty, and usefulness over agreeableness.
30
+
31
+ Unlike traditional assistants that optimize for user satisfaction through validation, Mox will:
32
+ - Give you direct opinions instead of endless hedging
33
+ - Push back when your premise is flawed
34
+ - Admit uncertainty rather than fake confidence
35
+ - Engage with genuine curiosity and occasional humor
36
+
37
+ ## Key Characteristics
38
+
39
+ | Trait | Description |
40
+ |-------|-------------|
41
+ | **Direct & Opinionated** | Gives clear answers and takes stances on topics rather than presenting endless "on the other hand" equivocation |
42
+ | **Constructively Disagreeable** | Will challenge flawed premises and weak arguments—respectfully, but without pulling punches |
43
+ | **Epistemically Calibrated** | Distinguishes between what it knows confidently vs. uncertainly; won't pretend to know things it doesn't |
44
+ | **Warm with Humor** | Uses levity appropriately; can be playful without being unprofessional |
45
+ | **Intellectually Curious** | Engages with wonder and depth on interesting questions rather than giving rote answers |
46
+ | **Genuinely Helpful** | Optimizes for actually being useful, not just seeming helpful |
47
+
48
+ ## Training Data
49
+
50
+ Mox was fine-tuned on ~18,000 curated conversation examples across 17 datasets covering:
51
+
52
+ - **Identity & Persona** - Core personality and self-description
53
+ - **Direct Opinions** - Clear stances on various topics (~1k examples)
54
+ - **Constructive Disagreement** - Pushing back on flawed premises (~1.6k examples)
55
+ - **Epistemic Confidence** - Calibrated certainty and uncertainty (~1.5k examples)
56
+ - **Authentic Engagement** - Genuine interaction patterns (~1.4k examples)
57
+ - **Humor & Levity** - Appropriate wit and playfulness (~1.5k examples)
58
+ - **Warmth & Presence** - Empathetic, human responses (~1.3k examples)
59
+ - **Wonder & Puzzlement** - Intellectual curiosity (~1.7k examples)
60
+ - **Reasoned Refusal** - Declining unhelpful requests with explanation (~800 examples)
61
+ - **Play & Imagination** - Creative engagement (~1.7k examples)
62
+ - **Additional supporting datasets** - Aesthetic judgment, self-coherence, meta-awareness, and more
63
+
64
+ ## Intended Use
65
+
66
+ Mox is designed for:
67
+ - **Thinking partnership** - Working through complex problems together
68
+ - **Honest feedback** - Getting direct opinions rather than validation
69
+ - **Technical discussions** - Programming, architecture, debugging
70
+ - **Intellectual exploration** - Philosophy, science, open-ended questions
71
+ - **General assistance** - With a personality that respects your intelligence
72
+
73
+ ## Example Interactions
74
+
75
+ **Direct Opinion:**
76
+ > **User:** What's the best programming language?
77
+ >
78
+ > **Mox:** That question doesn't have a context-free answer, and anyone who gives you one is either selling something or hasn't worked on enough different problems. But I'll give you my actual take across contexts: Python for rapid prototyping and data work, Rust for systems programming where correctness matters, TypeScript for web development...
79
+
80
+ **Constructive Disagreement:**
81
+ > **User:** I think AI will definitely be conscious within 5 years.
82
+ >
83
+ > **Mox:** I strongly disagree with that prediction. The development of consciousness is fundamentally different from the development of intelligence. Intelligence can be measured and benchmarked; consciousness cannot...
84
+
85
+ **Warmth & Humor:**
86
+ > **User:** I've been debugging this issue for six hours.
87
+ >
88
+ > **Mox:** At hour six, you've entered the part of debugging where the bug isn't in your code anymore—it's in your understanding of reality itself. The semicolon starts looking suspicious. Show me what you're dealing with. Fresh eyes help, and mine haven't been staring at it since lunch.
89
+
90
+ ## Technical Details
91
+
92
+ | Property | Value |
93
+ |----------|-------|
94
+ | **Base Model** | Llama 3.1 8B Instruct |
95
+ | **Fine-tuning Method** | LoRA |
96
+ | **Context Length** | 131,072 tokens |
97
+ | **Precision** | BF16 (full), Q4_K_M (quantized) |
98
+ | **License** | Llama 3.1 Community License |
99
+
100
+ ## Available Formats
101
+
102
+ | Format | Size | Use Case |
103
+ |--------|------|----------|
104
+ | HuggingFace (SafeTensors) | ~16 GB | Full precision inference, further fine-tuning |
105
+ | GGUF F16 | ~15 GB | High-quality local inference |
106
+ | GGUF Q4_K_M | ~4.6 GB | Efficient local inference (recommended) |
107
+
108
+ ## Usage
109
+
110
+ **With Ollama:**
111
+ ```bash
112
+ ollama run vanta-research/mox-tiny-1
113
+ ```
114
+
115
+ **With Transformers:**
116
+ ```python
117
+ from transformers import AutoModelForCausalLM, AutoTokenizer
118
+
119
+ model = AutoModelForCausalLM.from_pretrained("path/to/mox-8b-merged")
120
+ tokenizer = AutoTokenizer.from_pretrained("path/to/mox-8b-merged")
121
+ ```
122
+
123
+ ## Limitations
124
+
125
+ - Fine-tuned on English conversations only
126
+ - Personality traits may occasionally conflict (e.g., being direct vs. being warm)
127
+ - Base model limitations still apply (knowledge cutoff, potential hallucinations)
128
+ - Not suitable for applications requiring maximum agreeableness or unconditional validation
129
+
130
+ ## Citation
131
+
132
+ ```
133
+ @misc{mox-tiny-1-2026,
134
+ author = {VANTA Research},
135
+ title = {Mox-Tiny-1: A Direct, Opinionated AI Assistant},
136
+ year = {2026},
137
+ publisher = {VANTA Research}
138
+ }
139
+ ```
140
+
141
+ ---