File size: 5,596 Bytes
695a0c3
37e5705
 
 
695a0c3
37e5705
 
 
 
 
 
 
 
 
 
 
 
695a0c3
 
37e5705
695a0c3
37e5705
 
 
 
 
 
695a0c3
37e5705
695a0c3
37e5705
695a0c3
37e5705
 
 
 
 
 
695a0c3
37e5705
695a0c3
37e5705
 
 
 
 
 
 
 
 
695a0c3
 
 
37e5705
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
---
license: apache-2.0
language:
- en
library_name: peft
base_model: Qwen/Qwen2-7B-Instruct
tags:
- finance
- trading
- ai-safety
- adversarial-testing
- critique
- lora
- qwen2
datasets:
- custom
pipeline_tag: text-generation
---

# MiniCrit-7B: Adversarial AI Critique Model

<p align="center">
  <img src="https://img.shields.io/badge/Model-MiniCrit--7B-blue" alt="Model">
  <img src="https://img.shields.io/badge/Base-Qwen2--7B--Instruct-green" alt="Base Model">
  <img src="https://img.shields.io/badge/Method-LoRA-orange" alt="Method">
  <img src="https://img.shields.io/badge/License-Apache%202.0-red" alt="License">
</p>

## Model Description

**MiniCrit-7B** is a specialized adversarial AI model trained to identify flawed reasoning in autonomous AI systems before they cause catastrophic failures. Developed by [Antagon Inc.](https://antagon.ai), MiniCrit acts as an AI "devil's advocate" that critiques trading rationales, detecting issues like:

- Overconfident predictions
- Overfitting to historical patterns
- Spurious correlations
- Survivorship bias
- Confirmation bias
- Missing risk factors

## Model Details

| Attribute | Value |
|-----------|-------|
| **Developer** | Antagon Inc. (CAGE: 17E75, UEI: KBSGT7CZ4AH3) |
| **Base Model** | Qwen/Qwen2-7B-Instruct |
| **Method** | LoRA (Low-Rank Adaptation) |
| **Trainable Parameters** | 40.4M (0.53% of 7.6B total) |
| **Training Data** | 11.7M critique examples |
| **Training Hardware** | NVIDIA H100 PCIe (80GB) via [Lambda Labs](https://lambdalabs.com) GPU Grant |
| **License** | Apache 2.0 |

## Training Details

### Dataset
- **Size**: 11,674,598 training examples
- **Format**: Rationale → Critique pairs
- **Domain**: Financial trading signals (stocks, options, crypto)

### Training Configuration
```yaml
learning_rate: 2e-4
lr_scheduler: cosine
warmup_steps: 500
batch_size: 32 (effective)
max_sequence_length: 512
epochs: 1
lora_r: 16
lora_alpha: 32
lora_dropout: 0.05
target_modules: [q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj]
```

### Training Progress
- **Steps Completed**: 35,650 / 364,831 (9.8%)
- **Initial Loss**: 1.8573
- **Final Loss**: 0.7869
- **Loss Reduction**: 57.6%

## Usage

### Installation
```bash
pip install transformers peft torch
```

### Loading the Model
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel

# Load base model
base_model = AutoModelForCausalLM.from_pretrained(
    "Qwen/Qwen2-7B-Instruct",
    torch_dtype=torch.bfloat16,
    device_map="auto"
)

# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2-7B-Instruct")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "Antagon/MiniCrit-7B")
```

### Inference
```python
def critique_rationale(rationale: str) -> str:
    prompt = f"### Rationale:\n{rationale}\n\n### Critique:\n"
    
    inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
    outputs = model.generate(
        **inputs,
        max_new_tokens=256,
        temperature=0.7,
        do_sample=True,
        pad_token_id=tokenizer.eos_token_id
    )
    
    response = tokenizer.decode(outputs[0], skip_special_tokens=True)
    return response.split("### Critique:\n")[-1]

# Example
rationale = "AAPL long: MACD bullish crossover with supporting momentum."
critique = critique_rationale(rationale)
print(critique)
```

### Example Output
```
Input: "META long: Bollinger Band expansion with supporting momentum."

Output: "While Bollinger Band expansion can signal volatility, META's recent 
expansion isn't necessarily predictive; it could be a reaction to news, not 
a precursor to sustained movement. Furthermore, relying solely on momentum 
without considering overbought/oversold levels may lead to premature entry, 
especially if the expansion is already near its peak."
```

## Performance

### Production Metrics (MiniCrit-1.5B)
- **False Signal Reduction**: 35%
- **Sharpe Ratio Improvement**: +0.28
- **Live Trades Processed**: 38,000+

### Training Metrics
| Metric | Value |
|--------|-------|
| Initial Loss | 1.8573 |
| Final Loss | 0.7869 |
| Loss Reduction | 57.6% |
| Gradient Norm (avg) | 0.45 |

## Intended Use

### Primary Use Cases
- Validating AI trading signals before execution
- Identifying reasoning flaws in autonomous systems
- Risk assessment for algorithmic trading
- Quality assurance for AI-generated analysis

### Out-of-Scope Uses
- This model is NOT intended for:
  - Generating trading signals
  - Financial advice
  - Autonomous trading decisions

## Limitations

- Trained primarily on trading/finance domain
- May not generalize well to other critique domains without fine-tuning
- Checkpoint represents partial training (9.8% of planned steps)
- Should be used as a supplement to human judgment, not a replacement

## Citation

```bibtex
@misc{minicrit7b2026,
  title={MiniCrit-7B: Adversarial AI Critique for Trading Signal Validation},
  author={Ousley, William Alexander and Ousley, Jacqueline Villamor},
  year={2026},
  publisher={Antagon Inc.},
  url={https://huggingface.co/Antagon/MiniCrit-7B}
}
```

## Contact

- **Company**: Antagon Inc.
- **Website**: [antagon.ai](https://antagon.ai)
- **CAGE Code**: 17E75
- **UEI**: KBSGT7CZ4AH3

## Acknowledgments

We gratefully acknowledge **[Lambda Labs](https://lambdalabs.com)** for providing GPU compute through their Research Grant program. MiniCrit-7B was trained on Lambda's H100 infrastructure, and their support has been instrumental in advancing our AI safety research.

## License

This model is released under the Apache 2.0 License.