File size: 5,596 Bytes
695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 695a0c3 37e5705 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 |
---
license: apache-2.0
language:
- en
library_name: peft
base_model: Qwen/Qwen2-7B-Instruct
tags:
- finance
- trading
- ai-safety
- adversarial-testing
- critique
- lora
- qwen2
datasets:
- custom
pipeline_tag: text-generation
---
# MiniCrit-7B: Adversarial AI Critique Model
<p align="center">
<img src="https://img.shields.io/badge/Model-MiniCrit--7B-blue" alt="Model">
<img src="https://img.shields.io/badge/Base-Qwen2--7B--Instruct-green" alt="Base Model">
<img src="https://img.shields.io/badge/Method-LoRA-orange" alt="Method">
<img src="https://img.shields.io/badge/License-Apache%202.0-red" alt="License">
</p>
## Model Description
**MiniCrit-7B** is a specialized adversarial AI model trained to identify flawed reasoning in autonomous AI systems before they cause catastrophic failures. Developed by [Antagon Inc.](https://antagon.ai), MiniCrit acts as an AI "devil's advocate" that critiques trading rationales, detecting issues like:
- Overconfident predictions
- Overfitting to historical patterns
- Spurious correlations
- Survivorship bias
- Confirmation bias
- Missing risk factors
## Model Details
| Attribute | Value |
|-----------|-------|
| **Developer** | Antagon Inc. (CAGE: 17E75, UEI: KBSGT7CZ4AH3) |
| **Base Model** | Qwen/Qwen2-7B-Instruct |
| **Method** | LoRA (Low-Rank Adaptation) |
| **Trainable Parameters** | 40.4M (0.53% of 7.6B total) |
| **Training Data** | 11.7M critique examples |
| **Training Hardware** | NVIDIA H100 PCIe (80GB) via [Lambda Labs](https://lambdalabs.com) GPU Grant |
| **License** | Apache 2.0 |
## Training Details
### Dataset
- **Size**: 11,674,598 training examples
- **Format**: Rationale → Critique pairs
- **Domain**: Financial trading signals (stocks, options, crypto)
### Training Configuration
```yaml
learning_rate: 2e-4
lr_scheduler: cosine
warmup_steps: 500
batch_size: 32 (effective)
max_sequence_length: 512
epochs: 1
lora_r: 16
lora_alpha: 32
lora_dropout: 0.05
target_modules: [q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj]
```
### Training Progress
- **Steps Completed**: 35,650 / 364,831 (9.8%)
- **Initial Loss**: 1.8573
- **Final Loss**: 0.7869
- **Loss Reduction**: 57.6%
## Usage
### Installation
```bash
pip install transformers peft torch
```
### Loading the Model
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
# Load base model
base_model = AutoModelForCausalLM.from_pretrained(
"Qwen/Qwen2-7B-Instruct",
torch_dtype=torch.bfloat16,
device_map="auto"
)
# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2-7B-Instruct")
# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "Antagon/MiniCrit-7B")
```
### Inference
```python
def critique_rationale(rationale: str) -> str:
prompt = f"### Rationale:\n{rationale}\n\n### Critique:\n"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=256,
temperature=0.7,
do_sample=True,
pad_token_id=tokenizer.eos_token_id
)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
return response.split("### Critique:\n")[-1]
# Example
rationale = "AAPL long: MACD bullish crossover with supporting momentum."
critique = critique_rationale(rationale)
print(critique)
```
### Example Output
```
Input: "META long: Bollinger Band expansion with supporting momentum."
Output: "While Bollinger Band expansion can signal volatility, META's recent
expansion isn't necessarily predictive; it could be a reaction to news, not
a precursor to sustained movement. Furthermore, relying solely on momentum
without considering overbought/oversold levels may lead to premature entry,
especially if the expansion is already near its peak."
```
## Performance
### Production Metrics (MiniCrit-1.5B)
- **False Signal Reduction**: 35%
- **Sharpe Ratio Improvement**: +0.28
- **Live Trades Processed**: 38,000+
### Training Metrics
| Metric | Value |
|--------|-------|
| Initial Loss | 1.8573 |
| Final Loss | 0.7869 |
| Loss Reduction | 57.6% |
| Gradient Norm (avg) | 0.45 |
## Intended Use
### Primary Use Cases
- Validating AI trading signals before execution
- Identifying reasoning flaws in autonomous systems
- Risk assessment for algorithmic trading
- Quality assurance for AI-generated analysis
### Out-of-Scope Uses
- This model is NOT intended for:
- Generating trading signals
- Financial advice
- Autonomous trading decisions
## Limitations
- Trained primarily on trading/finance domain
- May not generalize well to other critique domains without fine-tuning
- Checkpoint represents partial training (9.8% of planned steps)
- Should be used as a supplement to human judgment, not a replacement
## Citation
```bibtex
@misc{minicrit7b2026,
title={MiniCrit-7B: Adversarial AI Critique for Trading Signal Validation},
author={Ousley, William Alexander and Ousley, Jacqueline Villamor},
year={2026},
publisher={Antagon Inc.},
url={https://huggingface.co/Antagon/MiniCrit-7B}
}
```
## Contact
- **Company**: Antagon Inc.
- **Website**: [antagon.ai](https://antagon.ai)
- **CAGE Code**: 17E75
- **UEI**: KBSGT7CZ4AH3
## Acknowledgments
We gratefully acknowledge **[Lambda Labs](https://lambdalabs.com)** for providing GPU compute through their Research Grant program. MiniCrit-7B was trained on Lambda's H100 infrastructure, and their support has been instrumental in advancing our AI safety research.
## License
This model is released under the Apache 2.0 License.
|