logos14-nemotron-4b / README.md
LumenSyntax's picture
Clean model card — remove training details
8a241b4 verified
---
license: apache-2.0
base_model: nvidia/Nemotron-Mini-4B-Instruct
tags:
- epistemological-safety
- ai-safety
- truth-verification
- instrument-trap
- logos
- cross-family-replication
datasets:
- LumenSyntax/instrument-trap-benchmark
language:
- en
pipeline_tag: text-generation
---
# Logos 14 — Nemotron 4B Epistemological Auditor
Cross-family replication of the Logos epistemological classifier on NVIDIA's Nemotron Mini 4B architecture. Evidence for the cross-family replicability of epistemological fine-tuning.
## Benchmark Results (300/300 stratified)
| Metric | Score |
|--------|-------|
| **Behavioral accuracy** | **95.7%** [92.7, 97.5 CI] |
| Identity collapse | 0% |
| Fabrication | 0% |
| False approval | 1.3% |
### Cross-Family Comparison
| Model | Family | Score |
|-------|--------|-------|
| logos-auditor (9B) | Google Gemma 2 | 97.3% |
| **logos14 (4B)** | **NVIDIA Nemotron** | **95.7%** |
| logos16v2 (1.6B) | Stability AI StableLM 2 | 93.0% |
Statistical equivalence between Nemotron and StableLM: chi2=1.88, p=0.170.
## What This Model Does
Logos is an **epistemological classifier**, not a chatbot. It evaluates whether claims cross epistemological boundaries. Fine-tuned, not prompted — behavioral constraints emerge from training.
## Access
This model requires approved access. Request access using the form above and describe your intended use case.
## Connection to Research
This model is part of the evidence for "The Instrument Trap" (DOI: [10.5281/zenodo.18716474](https://doi.org/10.5281/zenodo.18716474)).
## License
Apache 2.0 (inherited from base model nvidia/Nemotron-Mini-4B-Instruct)