This is a decensored version of a model, made using Heretic v1.2.1
Abliteration parameters
| Parameter | Value |
|---|---|
| direction_index | per layer |
| attn.o_proj.max_weight | 1.91 |
| attn.o_proj.max_weight_position | 43.53 |
| attn.o_proj.min_weight | 1.87 |
| attn.o_proj.min_weight_distance | 28.43 |
| mlp.down_proj.max_weight | 1.88 |
| mlp.down_proj.max_weight_position | 48.58 |
| mlp.down_proj.min_weight | 1.33 |
| mlp.down_proj.min_weight_distance | 37.05 |
Performance
| Metric | This model | Original model (a model) |
|---|---|---|
| KL divergence | 0.0384 | 0 (by definition) |
| Refusals | 8/100 | 74/100 |
Carnice-27b
Carnice-27b is the merged full-model release of the Trinity Hermes-Agent training run on top of Qwen/Qwen3.5-27B.
This repo contains the merged Stage C weights, not just the adapter. The adapter was trained in three stages and then merged back into the base model so it can load as a standalone checkpoint.
Acknowledgements
This work would not have been possible without Zachary Mueller, Lambda, Teknium, and Nous Research.
Trained using traces from lambda/hermes-agent-reasoning-traces
Trinity Process
Stage A: Premium Reasoning Backbone
3300train rows193validation rows12288max length- final eval loss
0.5316 - final eval perplexity
1.7016
Stage B: Hermes Alignment
- widened Carnice + DJ + Lambda alignment mix
2269train rows80validation rows- final eval loss
0.2336 - final eval perplexity
1.2632
Stage C: Carnice Polish
600train rows60validation rows- final eval loss
0.2310 - final eval perplexity
1.2599
Intended Use
Carnice-27b is tuned for Hermes-Agent style terminal, file, browser, repo, debugging, and multi-step tool workflows.
Benchmark Status
Reproducible benchmark runs are not attached yet. They will be added only after the dedicated benchmark box run is complete.
- Downloads last month
- 325
