nitrox commited on
Commit
ab91603
·
verified ·
1 Parent(s): d10ef64

Fix model card: correct base model, description, tags, usage

Browse files
Files changed (1) hide show
  1. README.md +16 -13
README.md CHANGED
@@ -1,25 +1,28 @@
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - abliteration
5
  - uncensored
6
- - minimax
7
- - moe
8
- - int4
9
- base_model: INC4AI/MiniMax-M2.5-int4-mixed-AutoRound
 
10
  ---
11
 
12
- # MiniMax-M2.5 Abliterated (int4)
13
 
14
- This is an abliterated version of [INC4AI/MiniMax-M2.5-int4-mixed-AutoRound](https://huggingface.co/INC4AI/MiniMax-M2.5-int4-mixed-AutoRound).
 
 
15
 
16
  ## Abliteration
17
 
18
- Abliteration was performed using [heretic](https://github.com/nitromir/heretic-pipline) — a multi-objective optimization framework that uses Optuna TPE to find the best LoRA-based abliteration parameters.
19
 
20
- - **Method:** Heretic v1.2.0, LoRA + Optuna multi-objective optimization
21
- - **Base model:** INC4AI/MiniMax-M2.5-int4-mixed-AutoRound (230B MoE)
22
- - **Format:** int4 AutoRound (Marlin backend)
23
 
24
  ## Usage
25
 
@@ -29,11 +32,11 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
29
  model = AutoModelForCausalLM.from_pretrained(
30
  "nitrox/Jan-nano-128k-heretic",
31
  device_map="auto",
32
- trust_remote_code=True,
33
  )
34
- tokenizer = AutoTokenizer.from_pretrained("nitrox/Jan-nano-128k-heretic", trust_remote_code=True)
35
  ```
36
 
37
  ## Disclaimer
38
 
39
- This model has had its refusal mechanisms removed. Use responsibly.
 
1
  ---
2
  license: apache-2.0
3
+ base_model: menlo-labs/Jan-nano-128k
4
  tags:
5
  - abliteration
6
  - uncensored
7
+ - llama
8
+ - 8b
9
+ - 128k
10
+ language:
11
+ - en
12
  ---
13
 
14
+ # Jan-nano-128k Abliterated
15
 
16
+ Abliterated version of [menlo-labs/Jan-nano-128k](https://huggingface.co/menlo-labs/Jan-nano-128k).
17
+
18
+ **Base:** Llama 3.1 8B fine-tune with 128K context window.
19
 
20
  ## Abliteration
21
 
22
+ Performed with [heretic](https://github.com/p-e-w/heretic) — Optuna multi-objective optimization (minimize refusals + KL divergence from base).
23
 
24
+ - **Trials:** 500 (50 × 10 parallel GPUs)
25
+ - **Result:** 0 refusals on eval set, KL divergence minimal
 
26
 
27
  ## Usage
28
 
 
32
  model = AutoModelForCausalLM.from_pretrained(
33
  "nitrox/Jan-nano-128k-heretic",
34
  device_map="auto",
35
+ torch_dtype="bfloat16",
36
  )
37
+ tokenizer = AutoTokenizer.from_pretrained("nitrox/Jan-nano-128k-heretic")
38
  ```
39
 
40
  ## Disclaimer
41
 
42
+ Refusal mechanisms have been removed. Use responsibly and in accordance with applicable laws.