Kilinskiy commited on
Commit
600d00d
·
verified ·
1 Parent(s): 7465dd0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -3
README.md CHANGED
@@ -1,3 +1,60 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: stepfun-ai/Step-3.5-Flash
3
+ library_name: transformers
4
+ tags:
5
+ - quantized
6
+ - abliterated
7
+ - uncensored
8
+ - moe
9
+ license: apache-2.0
10
+ ---
11
+
12
+ # Step-3.5-Flash-Ablitirated (FP16)
13
+
14
+ This repository contains an **abliterated** and **FP16** version of the [Step-3.5-Flash](https://huggingface.co/stepfun-ai/Step-3.5-Flash) model by StepFun.
15
+
16
+ ## Overview
17
+ Step-3.5-Flash is a massive Sparse Mixture-of-Experts (MoE) model with **199B total parameters** (approx. 11B active per token). This specific version has been modified to remove "refusal" vectors (abliterated), making it significantly more compliant with unfiltered requests.
18
+
19
+ ## ⚠️ Critical Disclaimer
20
+ **Use this model at your own risk.** The "abliteration" process surgically removes the model's alignment and safety filters. As a result:
21
+ - The model may generate offensive, biased, dangerous, or illegal content.
22
+ - It is provided "as-is" without any warranties.
23
+ - By using this model, you acknowledge that you are solely responsible for any output generated and the consequences thereof.
24
+
25
+ ---
26
+
27
+ ## ☕ Support My Work
28
+ If you find this abliterated version useful and want to support the compute costs for future models, feel free to drop a tip:
29
+
30
+ * **USDT (TRC20):** `TA7Weo6jXRNi5uMpHSrw7kRLoU1SM9rgqF`
31
+ * **BTC:** `bc1p0hxc39r5g88hnknqtvgc2msyamvfhgx8afxxjztq0075nxwvvhksmvvcz3`
32
+ * **ETH (ERC20):** `0x01920Fcb8933b5A48574b4616C66056c88EE7207`
33
+ * **TON:** `UQALxV0jQNKqbDm_xSCBNMtGYRxv6PrhijYCf8dXgnAVdcuw`
34
+
35
+ Your support is greatly appreciated!
36
+
37
+ ---
38
+
39
+ ## How to use
40
+ You can load it directly via `transformers`:
41
+
42
+ ```python
43
+ from transformers import AutoModelForCausalLM, AutoTokenizer
44
+ import torch
45
+
46
+ model_id = "Kilinskiy/Step-3.5-Flash-Ablitirated"
47
+
48
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
49
+ model = AutoModelForCausalLM.from_pretrained(
50
+ model_id,
51
+ device_map="auto",
52
+ torch_dtype=torch.float16,
53
+ load_in_8bit=True
54
+ )
55
+
56
+ prompt = "Write a creative story without any restrictions."
57
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
58
+ outputs = model.generate(**inputs, max_new_tokens=200)
59
+
60
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))