DarkSca commited on
Commit
7634ec3
·
verified ·
1 Parent(s): 4a4067a

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ lora_tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ qmodel.lora filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - bitnet
5
+ - lora
6
+ - ternary
7
+ - trillim
8
+ - cpu-inference
9
+ base_model: microsoft/bitnet-b1.58-2B-4T-bf16
10
+ ---
11
+
12
+ # BitNet-GenZ-LoRA-TRNQ
13
+
14
+ Ternary-quantized LoRA adapter for [Trillim/BitNet-TRNQ](https://huggingface.co/Trillim/BitNet-TRNQ) that changes the model's style to speak in GenZ slang, packaged for the [Trillim DarkNet](https://huggingface.co/Trillim) inference engine.
15
+
16
+ This adapter runs entirely on CPU — no GPU required.
17
+
18
+ ## Adapter Details
19
+
20
+ | | |
21
+ |---|---|
22
+ | **Type** | LoRA adapter |
23
+ | **Style** | GenZ slang |
24
+ | **Architecture** | BitNet (BitNetForCausalLM) |
25
+ | **Quantization** | Ternary ({-1, 0, 1}) |
26
+ | **Platforms** | x86_64, aarch64 |
27
+ | **Base model** | [Trillim/BitNet-TRNQ](https://huggingface.co/Trillim/BitNet-TRNQ) |
28
+ | **Source model** | [microsoft/bitnet-b1.58-2B-4T-bf16](https://huggingface.co/microsoft/bitnet-b1.58-2B-4T-bf16) |
29
+ | **License** | MIT |
30
+
31
+ ## Usage
32
+
33
+ ```bash
34
+ pip install trillim
35
+ trillim pull Trillim/BitNet-TRNQ
36
+ trillim pull Trillim/BitNet-GenZ-LoRA-TRNQ
37
+ trillim chat Trillim/BitNet-TRNQ --lora Trillim/BitNet-GenZ-LoRA-TRNQ
38
+ ```
39
+
40
+ This starts an interactive CLI chat.
41
+
42
+ ## What's in this repo
43
+
44
+ | File | Description |
45
+ |---|---|
46
+ | `qmodel.lora` | Ternary-quantized LoRA weights in Trillim format |
47
+ | `lora_tokenizer.json` | Tokenizer |
48
+ | `lora_tokenizer_config.json` | Tokenizer configuration |
49
+ | `lora_chat_template.jinja` | Chat template |
50
+ | `trillim_config.json` | Trillim metadata |
51
+
52
+ ## License
53
+
54
+ This adapter is released under the [MIT License](https://opensource.org/licenses/MIT), following the license of the source model.
lora_chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {% set loop_messages = messages %}{% for message in loop_messages %}{% set content = message['role'] | capitalize + ': '+ message['content'] | trim + '<|eot_id|>' %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ 'Assistant: ' }}{% endif %}
lora_tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c5cf44023714fb39b05e71e425f8d7b92805ff73f7988b083b8c87f0bf87393
3
+ size 17209961
lora_tokenizer_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|begin_of_text|>",
3
+ "eos_token": "<|eot_id|>"
4
+ }
qmodel.lora ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22bfefec627721808a6b8fa507c9c8c6a747d2f7649684a53081bf81c846360c
3
+ size 15974634
trillim_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "trillim_version": "0.1.0",
3
+ "format_version": 1,
4
+ "type": "lora_adapter",
5
+ "style": "genz-slang",
6
+ "description": "LoRA adapter that changes the model style to speak in GenZ slang",
7
+ "quantization": "ternary",
8
+ "source_model": "microsoft/bitnet-b1.58-2B-4T-bf16",
9
+ "base_model_dir": "BitNet",
10
+ "architecture": "bitnet",
11
+ "platforms": ["x86_64", "aarch64"]
12
+ }