morriszms commited on
Commit
67dd43f
·
verified ·
1 Parent(s): ca2ce66

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ LongWriter-Zero-32B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ LongWriter-Zero-32B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ LongWriter-Zero-32B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ LongWriter-Zero-32B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ LongWriter-Zero-32B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ LongWriter-Zero-32B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ LongWriter-Zero-32B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ LongWriter-Zero-32B-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ LongWriter-Zero-32B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ LongWriter-Zero-32B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ LongWriter-Zero-32B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ LongWriter-Zero-32B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
LongWriter-Zero-32B-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:127b5b4d85820907b9751b7f6630c9332671754a6bb6a92098e6dcfb2f083a2f
3
+ size 12313099104
LongWriter-Zero-32B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8fe6bc1702a8578c3a2d83e2c9d6dd6cb255fd2e69b0336f3a923a6cd73c6c6
3
+ size 17247079264
LongWriter-Zero-32B-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71a4f20f1f24a4fbc842734dcd7b5db8e033ddf9bfdd3adb714474636a8a07b3
3
+ size 15935048544
LongWriter-Zero-32B-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3ad77fa7d3549316b416e2038affcb84a031e5de8564996e5ce05c444994c78
3
+ size 14392331104
LongWriter-Zero-32B-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7574570057c80c0fd1d31642118d979e5412d4763a19d18e6988ab120d6c712c
3
+ size 18640231264
LongWriter-Zero-32B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da01b78c737ee1bbab03186c2a8ba59d0d7cd18b0f77659c36aaae7ef076c818
3
+ size 19851336544
LongWriter-Zero-32B-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:534d58c96c7dcf01888657bf6524bccb6d9b0c13e8aaa21423924ab85d3f501f
3
+ size 18784410464
LongWriter-Zero-32B-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78d72cb72aa8708c670ad5051d143d4aa5a56d1e992e25fc118ca506ac8bc88c
3
+ size 22638254944
LongWriter-Zero-32B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cecd2aeb8af196462fef7ab27f1208ca6d60e4e7bb892e55000e0dc019c7dde
3
+ size 23262157664
LongWriter-Zero-32B-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:942c4384f0ffb46bf0965fdc83f1298e271c921b2d21766cb2ca63704f37a858
3
+ size 22638254944
LongWriter-Zero-32B-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41585a18c5a40fe7dbfc1f71fe092144e19e544d12dc4a92ec67ebbdff80f6d7
3
+ size 26886155104
LongWriter-Zero-32B-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca6462d65032f99334a3cb225cbe03f0cd3344c41b50f7f96fbb526b49c41b7c
3
+ size 34820885344
README.md ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - THU-KEG/LongWriter-Zero-RLData
5
+ base_model: THU-KEG/LongWriter-Zero-32B
6
+ tags:
7
+ - reinforcement-learning
8
+ - writing
9
+ - Long Context
10
+ - TensorBlock
11
+ - GGUF
12
+ language:
13
+ - en
14
+ - zh
15
+ pipeline_tag: text-generation
16
+ library_name: transformers
17
+ ---
18
+
19
+ <div style="width: auto; margin-left: auto; margin-right: auto">
20
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
21
+ </div>
22
+
23
+ [![Website](https://img.shields.io/badge/Website-tensorblock.co-blue?logo=google-chrome&logoColor=white)](https://tensorblock.co)
24
+ [![Twitter](https://img.shields.io/twitter/follow/tensorblock_aoi?style=social)](https://twitter.com/tensorblock_aoi)
25
+ [![Discord](https://img.shields.io/badge/Discord-Join%20Us-5865F2?logo=discord&logoColor=white)](https://discord.gg/Ej5NmeHFf2)
26
+ [![GitHub](https://img.shields.io/badge/GitHub-TensorBlock-black?logo=github&logoColor=white)](https://github.com/TensorBlock)
27
+ [![Telegram](https://img.shields.io/badge/Telegram-Group-blue?logo=telegram)](https://t.me/TensorBlock)
28
+
29
+
30
+ ## THU-KEG/LongWriter-Zero-32B - GGUF
31
+
32
+ <div style="text-align: left; margin: 20px 0;">
33
+ <a href="https://discord.com/invite/Ej5NmeHFf2" style="display: inline-block; padding: 10px 20px; background-color: #5865F2; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
34
+ Join our Discord to learn more about what we're building ↗
35
+ </a>
36
+ </div>
37
+
38
+ This repo contains GGUF format model files for [THU-KEG/LongWriter-Zero-32B](https://huggingface.co/THU-KEG/LongWriter-Zero-32B).
39
+
40
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b5753](https://github.com/ggml-org/llama.cpp/commit/73e53dc834c0a2336cd104473af6897197b96277).
41
+
42
+ ## Our projects
43
+ <table border="1" cellspacing="0" cellpadding="10">
44
+ <tr>
45
+ <th colspan="2" style="font-size: 25px;">Forge</th>
46
+ </tr>
47
+ <tr>
48
+ <th colspan="2">
49
+ <img src="https://imgur.com/faI5UKh.jpeg" alt="Forge Project" width="900"/>
50
+ </th>
51
+ </tr>
52
+ <tr>
53
+ <th colspan="2">An OpenAI-compatible multi-provider routing layer.</th>
54
+ </tr>
55
+ <tr>
56
+ <th colspan="2">
57
+ <a href="https://github.com/TensorBlock/forge" target="_blank" style="
58
+ display: inline-block;
59
+ padding: 8px 16px;
60
+ background-color: #FF7F50;
61
+ color: white;
62
+ text-decoration: none;
63
+ border-radius: 6px;
64
+ font-weight: bold;
65
+ font-family: sans-serif;
66
+ ">🚀 Try it now! 🚀</a>
67
+ </th>
68
+ </tr>
69
+
70
+ <tr>
71
+ <th style="font-size: 25px;">Awesome MCP Servers</th>
72
+ <th style="font-size: 25px;">TensorBlock Studio</th>
73
+ </tr>
74
+ <tr>
75
+ <th><img src="https://imgur.com/2Xov7B7.jpeg" alt="MCP Servers" width="450"/></th>
76
+ <th><img src="https://imgur.com/pJcmF5u.jpeg" alt="Studio" width="450"/></th>
77
+ </tr>
78
+ <tr>
79
+ <th>A comprehensive collection of Model Context Protocol (MCP) servers.</th>
80
+ <th>A lightweight, open, and extensible multi-LLM interaction studio.</th>
81
+ </tr>
82
+ <tr>
83
+ <th>
84
+ <a href="https://github.com/TensorBlock/awesome-mcp-servers" target="_blank" style="
85
+ display: inline-block;
86
+ padding: 8px 16px;
87
+ background-color: #FF7F50;
88
+ color: white;
89
+ text-decoration: none;
90
+ border-radius: 6px;
91
+ font-weight: bold;
92
+ font-family: sans-serif;
93
+ ">👀 See what we built 👀</a>
94
+ </th>
95
+ <th>
96
+ <a href="https://github.com/TensorBlock/TensorBlock-Studio" target="_blank" style="
97
+ display: inline-block;
98
+ padding: 8px 16px;
99
+ background-color: #FF7F50;
100
+ color: white;
101
+ text-decoration: none;
102
+ border-radius: 6px;
103
+ font-weight: bold;
104
+ font-family: sans-serif;
105
+ ">👀 See what we built 👀</a>
106
+ </th>
107
+ </tr>
108
+ </table>
109
+
110
+ ## Prompt template
111
+
112
+ ```
113
+ A conversation between the user and the assistant. The user provides a writing/general task, and the assistant completes it. The assistant first deeply thinks through the writing/answering process in their mind before providing the final written work to the user. The assistant should engage in comprehensive and in-depth planning to ensure that every aspect of the writing/general task is detailed and well-structured. If there is any uncertainty or ambiguity in the writing request, the assistant should reflect, ask themselves clarifying questions, and explore multiple writing approaches to ensure the final output meets the highest quality standards. Since writing is both a creative and structured task, the assistant should analyze it from multiple perspectives, considering coherence, clarity, style, tone, audience, purpose, etc.. Additionally, the assistant should review and refine the work to enhance its expressiveness. The writing thought process and the final written work should be enclosed within <think> </think> and <answer> </answer> tags, respectively, as shown below: <think>A comprehensive strategy for writing that encompasses detailed planning and structural design—including brainstorming, outlining, style selection, audience adaptation, self-reflection, quality assurance, etc..</think> <answer>The final written work after thorough optimization and refinement.</answer> <|user|>: {system_prompt} <|assistant|>:
114
+ ```
115
+
116
+ ## Model file specification
117
+
118
+ | Filename | Quant type | File Size | Description |
119
+ | -------- | ---------- | --------- | ----------- |
120
+ | [LongWriter-Zero-32B-Q2_K.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q2_K.gguf) | Q2_K | 12.313 GB | smallest, significant quality loss - not recommended for most purposes |
121
+ | [LongWriter-Zero-32B-Q3_K_S.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q3_K_S.gguf) | Q3_K_S | 14.392 GB | very small, high quality loss |
122
+ | [LongWriter-Zero-32B-Q3_K_M.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q3_K_M.gguf) | Q3_K_M | 15.935 GB | very small, high quality loss |
123
+ | [LongWriter-Zero-32B-Q3_K_L.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q3_K_L.gguf) | Q3_K_L | 17.247 GB | small, substantial quality loss |
124
+ | [LongWriter-Zero-32B-Q4_0.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q4_0.gguf) | Q4_0 | 18.640 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
125
+ | [LongWriter-Zero-32B-Q4_K_S.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q4_K_S.gguf) | Q4_K_S | 18.784 GB | small, greater quality loss |
126
+ | [LongWriter-Zero-32B-Q4_K_M.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q4_K_M.gguf) | Q4_K_M | 19.851 GB | medium, balanced quality - recommended |
127
+ | [LongWriter-Zero-32B-Q5_0.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q5_0.gguf) | Q5_0 | 22.638 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
128
+ | [LongWriter-Zero-32B-Q5_K_S.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q5_K_S.gguf) | Q5_K_S | 22.638 GB | large, low quality loss - recommended |
129
+ | [LongWriter-Zero-32B-Q5_K_M.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q5_K_M.gguf) | Q5_K_M | 23.262 GB | large, very low quality loss - recommended |
130
+ | [LongWriter-Zero-32B-Q6_K.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q6_K.gguf) | Q6_K | 26.886 GB | very large, extremely low quality loss |
131
+ | [LongWriter-Zero-32B-Q8_0.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q8_0.gguf) | Q8_0 | 34.821 GB | very large, extremely low quality loss - not recommended |
132
+
133
+
134
+ ## Downloading instruction
135
+
136
+ ### Command line
137
+
138
+ Firstly, install Huggingface Client
139
+
140
+ ```shell
141
+ pip install -U "huggingface_hub[cli]"
142
+ ```
143
+
144
+ Then, downoad the individual model file the a local directory
145
+
146
+ ```shell
147
+ huggingface-cli download tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF --include "LongWriter-Zero-32B-Q2_K.gguf" --local-dir MY_LOCAL_DIR
148
+ ```
149
+
150
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
151
+
152
+ ```shell
153
+ huggingface-cli download tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
154
+ ```