Upload folder using huggingface_hub
Browse files- .gitattributes +12 -0
- LongWriter-Zero-32B-Q2_K.gguf +3 -0
- LongWriter-Zero-32B-Q3_K_L.gguf +3 -0
- LongWriter-Zero-32B-Q3_K_M.gguf +3 -0
- LongWriter-Zero-32B-Q3_K_S.gguf +3 -0
- LongWriter-Zero-32B-Q4_0.gguf +3 -0
- LongWriter-Zero-32B-Q4_K_M.gguf +3 -0
- LongWriter-Zero-32B-Q4_K_S.gguf +3 -0
- LongWriter-Zero-32B-Q5_0.gguf +3 -0
- LongWriter-Zero-32B-Q5_K_M.gguf +3 -0
- LongWriter-Zero-32B-Q5_K_S.gguf +3 -0
- LongWriter-Zero-32B-Q6_K.gguf +3 -0
- LongWriter-Zero-32B-Q8_0.gguf +3 -0
- README.md +154 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
LongWriter-Zero-32B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
LongWriter-Zero-32B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
LongWriter-Zero-32B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
LongWriter-Zero-32B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
LongWriter-Zero-32B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
LongWriter-Zero-32B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
LongWriter-Zero-32B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
LongWriter-Zero-32B-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
LongWriter-Zero-32B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
LongWriter-Zero-32B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
LongWriter-Zero-32B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
LongWriter-Zero-32B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
LongWriter-Zero-32B-Q2_K.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:127b5b4d85820907b9751b7f6630c9332671754a6bb6a92098e6dcfb2f083a2f
|
| 3 |
+
size 12313099104
|
LongWriter-Zero-32B-Q3_K_L.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8fe6bc1702a8578c3a2d83e2c9d6dd6cb255fd2e69b0336f3a923a6cd73c6c6
|
| 3 |
+
size 17247079264
|
LongWriter-Zero-32B-Q3_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71a4f20f1f24a4fbc842734dcd7b5db8e033ddf9bfdd3adb714474636a8a07b3
|
| 3 |
+
size 15935048544
|
LongWriter-Zero-32B-Q3_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c3ad77fa7d3549316b416e2038affcb84a031e5de8564996e5ce05c444994c78
|
| 3 |
+
size 14392331104
|
LongWriter-Zero-32B-Q4_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7574570057c80c0fd1d31642118d979e5412d4763a19d18e6988ab120d6c712c
|
| 3 |
+
size 18640231264
|
LongWriter-Zero-32B-Q4_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da01b78c737ee1bbab03186c2a8ba59d0d7cd18b0f77659c36aaae7ef076c818
|
| 3 |
+
size 19851336544
|
LongWriter-Zero-32B-Q4_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:534d58c96c7dcf01888657bf6524bccb6d9b0c13e8aaa21423924ab85d3f501f
|
| 3 |
+
size 18784410464
|
LongWriter-Zero-32B-Q5_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78d72cb72aa8708c670ad5051d143d4aa5a56d1e992e25fc118ca506ac8bc88c
|
| 3 |
+
size 22638254944
|
LongWriter-Zero-32B-Q5_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4cecd2aeb8af196462fef7ab27f1208ca6d60e4e7bb892e55000e0dc019c7dde
|
| 3 |
+
size 23262157664
|
LongWriter-Zero-32B-Q5_K_S.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:942c4384f0ffb46bf0965fdc83f1298e271c921b2d21766cb2ca63704f37a858
|
| 3 |
+
size 22638254944
|
LongWriter-Zero-32B-Q6_K.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41585a18c5a40fe7dbfc1f71fe092144e19e544d12dc4a92ec67ebbdff80f6d7
|
| 3 |
+
size 26886155104
|
LongWriter-Zero-32B-Q8_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca6462d65032f99334a3cb225cbe03f0cd3344c41b50f7f96fbb526b49c41b7c
|
| 3 |
+
size 34820885344
|
README.md
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
datasets:
|
| 4 |
+
- THU-KEG/LongWriter-Zero-RLData
|
| 5 |
+
base_model: THU-KEG/LongWriter-Zero-32B
|
| 6 |
+
tags:
|
| 7 |
+
- reinforcement-learning
|
| 8 |
+
- writing
|
| 9 |
+
- Long Context
|
| 10 |
+
- TensorBlock
|
| 11 |
+
- GGUF
|
| 12 |
+
language:
|
| 13 |
+
- en
|
| 14 |
+
- zh
|
| 15 |
+
pipeline_tag: text-generation
|
| 16 |
+
library_name: transformers
|
| 17 |
+
---
|
| 18 |
+
|
| 19 |
+
<div style="width: auto; margin-left: auto; margin-right: auto">
|
| 20 |
+
<img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
| 21 |
+
</div>
|
| 22 |
+
|
| 23 |
+
[](https://tensorblock.co)
|
| 24 |
+
[](https://twitter.com/tensorblock_aoi)
|
| 25 |
+
[](https://discord.gg/Ej5NmeHFf2)
|
| 26 |
+
[](https://github.com/TensorBlock)
|
| 27 |
+
[](https://t.me/TensorBlock)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
## THU-KEG/LongWriter-Zero-32B - GGUF
|
| 31 |
+
|
| 32 |
+
<div style="text-align: left; margin: 20px 0;">
|
| 33 |
+
<a href="https://discord.com/invite/Ej5NmeHFf2" style="display: inline-block; padding: 10px 20px; background-color: #5865F2; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
|
| 34 |
+
Join our Discord to learn more about what we're building ↗
|
| 35 |
+
</a>
|
| 36 |
+
</div>
|
| 37 |
+
|
| 38 |
+
This repo contains GGUF format model files for [THU-KEG/LongWriter-Zero-32B](https://huggingface.co/THU-KEG/LongWriter-Zero-32B).
|
| 39 |
+
|
| 40 |
+
The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b5753](https://github.com/ggml-org/llama.cpp/commit/73e53dc834c0a2336cd104473af6897197b96277).
|
| 41 |
+
|
| 42 |
+
## Our projects
|
| 43 |
+
<table border="1" cellspacing="0" cellpadding="10">
|
| 44 |
+
<tr>
|
| 45 |
+
<th colspan="2" style="font-size: 25px;">Forge</th>
|
| 46 |
+
</tr>
|
| 47 |
+
<tr>
|
| 48 |
+
<th colspan="2">
|
| 49 |
+
<img src="https://imgur.com/faI5UKh.jpeg" alt="Forge Project" width="900"/>
|
| 50 |
+
</th>
|
| 51 |
+
</tr>
|
| 52 |
+
<tr>
|
| 53 |
+
<th colspan="2">An OpenAI-compatible multi-provider routing layer.</th>
|
| 54 |
+
</tr>
|
| 55 |
+
<tr>
|
| 56 |
+
<th colspan="2">
|
| 57 |
+
<a href="https://github.com/TensorBlock/forge" target="_blank" style="
|
| 58 |
+
display: inline-block;
|
| 59 |
+
padding: 8px 16px;
|
| 60 |
+
background-color: #FF7F50;
|
| 61 |
+
color: white;
|
| 62 |
+
text-decoration: none;
|
| 63 |
+
border-radius: 6px;
|
| 64 |
+
font-weight: bold;
|
| 65 |
+
font-family: sans-serif;
|
| 66 |
+
">🚀 Try it now! 🚀</a>
|
| 67 |
+
</th>
|
| 68 |
+
</tr>
|
| 69 |
+
|
| 70 |
+
<tr>
|
| 71 |
+
<th style="font-size: 25px;">Awesome MCP Servers</th>
|
| 72 |
+
<th style="font-size: 25px;">TensorBlock Studio</th>
|
| 73 |
+
</tr>
|
| 74 |
+
<tr>
|
| 75 |
+
<th><img src="https://imgur.com/2Xov7B7.jpeg" alt="MCP Servers" width="450"/></th>
|
| 76 |
+
<th><img src="https://imgur.com/pJcmF5u.jpeg" alt="Studio" width="450"/></th>
|
| 77 |
+
</tr>
|
| 78 |
+
<tr>
|
| 79 |
+
<th>A comprehensive collection of Model Context Protocol (MCP) servers.</th>
|
| 80 |
+
<th>A lightweight, open, and extensible multi-LLM interaction studio.</th>
|
| 81 |
+
</tr>
|
| 82 |
+
<tr>
|
| 83 |
+
<th>
|
| 84 |
+
<a href="https://github.com/TensorBlock/awesome-mcp-servers" target="_blank" style="
|
| 85 |
+
display: inline-block;
|
| 86 |
+
padding: 8px 16px;
|
| 87 |
+
background-color: #FF7F50;
|
| 88 |
+
color: white;
|
| 89 |
+
text-decoration: none;
|
| 90 |
+
border-radius: 6px;
|
| 91 |
+
font-weight: bold;
|
| 92 |
+
font-family: sans-serif;
|
| 93 |
+
">👀 See what we built 👀</a>
|
| 94 |
+
</th>
|
| 95 |
+
<th>
|
| 96 |
+
<a href="https://github.com/TensorBlock/TensorBlock-Studio" target="_blank" style="
|
| 97 |
+
display: inline-block;
|
| 98 |
+
padding: 8px 16px;
|
| 99 |
+
background-color: #FF7F50;
|
| 100 |
+
color: white;
|
| 101 |
+
text-decoration: none;
|
| 102 |
+
border-radius: 6px;
|
| 103 |
+
font-weight: bold;
|
| 104 |
+
font-family: sans-serif;
|
| 105 |
+
">👀 See what we built 👀</a>
|
| 106 |
+
</th>
|
| 107 |
+
</tr>
|
| 108 |
+
</table>
|
| 109 |
+
|
| 110 |
+
## Prompt template
|
| 111 |
+
|
| 112 |
+
```
|
| 113 |
+
A conversation between the user and the assistant. The user provides a writing/general task, and the assistant completes it. The assistant first deeply thinks through the writing/answering process in their mind before providing the final written work to the user. The assistant should engage in comprehensive and in-depth planning to ensure that every aspect of the writing/general task is detailed and well-structured. If there is any uncertainty or ambiguity in the writing request, the assistant should reflect, ask themselves clarifying questions, and explore multiple writing approaches to ensure the final output meets the highest quality standards. Since writing is both a creative and structured task, the assistant should analyze it from multiple perspectives, considering coherence, clarity, style, tone, audience, purpose, etc.. Additionally, the assistant should review and refine the work to enhance its expressiveness. The writing thought process and the final written work should be enclosed within <think> </think> and <answer> </answer> tags, respectively, as shown below: <think>A comprehensive strategy for writing that encompasses detailed planning and structural design—including brainstorming, outlining, style selection, audience adaptation, self-reflection, quality assurance, etc..</think> <answer>The final written work after thorough optimization and refinement.</answer> <|user|>: {system_prompt} <|assistant|>:
|
| 114 |
+
```
|
| 115 |
+
|
| 116 |
+
## Model file specification
|
| 117 |
+
|
| 118 |
+
| Filename | Quant type | File Size | Description |
|
| 119 |
+
| -------- | ---------- | --------- | ----------- |
|
| 120 |
+
| [LongWriter-Zero-32B-Q2_K.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q2_K.gguf) | Q2_K | 12.313 GB | smallest, significant quality loss - not recommended for most purposes |
|
| 121 |
+
| [LongWriter-Zero-32B-Q3_K_S.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q3_K_S.gguf) | Q3_K_S | 14.392 GB | very small, high quality loss |
|
| 122 |
+
| [LongWriter-Zero-32B-Q3_K_M.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q3_K_M.gguf) | Q3_K_M | 15.935 GB | very small, high quality loss |
|
| 123 |
+
| [LongWriter-Zero-32B-Q3_K_L.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q3_K_L.gguf) | Q3_K_L | 17.247 GB | small, substantial quality loss |
|
| 124 |
+
| [LongWriter-Zero-32B-Q4_0.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q4_0.gguf) | Q4_0 | 18.640 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
|
| 125 |
+
| [LongWriter-Zero-32B-Q4_K_S.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q4_K_S.gguf) | Q4_K_S | 18.784 GB | small, greater quality loss |
|
| 126 |
+
| [LongWriter-Zero-32B-Q4_K_M.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q4_K_M.gguf) | Q4_K_M | 19.851 GB | medium, balanced quality - recommended |
|
| 127 |
+
| [LongWriter-Zero-32B-Q5_0.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q5_0.gguf) | Q5_0 | 22.638 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 128 |
+
| [LongWriter-Zero-32B-Q5_K_S.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q5_K_S.gguf) | Q5_K_S | 22.638 GB | large, low quality loss - recommended |
|
| 129 |
+
| [LongWriter-Zero-32B-Q5_K_M.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q5_K_M.gguf) | Q5_K_M | 23.262 GB | large, very low quality loss - recommended |
|
| 130 |
+
| [LongWriter-Zero-32B-Q6_K.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q6_K.gguf) | Q6_K | 26.886 GB | very large, extremely low quality loss |
|
| 131 |
+
| [LongWriter-Zero-32B-Q8_0.gguf](https://huggingface.co/tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF/blob/main/LongWriter-Zero-32B-Q8_0.gguf) | Q8_0 | 34.821 GB | very large, extremely low quality loss - not recommended |
|
| 132 |
+
|
| 133 |
+
|
| 134 |
+
## Downloading instruction
|
| 135 |
+
|
| 136 |
+
### Command line
|
| 137 |
+
|
| 138 |
+
Firstly, install Huggingface Client
|
| 139 |
+
|
| 140 |
+
```shell
|
| 141 |
+
pip install -U "huggingface_hub[cli]"
|
| 142 |
+
```
|
| 143 |
+
|
| 144 |
+
Then, downoad the individual model file the a local directory
|
| 145 |
+
|
| 146 |
+
```shell
|
| 147 |
+
huggingface-cli download tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF --include "LongWriter-Zero-32B-Q2_K.gguf" --local-dir MY_LOCAL_DIR
|
| 148 |
+
```
|
| 149 |
+
|
| 150 |
+
If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
|
| 151 |
+
|
| 152 |
+
```shell
|
| 153 |
+
huggingface-cli download tensorblock/THU-KEG_LongWriter-Zero-32B-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
|
| 154 |
+
```
|