Add files using upload-large-folder tool
Browse files- README.md +35 -0
- chat_template.jinja +1 -0
- config.json +36 -0
- model-00001-of-00013.safetensors +3 -0
- model-00002-of-00013.safetensors +3 -0
- model-00003-of-00013.safetensors +3 -0
- model-00004-of-00013.safetensors +3 -0
- model-00005-of-00013.safetensors +3 -0
- model-00006-of-00013.safetensors +3 -0
- model-00007-of-00013.safetensors +3 -0
- model-00008-of-00013.safetensors +3 -0
- model-00009-of-00013.safetensors +3 -0
- model-00010-of-00013.safetensors +3 -0
- model-00011-of-00013.safetensors +3 -0
- model-00012-of-00013.safetensors +3 -0
- model-00013-of-00013.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +23 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +0 -0
README.md
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: TheDrummer/Behemoth-R1-123B-v2
|
| 3 |
+
library_name: mlx
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- mlx
|
| 7 |
+
---
|
| 8 |
+
|
| 9 |
+
# beezu/Behemoth-R1-123B-v2-MLX-MXFP4
|
| 10 |
+
|
| 11 |
+
This model [beezu/Behemoth-R1-123B-v2-MLX-MXFP4](https://huggingface.co/beezu/Behemoth-R1-123B-v2-MLX-MXFP4) was
|
| 12 |
+
converted to MLX format from [TheDrummer/Behemoth-R1-123B-v2](https://huggingface.co/TheDrummer/Behemoth-R1-123B-v2)
|
| 13 |
+
using mlx-lm version **0.28.3**.
|
| 14 |
+
|
| 15 |
+
## Use with mlx
|
| 16 |
+
|
| 17 |
+
```bash
|
| 18 |
+
pip install mlx-lm
|
| 19 |
+
```
|
| 20 |
+
|
| 21 |
+
```python
|
| 22 |
+
from mlx_lm import load, generate
|
| 23 |
+
|
| 24 |
+
model, tokenizer = load("beezu/Behemoth-R1-123B-v2-MLX-MXFP4")
|
| 25 |
+
|
| 26 |
+
prompt = "hello"
|
| 27 |
+
|
| 28 |
+
if tokenizer.chat_template is not None:
|
| 29 |
+
messages = [{"role": "user", "content": prompt}]
|
| 30 |
+
prompt = tokenizer.apply_chat_template(
|
| 31 |
+
messages, add_generation_prompt=True
|
| 32 |
+
)
|
| 33 |
+
|
| 34 |
+
response = generate(model, tokenizer, prompt=prompt, verbose=True)
|
| 35 |
+
```
|
chat_template.jinja
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + '[/INST]' }}{% elif message['role'] == 'system' %}{{ '[SYSTEM_PROMPT] ' + message['content'] + '[/SYSTEM_PROMPT]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + eos_token }}{% else %}{{ raise_exception('Only user, system and assistant roles are supported!') }}{% endif %}{% endfor %}
|
config.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"MistralForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_dropout": 0.0,
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"head_dim": 128,
|
| 9 |
+
"hidden_act": "silu",
|
| 10 |
+
"hidden_size": 12288,
|
| 11 |
+
"initializer_range": 0.02,
|
| 12 |
+
"intermediate_size": 28672,
|
| 13 |
+
"max_position_embeddings": 131072,
|
| 14 |
+
"model_type": "mistral",
|
| 15 |
+
"num_attention_heads": 96,
|
| 16 |
+
"num_hidden_layers": 88,
|
| 17 |
+
"num_key_value_heads": 8,
|
| 18 |
+
"quantization": {
|
| 19 |
+
"group_size": 32,
|
| 20 |
+
"bits": 4,
|
| 21 |
+
"mode": "mxfp4"
|
| 22 |
+
},
|
| 23 |
+
"quantization_config": {
|
| 24 |
+
"group_size": 32,
|
| 25 |
+
"bits": 4,
|
| 26 |
+
"mode": "mxfp4"
|
| 27 |
+
},
|
| 28 |
+
"rms_norm_eps": 1e-05,
|
| 29 |
+
"rope_theta": 1000000.0,
|
| 30 |
+
"sliding_window": null,
|
| 31 |
+
"tie_word_embeddings": false,
|
| 32 |
+
"torch_dtype": "bfloat16",
|
| 33 |
+
"transformers_version": "4.55.2",
|
| 34 |
+
"use_cache": true,
|
| 35 |
+
"vocab_size": 32768
|
| 36 |
+
}
|
model-00001-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3ebbf6aa5ce27efdd81231f84e067795b28968cb9d7963e25de6520729657b1e
|
| 3 |
+
size 5361464013
|
model-00002-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7565382a2999b52d9982e268607045d1e419b1f381b68ced751c6f6b590d84f
|
| 3 |
+
size 5321356775
|
model-00003-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5bdea4271705cb9234a2c27b12c4b73fbb53f4873fa915b6c79a9d85922b98c5
|
| 3 |
+
size 5334725441
|
model-00004-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:20c1894eead71916c361bad394dd391d0321fb40ddea9993caae56c1ba62ae44
|
| 3 |
+
size 5334725473
|
model-00005-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eab3fddfd586a1befc2ac79c3dd3da6b432df31d62c73d5d3fd2faae8937704a
|
| 3 |
+
size 5334774832
|
model-00006-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5db6d8e456d9d97a4d917a86b40579c2e17ae5e798260d9f6457de5d21669227
|
| 3 |
+
size 5321356801
|
model-00007-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a448eb22b1e9c7a88216468c351246c907a48445e74661251a53bb0f0ea58a56
|
| 3 |
+
size 5334725461
|
model-00008-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e2042034bcb5b096ecf93fe4d6576152fd2b7587d83ddbe3b7d449bf4e7537e2
|
| 3 |
+
size 5334725479
|
model-00009-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:43e4203f648b809f51a3f38b9925585378f214119db549ae1612b8eb2383db0f
|
| 3 |
+
size 5334774850
|
model-00010-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9640b2be73ebfd2495522fc9362ba72cfa88c34dacd36e98f7a92e79320f05bd
|
| 3 |
+
size 5321356805
|
model-00011-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4d65aa494450ef153fd6624047d508f92c34fe33382a2e9ce4096cffc7b8ef0
|
| 3 |
+
size 5334725469
|
model-00012-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:44bd17a624ba0eeae8302a3dbc888f38e21189d980a1dc356b57400b8d900eb1
|
| 3 |
+
size 5334725467
|
model-00013-of-00013.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:de05d97a2d400548d4c095141a669b18983a2b32c9f083501fbbac198798613e
|
| 3 |
+
size 1136519690
|
model.safetensors.index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"unk_token": {
|
| 17 |
+
"content": "<unk>",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
}
|
| 23 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b968b8dc352f42192367337c78ccc61e1eaddc6d641a579372d4f20694beb7a
|
| 3 |
+
size 587562
|
tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|