prince-canuma commited on
Commit
9a6b50f
·
verified ·
1 Parent(s): c049a3b

Upload folder using huggingface_hub (#1)

Browse files

- 67d3b889ec1cb557b2972d1c0075fd1602a7b75dba2ba90f5108b2d714639886 (61c0e14c646382c1f918852960a805f2dedc997c)
- edfec6ed588944c71450f7b524a0768f75e70be6216da936a7500a43d0509af7 (b47189566b9500c098823cb2eb9af593d4f8c489)
- f134b89daea17165299704d1f44b6fec5d7c3a93fcc163c72a402acdae915afa (123cb60b65a3c663bd858d7f3e103473efbe83ce)
- fc85e5fd4808c432af7bd75ae8e72fc7c97b024a5d22a2e932ed7bf7be0d8b7f (c56638fc80cbec80087f70ae874e1e38b9a84931)
- b490dcd25219f9899dac5b26d4f124b13e7abafa0e83f5126a8e742ba08333f8 (fb5a25955c7c696182e5a0a8467c7f22b3f572a7)
- 5a57103480713e8a889e4ea485ffdb1cf7ac5c2845add3077296bbde5e4c0709 (cc862072af59d0573b122df170c3d17ef28a3e71)
- 1ec52663ba922c6eb8602b914219a67d02a9099d7ff52f01528f0d4f80e5e0f5 (22d41602d2ebb9a7943eefb6f0770c401d0780a1)
- b69cf8822739a60f961499bf52d3ec4d9bb6ed8cba9cc27d324942ff71eef3ad (38e9b1ba8184b978b016048fe7109182779eea92)
- 182970dd5abe4fe350cb750846cf43694dad335a8306136c4b69820569b7a18b (e813a3771c2a33d1b2c832bdce8e653d4122ac7a)
- 74b15acf354c2ed5ce75dce9de1d085955c115639d00134e1de908bf4238fa60 (f2c22004798f40f428e52ed4fe2a49e548c953e6)
- 8c9cb713310efd6d5cf5af31f8e02085fbb38df310e0726bb701aa81899a6b5d (5d78aca8ea7c04fe8f2599beb9cc446b5640af93)
- 1a400eaa4b3cf8946d54b58345586c43d84d432ce999feac84f0fa639436902c (1886640b2a4e31014ecf148d241764ff29a389ee)
- 320f7eb9626de200a1f0380f767685f714b93d4f4fbf86d69ccd304d1843cf90 (e524825f013686796c1144a497490d77677b334d)
- 8fbc545c0fcf28eac2e40a631ada1ac59ef5bdbf74199a1f9f8e90a2da6c40be (54b48d26f1a4fe583dac6f88496241472ae54610)
- 4ea2af8def5b3450e1d1968aaa6dc74ebd544affc5649a8ea6d837029568481d (f43bce9b12550fd5044b84fac961e70da14f6a57)
- b8d613e988a63e42a08e64e83e9c058b219ab9816d6d9aa642e36a2985537c27 (e12dd1e708a5b34f1438babcce8864eddcafb1fc)

README.md ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - chat
7
+ - mlx
8
+ license_name: tongyi-qianwen
9
+ license_link: https://huggingface.co/Qwen/Qwen2-Math-72B-Instruct/blob/main/LICENSE
10
+ pipeline_tag: text-generation
11
+ ---
12
+
13
+ # mlx-community/Qwen2-Math-72B-Instruct-8bit
14
+
15
+ The Model [mlx-community/Qwen2-Math-72B-Instruct-8bit](https://huggingface.co/mlx-community/Qwen2-Math-72B-Instruct-8bit) was converted to MLX format from [Qwen/Qwen2-Math-72B-Instruct](https://huggingface.co/Qwen/Qwen2-Math-72B-Instruct) using mlx-lm version **0.16.1**.
16
+
17
+ ## Use with mlx
18
+
19
+ ```bash
20
+ pip install mlx-lm
21
+ ```
22
+
23
+ ```python
24
+ from mlx_lm import load, generate
25
+
26
+ model, tokenizer = load("mlx-community/Qwen2-Math-72B-Instruct-8bit")
27
+ response = generate(model, tokenizer, prompt="hello", verbose=True)
28
+ ```
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 8192,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 29568,
12
+ "max_position_embeddings": 4096,
13
+ "max_window_layers": 70,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 64,
16
+ "num_hidden_layers": 80,
17
+ "num_key_value_heads": 8,
18
+ "quantization": {
19
+ "group_size": 64,
20
+ "bits": 8
21
+ },
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_theta": 10000.0,
24
+ "sliding_window": 32768,
25
+ "tie_word_embeddings": false,
26
+ "torch_dtype": "bfloat16",
27
+ "transformers_version": "4.43.1",
28
+ "use_cache": true,
29
+ "use_mrope": false,
30
+ "use_sliding_window": false,
31
+ "vocab_size": 152064
32
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b233c262df637f24b4b1ca5fc90382d47a919df9172ed6ffd28da6e8dc79cd4
3
+ size 5214291418
model-00002-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93971bd6d4f335e798dac79fe2a4432791faddcb6bf74b5ee19a68399b46c84d
3
+ size 5177560124
model-00003-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9bcf4d81280774361320bcfc834c402bf0ded2c778ace461dd6973299231e75
3
+ size 5338047575
model-00004-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9e20ff6ea021cd2b4e81492f27580edb9f2188a0dcd7949d00e0b8a3ea28748
3
+ size 5338047581
model-00005-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d78016b620d67dfbd13dd9d039a6cacef1620676d57368751786444df383509
3
+ size 5177560270
model-00006-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:006642406c4cba51a902649069da4ce746e56eb02520494bc5e92b75742c4dd1
3
+ size 5338047562
model-00007-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:234529ca9ea410777f50af20aefa6171ed824fe46518da9fb4d35835e227ea01
3
+ size 5338047557
model-00008-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:470fc497fbe3fa8e3492ddae721d07be71873b5d7559b75fb7f4331b75782454
3
+ size 5177560278
model-00009-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84264319b81c5c8ba5bf12675e9d6a695f4fb54f69d4fe0f819482c61ced9637
3
+ size 5338047594
model-00010-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9860f1e0f5bf86ef8219580afd8e986ae267e158bea3f2c3eed95a3fbaf2b16a
3
+ size 5338047549
model-00011-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9547e2242539c0c94a774b5f57b51f099d1b212dc0532fff57eaca09a008009e
3
+ size 5177560280
model-00012-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef21479e8a3508fa7e9407feeac9b257055cf60e24edb4190902dd538fa215fe
3
+ size 5338047560
model-00013-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b266dd16570d66b074d3448bcd4951932d4de603b382fe2ec29f8a396cdb4d1
3
+ size 5338047535
model-00014-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6057ee1890acbde41f0a98e1dc2831af7cc34895264e65a9b8ad1aa6035843e5
3
+ size 5177560290
model-00015-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d45d088407ee4afb073148499654c3317bc2ca6957970b3bc731bed599c984d
3
+ size 3446110912
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff