bibproj commited on
Commit
d52a51b
·
verified ·
1 Parent(s): b7f27f9
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,7 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ pipeline_tag: text-generation
4
+ tags:
5
+ - mlx
6
+ library_name: mlx
7
+ ---
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {% for message in messages %}{% if message['role'] == 'assistant' %}{% set role = 'assistant' %}{% else %}{% set role = message['role'] %}{% endif %}<|im_start|>{{ role }}
2
+ {{ message['content'] | trim }}<|im_end|>
3
+ {% endfor %}{% if add_generation_prompt %}{{'<|im_start|>assistant
4
+ '}}{% endif %}
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 4,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 6144,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 16384,
14
+ "max_position_embeddings": 32768,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 48,
18
+ "num_hidden_layers": 54,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "quantization": {
22
+ "group_size": 64,
23
+ "bits": 8,
24
+ "mode": "affine"
25
+ },
26
+ "quantization_config": {
27
+ "group_size": 64,
28
+ "bits": 8,
29
+ "mode": "affine"
30
+ },
31
+ "rms_norm_eps": 1e-05,
32
+ "rope_scaling": null,
33
+ "rope_theta": 1000000,
34
+ "tie_word_embeddings": false,
35
+ "torch_dtype": "bfloat16",
36
+ "transformers_version": "4.51.3",
37
+ "use_cache": false,
38
+ "vocab_size": 128000
39
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 3,
4
+ "do_sample": true,
5
+ "eos_token_id": 4,
6
+ "transformers_version": "4.51.3"
7
+ }
model-00001-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5da5bbce1fb5c0aa9de54c749727352a25a0459cca7a9cf63773dc6d866640f
3
+ size 5287849928
model-00002-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65bef64c37fa9fb9740e900e6bd08a5d777032096ce587c18ac6ec5806227f0d
3
+ size 5281244401
model-00003-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61b0b801e9323ff5f8ae18daafda873529bc29fa1cd1c46481ed906b78e20ef7
3
+ size 5281244390
model-00004-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72eed5b02905cd4b334f778c602bf42a2448e0e3582ea55aa99d09e57a0a2a2d
3
+ size 5348091098
model-00005-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a79e7a8caaccce36997e98c99df951f8f82872337b36c9615c0320349bf95ba7
3
+ size 2854502703
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ced64a98f4d7db5b270328757bbdf4e3051c7a1bb64c5dc7223f1d308738535
3
+ size 15783106
tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "additional_special_tokens": null,
4
+ "backend": "tokenizers",
5
+ "bos_token": "<s>",
6
+ "clean_up_tokenization_spaces": false,
7
+ "eos_token": "<|im_end|>",
8
+ "is_local": true,
9
+ "model_max_length": 1000000000000000019884624838656,
10
+ "model_specific_special_tokens": {},
11
+ "pad_token": "</s>",
12
+ "sp_model_kwargs": {},
13
+ "spaces_between_special_tokens": false,
14
+ "tokenizer_class": "LlamaTokenizer",
15
+ "unk_token": "<unk>",
16
+ "use_default_system_prompt": false
17
+ }