JonathanMiddleton commited on
Commit
8f38c1d
·
verified ·
1 Parent(s): 4d3e020

Upload folder using huggingface_hub

Browse files
chat_template.jinja ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% for message in messages -%}
2
+ {% if message['role'] == 'assistant' -%}
3
+ <|im_start|>assistant
4
+ {% generation %}{{ message['content'] }}{% endgeneration %}<|im_end|>
5
+ {% else -%}
6
+ <|im_start|>{{ message['role'] }}
7
+ {{ message['content'] }}<|im_end|>
8
+ {% endif -%}
9
+ {% endfor -%}
10
+ {% if add_generation_prompt -%}
11
+ <|im_start|>assistant
12
+ {% generation %}{% endgeneration %}
13
+ {% endif -%}
checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452db7d3ba44a071c1dc9c4ca4df773b557562ed96e61ee7ce8ae25bf28eb7a8
3
+ size 3803045897
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DaisyForCausalLM"
4
+ ],
5
+ "attn_all_layers": true,
6
+ "attn_impl": "standard",
7
+ "bos_token_id": 50256,
8
+ "dtype": "float32",
9
+ "eos_token_id": 50256,
10
+ "head_dim": 128,
11
+ "kd_alpha": null,
12
+ "kd_eps": null,
13
+ "kd_temperature": null,
14
+ "max_position_embeddings": 131072,
15
+ "model_dim": 1280,
16
+ "model_type": "daisy",
17
+ "num_heads": 10,
18
+ "num_layers": 20,
19
+ "padded_embeddings": false,
20
+ "skip_mix_mode": "linear",
21
+ "tokenizer_name": "jonathanmiddleton/daisy",
22
+ "transformers_version": "5.0.0rc1",
23
+ "use_tied_embeddings": false,
24
+ "use_value_embeddings": true,
25
+ "vocab_size": 50259,
26
+ "window_size": 2048
27
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb0bf5f76261a54e37881af985d4bee0618d20d369bc684ad9787653a344f5a3
3
+ size 1687085816
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "additional_special_tokens": [
4
+ "<|im_start|>",
5
+ "<|im_end|>"
6
+ ],
7
+ "backend": "tokenizers",
8
+ "bos_token": "<|endoftext|>",
9
+ "eos_token": "<|endoftext|>",
10
+ "errors": "replace",
11
+ "extra_special_tokens": [
12
+ "<|im_start|>",
13
+ "<|im_end|>"
14
+ ],
15
+ "is_local": false,
16
+ "model_max_length": 131072,
17
+ "pad_token": "<|endoftext|>",
18
+ "tokenizer_class": "GPT2Tokenizer",
19
+ "unk_token": "<|endoftext|>"
20
+ }