AmitMY commited on
Commit
beef6fa
·
verified ·
1 Parent(s): 1557d84

Upload folder using huggingface_hub

Browse files
chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {# -------------------------------Optional: Define available tools------------------------------- #}{%- if tools %}{%- for tool in tools %}{{- "\x11" -}} {# Start Of Tool Definition #}{{- "\x0E" -}} {# Start Of Attention Block #}{{- tool['function']['name'] + '\n' -}}{# Loop through all argument names and their descriptions #}{%- for argument in tool['function']['parameters']['properties'] %}{{- argument + ': ' + tool['function']['parameters']['properties'][argument]['description'] + '\n' -}}{%- endfor %}{{- '\x0F' -}} {# End Of Attention Block #}{%- endfor %}{%- endif %}{# -------------------------------Main conversation message loop------------------------------- #}{% for message in messages %}{{- "\x01" -}} {# Start Of Text Block #}{# Print the role tag, e.g. user or assistant #}{{- message.role + "\n" -}}{# If not the "assistant", we wrap with an attention block #}{% if message.role != "assistant" %}{{- "\x0E" -}}{% endif %}{# If the message contains normal content, print it #}{% if message.content %}{{- message.content -}}{% endif %}{# If the assistant called any tools, print those tool calls #}{% if message.tool_calls %}{% for call in message.tool_calls %}{{- "\x1A" -}} {# Start Of Tool Call #}{{- call.type + " " + call.function.name+ "(" + call.function.arguments | tojson + ")" -}}{{- "\x1B" -}} {# End Of Tool Call #}{% endfor %}{% endif %}{# If not the "assistant", we wrap with an attention block #}{% if message.role != "assistant" %}{{- "\x0F" -}}{% endif %}{{- "\x17" -}} {# End Of Text Block #}{{- "\n" -}} {# Newline after each message #}{% endfor %}{# -------------------------------Add a final assistant prompt marker to continue generation------------------------------- #}{%- if add_generation_prompt %}{{- "\x01assistant\n" -}}{%- endif %}
config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CharacterCausalLMWrapper"
4
+ ],
5
+ "base_model_name_or_path": "sign/utf8-lm-tiny",
6
+ "bos_token_id": 2,
7
+ "dtype": "float32",
8
+ "eos_token_id": 3,
9
+ "model_type": "character_causal_lm",
10
+ "num_bytes": 4,
11
+ "pad_token_id": 0,
12
+ "transformers_version": "4.57.3"
13
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3f434379d279f799d15f3f44f9b5722c4adfe65a0e48d9a8f3a1573c1a9d720
3
+ size 13000480
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9c02aa1d0972610f5fe2f3845c15451b08bc0d41ca4e8bfd06eae04daf61f04
3
+ size 24962187
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c19bab1174704a4a4441475683bf1270277af15d2e2c95e964789128e482c4
3
+ size 14645
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0b5e80164025bc0b1fea364c5130f846e3294d9ff1d2acdb4f8586c6891d0d5
3
+ size 1465
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "\u0002",
3
+ "eos_token": "\u0003",
4
+ "pad_token": "\u0000"
5
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "\u0000",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "2": {
12
+ "content": "\u0002",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "3": {
20
+ "content": "\u0003",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "\u0002",
29
+ "bos_token_id": 2,
30
+ "clean_up_tokenization_spaces": false,
31
+ "eos_token": "\u0003",
32
+ "eos_token_id": 3,
33
+ "extra_special_tokens": {},
34
+ "model_max_length": 1000000000000000019884624838656,
35
+ "pad_token": "\u0000",
36
+ "pad_token_id": 0,
37
+ "tokenizer_class": "UTF32Tokenizer"
38
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147a1faad7736ad7aa880bcde2ba16d2fbc4b889914051c5b14c52a166fe0dbe
3
+ size 5841