update axcl aarch64 bin
Browse files- .gitattributes +1 -0
- main +3 -0
- run_smollm2_360m_axcl_aarch64.sh +3 -6
- smollm2_tokenizer_uid.py +2 -2
.gitattributes
CHANGED
|
@@ -38,3 +38,4 @@ main_axcl_aarch64 filter=lfs diff=lfs merge=lfs -text
|
|
| 38 |
main_axcl_x86 filter=lfs diff=lfs merge=lfs -text
|
| 39 |
main_prefill filter=lfs diff=lfs merge=lfs -text
|
| 40 |
main_ax650 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 38 |
main_axcl_x86 filter=lfs diff=lfs merge=lfs -text
|
| 39 |
main_prefill filter=lfs diff=lfs merge=lfs -text
|
| 40 |
main_ax650 filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
main filter=lfs diff=lfs merge=lfs -text
|
main
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9396632c6f0ad5dde9f93a5e4ac57cd3cbb3d2fd1b3748cfdd188bbbce0bfe62
|
| 3 |
+
size 1787328
|
run_smollm2_360m_axcl_aarch64.sh
CHANGED
|
@@ -1,14 +1,11 @@
|
|
| 1 |
-
./
|
| 2 |
--template_filename_axmodel "smollm2-360m-ax650/llama_p128_l%d_together.axmodel" \
|
| 3 |
--axmodel_num 32 \
|
| 4 |
-
--
|
| 5 |
-
--filename_tokenizer_model "http://127.0.0.1:12345" \
|
| 6 |
-
--bos 0 --eos 0 \
|
| 7 |
--filename_post_axmodel "smollm2-360m-ax650/llama_post.axmodel" \
|
| 8 |
--filename_tokens_embed "smollm2-360m-ax650/model.embed_tokens.weight.bfloat16.bin" \
|
| 9 |
--tokens_embed_num 49152 \
|
| 10 |
--tokens_embed_size 960 \
|
| 11 |
--use_mmap_load_embed 0 \
|
| 12 |
--live_print 1 \
|
| 13 |
-
--
|
| 14 |
-
--prompt "$1"
|
|
|
|
| 1 |
+
./main \
|
| 2 |
--template_filename_axmodel "smollm2-360m-ax650/llama_p128_l%d_together.axmodel" \
|
| 3 |
--axmodel_num 32 \
|
| 4 |
+
--url_tokenizer_model "http://127.0.0.1:12345" \
|
|
|
|
|
|
|
| 5 |
--filename_post_axmodel "smollm2-360m-ax650/llama_post.axmodel" \
|
| 6 |
--filename_tokens_embed "smollm2-360m-ax650/model.embed_tokens.weight.bfloat16.bin" \
|
| 7 |
--tokens_embed_num 49152 \
|
| 8 |
--tokens_embed_size 960 \
|
| 9 |
--use_mmap_load_embed 0 \
|
| 10 |
--live_print 1 \
|
| 11 |
+
--devices 0
|
|
|
smollm2_tokenizer_uid.py
CHANGED
|
@@ -11,7 +11,7 @@ class Tokenizer_Http():
|
|
| 11 |
def __init__(self):
|
| 12 |
model_id = "smollm2_tokenizer"
|
| 13 |
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 14 |
-
self.messages = [{"role": "system", "content": "
|
| 15 |
self.token_ids = []
|
| 16 |
|
| 17 |
def encode(self, prompt, last_reply=None):
|
|
@@ -58,7 +58,7 @@ class Tokenizer_Http():
|
|
| 58 |
def eos_token(self):
|
| 59 |
return self.tokenizer.eos_token
|
| 60 |
|
| 61 |
-
def reset(self, system_prompt="
|
| 62 |
self.messages = [
|
| 63 |
{"role": "system", "content": system_prompt},
|
| 64 |
]
|
|
|
|
| 11 |
def __init__(self):
|
| 12 |
model_id = "smollm2_tokenizer"
|
| 13 |
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 14 |
+
self.messages = [{"role": "system", "content": "You are a AI assistant, created by HuggingfaceTB"}]
|
| 15 |
self.token_ids = []
|
| 16 |
|
| 17 |
def encode(self, prompt, last_reply=None):
|
|
|
|
| 58 |
def eos_token(self):
|
| 59 |
return self.tokenizer.eos_token
|
| 60 |
|
| 61 |
+
def reset(self, system_prompt="You are a AI assistant, created by HuggingfaceTB"):
|
| 62 |
self.messages = [
|
| 63 |
{"role": "system", "content": system_prompt},
|
| 64 |
]
|