wli1995 commited on
Commit
108f6d0
·
verified ·
1 Parent(s): 0bc522d

update axcl aarch64 bin

Browse files
.gitattributes CHANGED
@@ -38,3 +38,4 @@ main_axcl_aarch64 filter=lfs diff=lfs merge=lfs -text
38
  main_axcl_x86 filter=lfs diff=lfs merge=lfs -text
39
  main_prefill filter=lfs diff=lfs merge=lfs -text
40
  main_ax650 filter=lfs diff=lfs merge=lfs -text
 
 
38
  main_axcl_x86 filter=lfs diff=lfs merge=lfs -text
39
  main_prefill filter=lfs diff=lfs merge=lfs -text
40
  main_ax650 filter=lfs diff=lfs merge=lfs -text
41
+ main filter=lfs diff=lfs merge=lfs -text
main ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9396632c6f0ad5dde9f93a5e4ac57cd3cbb3d2fd1b3748cfdd188bbbce0bfe62
3
+ size 1787328
run_smollm2_360m_axcl_aarch64.sh CHANGED
@@ -1,14 +1,11 @@
1
- ./main_axcl_aarch64 \
2
  --template_filename_axmodel "smollm2-360m-ax650/llama_p128_l%d_together.axmodel" \
3
  --axmodel_num 32 \
4
- --tokenizer_type 2 \
5
- --filename_tokenizer_model "http://127.0.0.1:12345" \
6
- --bos 0 --eos 0 \
7
  --filename_post_axmodel "smollm2-360m-ax650/llama_post.axmodel" \
8
  --filename_tokens_embed "smollm2-360m-ax650/model.embed_tokens.weight.bfloat16.bin" \
9
  --tokens_embed_num 49152 \
10
  --tokens_embed_size 960 \
11
  --use_mmap_load_embed 0 \
12
  --live_print 1 \
13
- --continue 1 \
14
- --prompt "$1"
 
1
+ ./main \
2
  --template_filename_axmodel "smollm2-360m-ax650/llama_p128_l%d_together.axmodel" \
3
  --axmodel_num 32 \
4
+ --url_tokenizer_model "http://127.0.0.1:12345" \
 
 
5
  --filename_post_axmodel "smollm2-360m-ax650/llama_post.axmodel" \
6
  --filename_tokens_embed "smollm2-360m-ax650/model.embed_tokens.weight.bfloat16.bin" \
7
  --tokens_embed_num 49152 \
8
  --tokens_embed_size 960 \
9
  --use_mmap_load_embed 0 \
10
  --live_print 1 \
11
+ --devices 0
 
smollm2_tokenizer_uid.py CHANGED
@@ -11,7 +11,7 @@ class Tokenizer_Http():
11
  def __init__(self):
12
  model_id = "smollm2_tokenizer"
13
  self.tokenizer = AutoTokenizer.from_pretrained(model_id)
14
- self.messages = [{"role": "system", "content": "I are a AI assistant, created by HuggingfaceTB"}]
15
  self.token_ids = []
16
 
17
  def encode(self, prompt, last_reply=None):
@@ -58,7 +58,7 @@ class Tokenizer_Http():
58
  def eos_token(self):
59
  return self.tokenizer.eos_token
60
 
61
- def reset(self, system_prompt="I are a AI assistant, created by HuggingfaceTB"):
62
  self.messages = [
63
  {"role": "system", "content": system_prompt},
64
  ]
 
11
  def __init__(self):
12
  model_id = "smollm2_tokenizer"
13
  self.tokenizer = AutoTokenizer.from_pretrained(model_id)
14
+ self.messages = [{"role": "system", "content": "You are a AI assistant, created by HuggingfaceTB"}]
15
  self.token_ids = []
16
 
17
  def encode(self, prompt, last_reply=None):
 
58
  def eos_token(self):
59
  return self.tokenizer.eos_token
60
 
61
+ def reset(self, system_prompt="You are a AI assistant, created by HuggingfaceTB"):
62
  self.messages = [
63
  {"role": "system", "content": system_prompt},
64
  ]