File size: 1,224 Bytes
5f923cd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
NB: For significant changes, please remake the testdata as follows (from the schema/ directory)

Remake test_tflite_tokenizer.litertlm :
bazel run -c opt //schema:litertlm_export_main -- \
 --tokenizer_file=$PWD/schema/testdata//gemma3_tokenizer.spiece  \
 --tflite_file=$PWD/schema/testdata/attention.tflite \
 --output_path=$PWD/schema/testdata/test_tokenizer_tflite.litertlm  \
 --section_metadata="tokenizer:vocab_size=10000,algorithm=bpe;tflite:quantized=true,model_size=1234567"

Remake test_tok_tfl_llm.litertlm
bazel run -c opt //schema:litertlm_export_main -- \
 --tokenizer_file=$PWD/schema/testdata/gemma3_tokenizer.spiece \
 --tflite_file=$PWD/schema/testdata/attention.tflite \
 --llm_metadata=$PWD/schema/testdata/llm_metadata.pb \
 --binary_data=$PWD/schema/testdata/data.bin \
 --output_path=$PWD/schema/testdata/test_tok_tfl_llm.litertlm \
 --section_metadata="tokenizer:vocab_size=10000,algorithm=bpe;tflite:quantized=true,model_size=1234567;llm_metadata:model=gemma3;binary_data:type=abc"

Remake test_hf_tokenizer.litertlm :
bazel run -c opt //schema:litertlm_export_main -- \
 --hf_tokenizer_json_file=$PWD/schema/testdata/tokenizer.json \
 --output_path=$PWD/schema/testdata/test_hf_tokenizer.litertlm