Spaces:
Running
Running
| NB: For significant changes, please remake the testdata as follows (from the schema/ directory) | |
| Remake test_tflite_tokenizer.litertlm : | |
| bazel run -c opt //schema:litertlm_export_main -- \ | |
| --tokenizer_file=$PWD/schema/testdata//gemma3_tokenizer.spiece \ | |
| --tflite_file=$PWD/schema/testdata/attention.tflite \ | |
| --output_path=$PWD/schema/testdata/test_tokenizer_tflite.litertlm \ | |
| --section_metadata="tokenizer:vocab_size=10000,algorithm=bpe;tflite:quantized=true,model_size=1234567" | |
| Remake test_tok_tfl_llm.litertlm | |
| bazel run -c opt //schema:litertlm_export_main -- \ | |
| --tokenizer_file=$PWD/schema/testdata/gemma3_tokenizer.spiece \ | |
| --tflite_file=$PWD/schema/testdata/attention.tflite \ | |
| --llm_metadata=$PWD/schema/testdata/llm_metadata.pb \ | |
| --binary_data=$PWD/schema/testdata/data.bin \ | |
| --output_path=$PWD/schema/testdata/test_tok_tfl_llm.litertlm \ | |
| --section_metadata="tokenizer:vocab_size=10000,algorithm=bpe;tflite:quantized=true,model_size=1234567;llm_metadata:model=gemma3;binary_data:type=abc" | |
| Remake test_hf_tokenizer.litertlm : | |
| bazel run -c opt //schema:litertlm_export_main -- \ | |
| --hf_tokenizer_json_file=$PWD/schema/testdata/tokenizer.json \ | |
| --output_path=$PWD/schema/testdata/test_hf_tokenizer.litertlm | |