broadfield-dev commited on
Commit
7a0556a
·
verified ·
1 Parent(s): f388585

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +2 -5
  2. model.onnx +2 -2
  3. tokenizer_config.json +1 -0
README.md CHANGED
@@ -7,9 +7,6 @@ tags:
7
  - tokenizers
8
  - optimum
9
  - text-generation
10
- - int8
11
- - quantized
12
- - mobile
13
  language: en
14
  pipeline_tag: text-generation
15
  ---
@@ -20,7 +17,7 @@ This is a version of [Qwen/Qwen3-0.6B](https://huggingface.co/Qwen/Qwen3-0.6B) t
20
  - **Base Model:** `Qwen/Qwen3-0.6B`
21
  - **Task:** `text-generation`
22
  - **Opset Version:** `17`
23
- - **Optimization:** `INT8 - Optimized for Mobile (ARM64)`
24
  ## Usage
25
  ### Installation
26
  For a lightweight mobile/serverless setup, you only need `onnxruntime` and `tokenizers`.
@@ -58,4 +55,4 @@ print("Output logits shape:", outputs[0].shape)
58
  ```
59
  ## About this Export
60
  This model was exported using [Optimum](https://huggingface.co/docs/optimum/index).
61
- It includes the `INT8 - Optimized for Mobile (ARM64)` quantization settings and a pre-compiled `tokenizer.json` for fast loading.
 
7
  - tokenizers
8
  - optimum
9
  - text-generation
 
 
 
10
  language: en
11
  pipeline_tag: text-generation
12
  ---
 
17
  - **Base Model:** `Qwen/Qwen3-0.6B`
18
  - **Task:** `text-generation`
19
  - **Opset Version:** `17`
20
+ - **Optimization:** `FP32 (No Quantization)`
21
  ## Usage
22
  ### Installation
23
  For a lightweight mobile/serverless setup, you only need `onnxruntime` and `tokenizers`.
 
55
  ```
56
  ## About this Export
57
  This model was exported using [Optimum](https://huggingface.co/docs/optimum/index).
58
+ It includes the `FP32 (No Quantization)` quantization settings and a pre-compiled `tokenizer.json` for fast loading.
model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11cd9918130937ca393d358ed28b488709eff8acc1715461d3032dd32796c35e
3
- size 754045218
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a00ac8fcba8de4719e7205a4cc1c4b9b3f01f88c0cf3c4fc90e59cdbce8a20
3
+ size 1403685
tokenizer_config.json CHANGED
@@ -231,6 +231,7 @@
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
 
234
  "model_max_length": 131072,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
 
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
+ "fix_mistral_regex": true,
235
  "model_max_length": 131072,
236
  "pad_token": "<|endoftext|>",
237
  "split_special_tokens": false,