MapleJc commited on
Commit
783c540
·
verified ·
1 Parent(s): 4ad62dd

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. onnx/model_quantized.onnx +3 -0
  2. onnx/ort_config.json +27 -31
onnx/model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee0de17cffc413323b61d8a8e602af1cbb37b77e6bf517d568c2054cc89f8b53
3
+ size 11414361
onnx/ort_config.json CHANGED
@@ -1,37 +1,33 @@
1
  {
2
  "one_external_file": true,
3
  "opset": null,
4
- "optimization": {
5
- "disable_attention": null,
6
- "disable_attention_fusion": false,
7
- "disable_bias_gelu": null,
8
- "disable_bias_gelu_fusion": false,
9
- "disable_bias_skip_layer_norm": null,
10
- "disable_bias_skip_layer_norm_fusion": false,
11
- "disable_embed_layer_norm": true,
12
- "disable_embed_layer_norm_fusion": true,
13
- "disable_gelu": null,
14
- "disable_gelu_fusion": false,
15
- "disable_group_norm_fusion": true,
16
- "disable_layer_norm": null,
17
- "disable_layer_norm_fusion": false,
18
- "disable_packed_kv": true,
19
- "disable_rotary_embeddings": false,
20
- "disable_shape_inference": true,
21
- "disable_skip_layer_norm": null,
22
- "disable_skip_layer_norm_fusion": false,
23
- "enable_gelu_approximation": false,
24
- "enable_gemm_fast_gelu_fusion": false,
25
- "enable_transformers_specific_optimizations": true,
26
- "fp16": false,
27
- "no_attention_mask": false,
28
- "optimization_level": 2,
29
- "optimize_for_gpu": false,
30
- "optimize_with_onnxruntime_only": null,
31
- "use_mask_index": false,
32
- "use_multi_head_attention": false,
33
- "use_raw_attention_mask": false
34
  },
35
- "quantization": {},
36
  "use_external_data_format": false
37
  }
 
1
  {
2
  "one_external_file": true,
3
  "opset": null,
4
+ "optimization": {},
5
+ "quantization": {
6
+ "activations_dtype": "QUInt8",
7
+ "activations_symmetric": false,
8
+ "format": "QOperator",
9
+ "is_static": false,
10
+ "mode": "IntegerOps",
11
+ "nodes_to_exclude": [],
12
+ "nodes_to_quantize": [],
13
+ "operators_to_quantize": [
14
+ "Conv",
15
+ "MatMul",
16
+ "Attention",
17
+ "LSTM",
18
+ "Gather",
19
+ "Transpose",
20
+ "EmbedLayerNormalization"
21
+ ],
22
+ "per_channel": false,
23
+ "qdq_add_pair_to_weight": false,
24
+ "qdq_dedicated_pair": false,
25
+ "qdq_op_type_per_channel_support_to_axis": {
26
+ "MatMul": 1
27
+ },
28
+ "reduce_range": false,
29
+ "weights_dtype": "QInt8",
30
+ "weights_symmetric": true
 
 
 
31
  },
 
32
  "use_external_data_format": false
33
  }