yongqiang commited on
Commit
d5f9fc5
·
1 Parent(s): b30c13a

Update readme

Browse files
Files changed (1) hide show
  1. README.md +11 -9
README.md CHANGED
@@ -2,10 +2,10 @@
2
  library_name: transformers
3
  license: bsd-3-clause
4
  base_model:
5
- - OpenGVLab/InternVL3_5-1B
6
  tags:
7
  - InternVL3
8
- - InternVL3_5-1B
9
  - Int8
10
  - VLM
11
  pipeline_tag: image-text-to-text
@@ -13,21 +13,23 @@ language:
13
  - en
14
  ---
15
 
16
- # InternVL3_5-1B
17
 
18
- This version of InternVL3_5-1B has been converted to run on the Axera NPU using **w8a16** quantization.
19
 
20
  This model has been optimized with the following LoRA:
21
 
22
- Compatible with Pulsar2 version: 4.1
 
 
23
 
24
  ## Convert tools links:
25
 
26
  For those who are interested in model conversion, you can try to export axmodel through the original repo:
27
 
28
- https://huggingface.co/OpenGVLab/InternVL3_5-1B
29
 
30
- [How to Convert LLM from Huggingface to axmodel](https://github.com/AXERA-TECH/InternVL3_5-1B.axera/tree/main/model_convert)
31
 
32
  [AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/ax-internvl)
33
 
@@ -94,7 +96,7 @@ Image understanding:
94
  Run the following command on the Axera board to start a chat conversation:
95
 
96
  ```sh
97
- $ cd InternVL3_5-1B.axera/python
98
  $ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请计算函数[y=2x^2+2]的导数, 并提供 markdown 格式的推理过程"
99
  ```
100
 
@@ -127,7 +129,7 @@ y' = 4x
127
  Enter the following command to perform the single-image understanding task:
128
 
129
  ```sh
130
- $ cd InternVL3_5-1B.axera/python
131
  $ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请描述这幅图" -i examples/image_0.jpg --vit_model vit-models/internvl_vit_model_1x3x448x448.axmodel
132
  ```
133
 
 
2
  library_name: transformers
3
  license: bsd-3-clause
4
  base_model:
5
+ - OpenGVLab/InternVL3_5-2B
6
  tags:
7
  - InternVL3
8
+ - InternVL3_5-2B
9
  - Int8
10
  - VLM
11
  pipeline_tag: image-text-to-text
 
13
  - en
14
  ---
15
 
16
+ # InternVL3_5-2B
17
 
18
+ This version of InternVL3_5-2B has been converted to run on the Axera NPU using **w8a16** quantization.
19
 
20
  This model has been optimized with the following LoRA:
21
 
22
+ Compatible with Pulsar2 version: 5.1-patch1.
23
+
24
+ Please note that the context of the model is 2k and the maximum prefill length is 1k.
25
 
26
  ## Convert tools links:
27
 
28
  For those who are interested in model conversion, you can try to export axmodel through the original repo:
29
 
30
+ https://huggingface.co/OpenGVLab/InternVL3_5-2B
31
 
32
+ [How to Convert LLM from Huggingface to axmodel](https://github.com/AXERA-TECH/InternVL3_5-2B.axera/tree/main/model_convert)
33
 
34
  [AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/ax-internvl)
35
 
 
96
  Run the following command on the Axera board to start a chat conversation:
97
 
98
  ```sh
99
+ $ cd InternVL3_5-2B.axera/python
100
  $ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请计算函数[y=2x^2+2]的导数, 并提供 markdown 格式的推理过程"
101
  ```
102
 
 
129
  Enter the following command to perform the single-image understanding task:
130
 
131
  ```sh
132
+ $ cd InternVL3_5-2B.axera/python
133
  $ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请描述这幅图" -i examples/image_0.jpg --vit_model vit-models/internvl_vit_model_1x3x448x448.axmodel
134
  ```
135