yongqiang
commited on
Commit
·
d5f9fc5
1
Parent(s):
b30c13a
Update readme
Browse files
README.md
CHANGED
|
@@ -2,10 +2,10 @@
|
|
| 2 |
library_name: transformers
|
| 3 |
license: bsd-3-clause
|
| 4 |
base_model:
|
| 5 |
-
- OpenGVLab/InternVL3_5-
|
| 6 |
tags:
|
| 7 |
- InternVL3
|
| 8 |
-
- InternVL3_5-
|
| 9 |
- Int8
|
| 10 |
- VLM
|
| 11 |
pipeline_tag: image-text-to-text
|
|
@@ -13,21 +13,23 @@ language:
|
|
| 13 |
- en
|
| 14 |
---
|
| 15 |
|
| 16 |
-
# InternVL3_5-
|
| 17 |
|
| 18 |
-
This version of InternVL3_5-
|
| 19 |
|
| 20 |
This model has been optimized with the following LoRA:
|
| 21 |
|
| 22 |
-
Compatible with Pulsar2 version:
|
|
|
|
|
|
|
| 23 |
|
| 24 |
## Convert tools links:
|
| 25 |
|
| 26 |
For those who are interested in model conversion, you can try to export axmodel through the original repo:
|
| 27 |
|
| 28 |
-
https://huggingface.co/OpenGVLab/InternVL3_5-
|
| 29 |
|
| 30 |
-
[How to Convert LLM from Huggingface to axmodel](https://github.com/AXERA-TECH/InternVL3_5-
|
| 31 |
|
| 32 |
[AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/ax-internvl)
|
| 33 |
|
|
@@ -94,7 +96,7 @@ Image understanding:
|
|
| 94 |
Run the following command on the Axera board to start a chat conversation:
|
| 95 |
|
| 96 |
```sh
|
| 97 |
-
$ cd InternVL3_5-
|
| 98 |
$ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请计算函数[y=2x^2+2]的导数, 并提供 markdown 格式的推理过程"
|
| 99 |
```
|
| 100 |
|
|
@@ -127,7 +129,7 @@ y' = 4x
|
|
| 127 |
Enter the following command to perform the single-image understanding task:
|
| 128 |
|
| 129 |
```sh
|
| 130 |
-
$ cd InternVL3_5-
|
| 131 |
$ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请描述这幅图" -i examples/image_0.jpg --vit_model vit-models/internvl_vit_model_1x3x448x448.axmodel
|
| 132 |
```
|
| 133 |
|
|
|
|
| 2 |
library_name: transformers
|
| 3 |
license: bsd-3-clause
|
| 4 |
base_model:
|
| 5 |
+
- OpenGVLab/InternVL3_5-2B
|
| 6 |
tags:
|
| 7 |
- InternVL3
|
| 8 |
+
- InternVL3_5-2B
|
| 9 |
- Int8
|
| 10 |
- VLM
|
| 11 |
pipeline_tag: image-text-to-text
|
|
|
|
| 13 |
- en
|
| 14 |
---
|
| 15 |
|
| 16 |
+
# InternVL3_5-2B
|
| 17 |
|
| 18 |
+
This version of InternVL3_5-2B has been converted to run on the Axera NPU using **w8a16** quantization.
|
| 19 |
|
| 20 |
This model has been optimized with the following LoRA:
|
| 21 |
|
| 22 |
+
Compatible with Pulsar2 version: 5.1-patch1.
|
| 23 |
+
|
| 24 |
+
Please note that the context of the model is 2k and the maximum prefill length is 1k.
|
| 25 |
|
| 26 |
## Convert tools links:
|
| 27 |
|
| 28 |
For those who are interested in model conversion, you can try to export axmodel through the original repo:
|
| 29 |
|
| 30 |
+
https://huggingface.co/OpenGVLab/InternVL3_5-2B
|
| 31 |
|
| 32 |
+
[How to Convert LLM from Huggingface to axmodel](https://github.com/AXERA-TECH/InternVL3_5-2B.axera/tree/main/model_convert)
|
| 33 |
|
| 34 |
[AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/ax-internvl)
|
| 35 |
|
|
|
|
| 96 |
Run the following command on the Axera board to start a chat conversation:
|
| 97 |
|
| 98 |
```sh
|
| 99 |
+
$ cd InternVL3_5-2B.axera/python
|
| 100 |
$ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请计算函数[y=2x^2+2]的导数, 并提供 markdown 格式的推理过程"
|
| 101 |
```
|
| 102 |
|
|
|
|
| 129 |
Enter the following command to perform the single-image understanding task:
|
| 130 |
|
| 131 |
```sh
|
| 132 |
+
$ cd InternVL3_5-2B.axera/python
|
| 133 |
$ python3 infer_axmodel.py --hf_model internvl3-5_tokenizer/ --axmodel_path internvl3-5_axmodel/ --question "请描述这幅图" -i examples/image_0.jpg --vit_model vit-models/internvl_vit_model_1x3x448x448.axmodel
|
| 134 |
```
|
| 135 |
|