czczup commited on
Commit
1165597
·
verified ·
1 Parent(s): f92cfa6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -10,16 +10,19 @@ datasets:
10
  pipeline_tag: visual-question-answering
11
  ---
12
 
13
- # Model Card for InternVL-Chat-Chinese-V1.1
14
-
15
- <img src="https://cdn-uploads.huggingface.co/production/uploads/64119264f0f81eb569e0d569/-N5Kz3SQM2KOxN0m70ecj.webp" alt="Image Description" width="300" height="300">
16
 
 
17
 
18
- ## What is InternVL?
19
 
20
- \[[Paper](https://arxiv.org/abs/2312.14238)\] \[[GitHub](https://github.com/OpenGVLab/InternVL)\] \[[Chat Demo](https://internvl.opengvlab.com/)\]
21
 
22
- InternVL scales up the ViT to _**6B parameters**_ and aligns it with LLM.
 
 
 
 
 
23
 
24
  ## Model Details
25
  - **Model Type:** multimodal large language model (MLLM)
@@ -40,13 +43,10 @@ InternVL scales up the ViT to _**6B parameters**_ and aligns it with LLM.
40
 
41
  ## Model Usage
42
 
43
- We provide a minimum code example to run InternVL-Chat using only the `transformers` library.
44
 
45
  You also can use our [online demo](https://internvl.opengvlab.com/) for a quick experience of this model.
46
 
47
- Note: If you meet this error `ImportError: This modeling file requires the following packages that were not found in your environment: fastchat`, please run `pip install fschat`.
48
-
49
-
50
  ```python
51
  import torch
52
  from PIL import Image
 
10
  pipeline_tag: visual-question-answering
11
  ---
12
 
 
 
 
13
 
14
+ # Model Card for InternVL-Chat-V1.1
15
 
16
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/64119264f0f81eb569e0d569/-N5Kz3SQM2KOxN0m70ecj.webp" alt="Image Description" width="300" height="300">
17
 
18
+ \[[Paper](https://arxiv.org/abs/2312.14238)\] \[[GitHub](https://github.com/OpenGVLab/InternVL)\] \[[Chat Demo](https://internvl.opengvlab.com/)\] \[[中文解读](https://zhuanlan.zhihu.com/p/675877376)]
19
 
20
+ | Model | Date | Download | Note |
21
+ | ----------------------- | ---------- | ------------------------------------------------------------------------------------ | ---------------------------------- |
22
+ | InternVL-Chat-V1.5 | 2024.04.18 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5) | support 4K image; super strong OCR; Approaching the performance of GPT-4V and Gemini Pro on various benchmarks like MMMU, DocVQA, ChartQA, MathVista, etc. (🔥new)|
23
+ | InternVL-Chat-V1.2-Plus | 2024.02.21 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-Chinese-V1-2-Plus) | more SFT data and stronger |
24
+ | InternVL-Chat-V1.2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-Chinese-V1-2) | scaling up LLM to 34B |
25
+ | InternVL-Chat-V1.1 | 2024.01.24 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-Chinese-V1-1) | support Chinese and stronger OCR |
26
 
27
  ## Model Details
28
  - **Model Type:** multimodal large language model (MLLM)
 
43
 
44
  ## Model Usage
45
 
46
+ We provide an example code to run InternVL-Chat-V1.1 using only the `transformers` library.
47
 
48
  You also can use our [online demo](https://internvl.opengvlab.com/) for a quick experience of this model.
49
 
 
 
 
50
  ```python
51
  import torch
52
  from PIL import Image