Update README.md
Browse files
README.md
CHANGED
|
@@ -28,7 +28,7 @@ It is _**the largest open-source vision/vision-language foundation model (14B)**
|
|
| 28 |
- Params (M): 5540 (the last 3 blocks are discarded)
|
| 29 |
- Image size: 448 x 448
|
| 30 |
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR data
|
| 31 |
-
- **Note:** InternViT-6B originally had 48 blocks, and we found that using the output after the fourth-to-last block worked best for VLLM. For ease of use and to save GPU memory, we simply discarded the last 3 blocks. Now, the model has only 45 blocks and the number of parameters has been reduced from 5.9B to 5.5B. Therefore, if you want to build a VLLM based on this model, **please use the last layer
|
| 32 |
## Model Usage (Image Embeddings)
|
| 33 |
|
| 34 |
```python
|
|
|
|
| 28 |
- Params (M): 5540 (the last 3 blocks are discarded)
|
| 29 |
- Image size: 448 x 448
|
| 30 |
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR data
|
| 31 |
+
- **Note:** InternViT-6B originally had 48 blocks, and we found that using the output after the fourth-to-last block worked best for VLLM. For ease of use and to save GPU memory, we simply discarded the last 3 blocks. Now, the model has only 45 blocks and the number of parameters has been reduced from 5.9B to 5.5B. Therefore, if you want to build a VLLM based on this model, **please make use of the features from the last layer.**
|
| 32 |
## Model Usage (Image Embeddings)
|
| 33 |
|
| 34 |
```python
|