Update README.md
Browse files
README.md
CHANGED
|
@@ -26,25 +26,6 @@ We release InternViT-6B-448px-V1-0, which is integrated into [InternVL-Chat-V1-1
|
|
| 26 |
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR-related datasets.
|
| 27 |
- **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for MLLM. Therefore, when building a MLLM with this model, **please use the features from the fourth-to-last layer.**
|
| 28 |
|
| 29 |
-
## Released Models
|
| 30 |
-
### Vision Foundation model
|
| 31 |
-
| Model | Date | Download | Note |
|
| 32 |
-
| ----------------------- | ---------- | ---------------------------------------------------------------------- | -------------------------------- |
|
| 33 |
-
| InternViT-6B-448px-V1-5 | 2024.04.20 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-5) | support dynamic resolution, super strong OCR (🔥new) |
|
| 34 |
-
| InternViT-6B-448px-V1-2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-2) | 448 resolution |
|
| 35 |
-
| InternViT-6B-448px-V1-0 | 2024.01.30 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-0) | 448 resolution |
|
| 36 |
-
| InternViT-6B-224px | 2023.12.22 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-224px) | vision foundation model |
|
| 37 |
-
| InternVL-14B-224px | 2023.12.22 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-14B-224px) | vision-language foundation model |
|
| 38 |
-
|
| 39 |
-
### Multimodal Large Language Model (MLLM)
|
| 40 |
-
| Model | Date | Download | Note |
|
| 41 |
-
| ----------------------- | ---------- | --------------------------------------------------------------------------- | ---------------------------------- |
|
| 42 |
-
| InternVL-Chat-V1-5 | 2024.04.18 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5) | support 4K image; super strong OCR; Approaching the performance of GPT-4V and Gemini Pro on various benchmarks like MMMU, DocVQA, ChartQA, MathVista, etc. (🔥new)|
|
| 43 |
-
| InternVL-Chat-V1-2-Plus | 2024.02.21 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2-Plus) | more SFT data and stronger |
|
| 44 |
-
| InternVL-Chat-V1-2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2) | scaling up LLM to 34B |
|
| 45 |
-
| InternVL-Chat-V1-1 | 2024.01.24 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-1) | support Chinese and stronger OCR |
|
| 46 |
-
|
| 47 |
-
|
| 48 |
## Model Usage (Image Embeddings)
|
| 49 |
|
| 50 |
```python
|
|
|
|
| 26 |
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR-related datasets.
|
| 27 |
- **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for MLLM. Therefore, when building a MLLM with this model, **please use the features from the fourth-to-last layer.**
|
| 28 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
## Model Usage (Image Embeddings)
|
| 30 |
|
| 31 |
```python
|