|
2 | 2 |
|
3 | 3 | LMDeploy abstracts the complex inference process of multi-modal Vision-Language Models (VLM) into an easy-to-use pipeline, similar to the Large Language Model (LLM) inference [pipeline](../llm/pipeline.md). |
4 | 4 |
|
5 | | -Currently, it supports the following models. |
6 | | - |
7 | | -- [Qwen-VL-Chat](https://huggingface.co/Qwen/Qwen-VL-Chat) |
8 | | -- LLaVA series: [v1.5](https://huggingface.co/collections/liuhaotian/llava-15-653aac15d994e992e2677a7e), [v1.6](https://huggingface.co/collections/liuhaotian/llava-16-65b9e40155f60fd046a5ccf2) |
9 | | -- [Yi-VL](https://huggingface.co/01-ai/Yi-VL-6B) |
10 | | -- [DeepSeek-VL](https://huggingface.co/deepseek-ai/deepseek-vl-7b-chat) |
11 | | -- [InternVL](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5) |
12 | | -- [Mono-InternVL](https://huggingface.co/OpenGVLab/Mono-InternVL-2B) |
13 | | -- [MGM](https://huggingface.co/YanweiLi/MGM-7B) |
14 | | -- [XComposer](https://huggingface.co/internlm/internlm-xcomposer2-vl-7b) |
15 | | -- [CogVLM](https://github.com/InternLM/lmdeploy/tree/main/docs/en/multi_modal/cogvlm.md) |
16 | | - |
17 | | -We genuinely invite the community to contribute new VLM support to LMDeploy. Your involvement is truly appreciated. |
| 5 | +The supported models are listed [here](../supported_models/supported_models.md). We genuinely invite the community to contribute new VLM support to LMDeploy. Your involvement is truly appreciated. |
18 | 6 |
|
19 | 7 | This article showcases the VLM pipeline using the [liuhaotian/llava-v1.6-vicuna-7b](https://huggingface.co/liuhaotian/llava-v1.6-vicuna-7b) model as a case study. |
20 | 8 | You'll learn about the simplest ways to leverage the pipeline and how to gradually unlock more advanced features by adjusting engine parameters and generation arguments, such as tensor parallelism, context window sizing, random sampling, and chat template customization. |
21 | 9 | Moreover, we will provide practical inference examples tailored to scenarios with multiple images, batch prompts etc. |
22 | 10 |
|
| 11 | +Using the pipeline interface to infer other VLM models is similar, with the main difference being the configuration and installation dependencies of the models. You can read [here](https://lmdeploy.readthedocs.io/en/latest/multi_modal/index.html) for environment installation and configuration methods for different models. |
| 12 | + |
23 | 13 | ## A 'Hello, world' example |
24 | 14 |
|
25 | 15 | ```python |
|
0 commit comments