https://github.com/SXQBW/ComfyUI-Qwen-VL
https://github.com/SXQBW/ComfyUI-Qwen-VL
Last synced: 7 days ago
JSON representation
- Host: GitHub
- URL: https://github.com/SXQBW/ComfyUI-Qwen-VL
- Owner: SXQBW
- License: gpl-3.0
- Created: 2025-05-05T15:21:53.000Z (about 1 month ago)
- Default Branch: main
- Last Pushed: 2025-05-26T06:11:20.000Z (23 days ago)
- Last Synced: 2025-05-26T07:52:46.239Z (23 days ago)
- Language: Python
- Size: 5.79 MB
- Stars: 4
- Watchers: 1
- Forks: 1
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- awesome-comfyui - **ComfyUI-Qwen-VL** - VL series large language models, supporting multi-modal functions such as text generation, image understanding, and video analysis.Support for Qwen2-VL, Qwen2.5-VL. (All Workflows Sorted by GitHub Stars)
README
# ComfyUI-Qwen-VL πΌ
δΈζ  ο½   English**Where Figma meets VSCode: Artistic vision meets engineering precision ββ a romantic manifesto from designers to the code world.**
β¨ A ComfyUI extension for Qwen2.5-VL series large language models, supporting multimodal capabilities such as text generation, image understanding, and video analysis. β¨
[](https://github.com/SXQBW/ComfyUI-Qwen-VL/stargazers)
[](https://huggingface.co/Qwen)
![]()
---
### π Features
- Supports Qwen2-VL, Qwen2.5-VL and other series models
- Provides various functional nodes for text generation, image understanding, video analysis, etc.
- Supports model quantization configuration to optimize memory usage
- Offers an intuitive user interface for easy parameter adjustment### π Installation
1. Navigate to the `custom_nodes` directory of ComfyUI
2. Clone this repository:
```bash
git clone https://github.com/SXQBW/ComfyUI-Qwen-VL.git
```
3. Install dependencies:
```bash
cd ComfyUI-Qwen-VL
pip install -r requirements.txt
```
4. Restart ComfyUI### π Usage
1. In the ComfyUI interface, locate the Qwen-VL related nodes
2. Select the model and quantization method you want to use
3. Configure generation parameters such as temperature and maximum tokens
4. Connect input (text, image, or video) and output nodes
5. Run the workflow### π¦ Supported Models
The following models are currently supported:
- Qwen/Qwen2.5-VL-3B-Instruct
- Qwen/Qwen2.5-VL-3B-Instruct-AWQ
- Qwen/Qwen2.5-VL-7B-Instruct
- Qwen/Qwen2.5-VL-7B-Instruct-AWQ
- Qwen/Qwen2.5-VL-32B-Instruct
- Qwen/Qwen2.5-VL-32B-Instruct-AWQ
- Qwen/Qwen2.5-VL-72B-Instruct
- Qwen/Qwen2.5-VL-72B-Instruct-AWQ
- Qwen/Qwen2-VL-2B
- Qwen/Qwen2-VL-2B-Instruct
- Qwen/Qwen2-VL-7B-Instruct
- Qwen/Qwen2-VL-72B-Instruct
- Qwen/Qwen2-VL-2B-Instruct-AWQ
- Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4
- Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8
- Qwen/Qwen2-VL-7B-Instruct-AWQ
- Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int4
- Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8
- Qwen/Qwen2-VL-72B-Instruct-AWQ
- Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int4
- Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int8
- huihui-ai/Qwen2.5-VL-7B-Instruct-abliterated### Quantization Options
- π« None (Original Precision): Use original precision
- π 4-bit (VRAM-friendly): Use 4-bit quantization to save VRAM
- βοΈ 8-bit (Balanced Precision): Use 8-bit quantization for balanced precision and performance### π Example Workflows
Here's a simple example workflow for image understanding:


### π FAQ
#### Model Loading Issues
If you encounter errors loading the model, ensure:
1. The model file path is correct
2. You have sufficient GPU memory (Choose an appropriate model based on your VRAM size. Don't jump straight to the 72B model β brute force won't work here, it'll just crash your VRAM)
3. All necessary dependencies are installed#### About Quantization
When using pre-quantized models (e.g., AWQ versions), you may see the following warning: "Model Qwen2.5-VL-3B-Instruct-AWQ is already quantized, user quantization settings will be ignored." This is normal, and the plugin will automatically use the model's pre-quantized version.
### π€ Contributing
Contributions, issues, and feature requests are welcome!
### π Acknowledgments
Special thanks to the Qwen team for developing these powerful models, and to the ComfyUI community for their support!
**The star you're about to click β¨**
Is not just a gesture of approval, but a cosmic explosion where design thinking meets the code universe. When an artist's aesthetic obsession collides with a programmer's geek spirit β this might just be the most romantic chemical reaction on GitHub.[Click to Star and Witness the Cross-Disciplinary Revolution](https://github.com/SXQBW/ComfyUI-Qwen-VL)