⚠ This tool appears inactive — no commits in 90+ days. Consider an alternative.
MultimodalOpen Source✦ Free Tier

Qwen-VL

Alibaba's open-weight vision-language model

15,000 stars● Health 40SlowingApp Infrastructure

About

Qwen Visual Language model series from Alibaba. Strong at multilingual visual understanding, document parsing, and chart reading. Available as open weights on HuggingFace. Runs via vLLM.

Choose Qwen-VL when…

  • You need multilingual visual understanding (especially CJK languages)
  • Chart, table, and document parsing is the primary use case
  • You want strong performance across multiple model sizes

Builder Slot

How does your AI see and understand images?Optional for most stacks

Vision-language models for image understanding, captioning, visual QA, and document parsing

Dev Tools
Not applicable
App Infra
Optional
Hybrid
Optional

Other tools in this slot:

Stack Genome Detection

AIchitect's Genome scanner detects Qwen-VL in your project via these signals:

pip packages
transformers

Integrates with (1)

vLLMLLM Infrastructure
Compare →

Alternatives to consider (3)

Pricing

✦ Free tier available

Badge

Add to your GitHub README

Qwen-VL on AIchitect[![Qwen-VL](https://aichitect.dev/badge/tool/qwen-vl)](https://aichitect.dev/tool/qwen-vl)

Explore the full AI landscape

See how Qwen-VL fits into the bigger picture — browse all 207 tools and their relationships.

Explore graph →