qwen

Qwen3 VL 4B - Instruct

Fine-tunable
text-to-textimage-to-text
Multimodal LLM for text and images, excelling in visual QA, document/UI understanding, spatial reasoning, image captioning, and multimodal coding.
About
Released: 8/27/2025

Qwen/Qwen3-VL-4B-Instruct is a multimodal LLM that processes both text and images, offering a relatively lightweight option for vision-language tasks while maintaining strong general language capabilities.

It excels in visual question answering, document and UI understanding, spatial reasoning over images, and general instruction-following dialogue, making it suitable when you need a compact model that can both see and read.

Some other noteworthy use cases of Qwen/Qwen3-VL-4B-Instruct include image captioning and explanation, multimodal coding assistance from designs or screenshots, and agentic visual assistants that can reason about interfaces and complex scenes.

MetricValue
Parameter Count4 billion
Mixture of ExpertsNo
Context Length256,000 tokens (up to 1M with extension)
MultilingualYes
Quantized*No

*Quantization is specific to the inference provider and the model may be offered with different quantization levels by other providers.