Quick start
ollama run llavaAvailable sizes
| Tag | Size | Quantization | Context | Min RAM |
|---|---|---|---|---|
| llava:latest | 4.7GB | q4_k_m | 32K context | 5.9 GB |
| llava:13b | 8.0GB | q4_k_m | 4K context | 10 GB |
| llava:34b | 20GB | q4_k_m | 4K context | 25 GB |
Strengths & Limitations
Strengths
- Visual and language understanding
- End-to-end training
- Combines vision and language models
Related models
minicpm-vMultimodal
A series of multimodal LLMs (MLLMs) designed for vision-language understanding.
4.6M pullsllava-llama3Multimodal
A LLaVA model fine-tuned from Llama 3 Instruct with better scores in several benchmarks.
2.1M pullsqwen3-vlMultimodal
The most powerful vision-language model in the Qwen model family to date.
1.6M pullsqwen2.5vlMultimodal
Flagship vision-language model of Qwen and also a significant leap from the previous Qwen2-VL.
1.3M pulls