Llama 3.2 11B Vision is a 11 billion parameter language model from Meta, part of Meta's open-weight Llama family. It features accepts image inputs alongside text. It is released under Meta's Llama Community License.
Input
Output
Context
—
Max Output
8K
Parameters
11B
Input Modalities
Output Modalities
Features
Estimates based on INT8 quantization. Actual requirements vary by framework and configuration.
Data sourced from official provider APIs and documentation
Last updated: May 5, 2026
From model selection to production, one platform, no fragmentation.