A lightweight 1 billion parameter open-source model from Google, based on the Gemma 3 architecture. Designed for on-device and resource-constrained deployments with a 32K context window. Small enough to run on consumer hardware while still handling basic text generation tasks.
Input
Output
Context
33K
Max Output
8K
Parameters
1B
Input Modalities
Output Modalities
Features
Estimates based on INT8 quantization. Actual requirements vary by framework and configuration.
Data sourced from official provider APIs and documentation
Last updated: May 5, 2026
From model selection to production, one platform, no fragmentation.