Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, with open weights for both pre-trained variants and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as a laptop, desktop or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.
Input: Text
Output: Text
Providers
chutes
Credits
Context8k
Max Output8k
Input$0.010/1M
Output$0.030/1M
Cache Read—
Cache Write—
openrouter
Credits
Context8k
Max Output8k
Input (Max)$0.210/1M
Output (Max)$0.210/1M
Cache Read—
Cache Write—
Quick Start
Use Google Gemma 2 through Helicone's AI Gateway with automatic logging and monitoring.