meta-llama: Meta Llama 3.1 8B Instant

llama-3.1-8b-instant
Context: 131k
Max Output: 33k
Input: $0.050/1M
Output: $0.080/1M
Compact 8B general-purpose model offering efficient inference for chat, coding, and RAG workflows on limited compute.
Input: Text
Output: Text

Providers

groq
Credits
Context131k
Max Output33k
Input$0.050/1M
Output$0.080/1M
Cache Read
Cache Write
openrouter
Credits
Context131k
Max Output33k
Input (Max)$0.210/1M
Output (Max)$0.310/1M
Cache Read
Cache Write

Quick Start

Use Meta Llama 3.1 8B Instant through Helicone's AI Gateway with automatic logging and monitoring.