llama-4-scout-17b-16e-instruct
by meta-llamaLlama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model from Meta, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, Scout uses 16 experts per forward pass and features a context length of 10 million tokens, with a training corpus of ~40 trillion tokens. Built for high efficiency and local or commercial deployment, it is instruction-tuned for multilingual chat, captioning, and image understanding.
Pricing
Pay-as-you-go rates for this model. More details can be found here.
Input Tokens (1M)
$0.24
Output Tokens (1M)
$0.96
Capabilities
Input Modalities
Output Modalities
Rate Limits
Requests per minute (RPM) and per day (RPD) by tier. More about tiers here
Tier | RPM | RPD |
---|---|---|
Free | 2 | 50 |
Tier 1 | 10 | — |
Tier 2 | 15 | — |
Tier 3 | 25 | — |
Tier 4 | 50 | — |
Usage Analytics
Token usage across the last 30 active days