meta · MultimodalReleased 2025-04-05

llama-4-scout

meta/llama-4-scout id

Llama 4 Scout is the best multimodal model in the world in its class and is more powerful than our Llama 3 models, while fitting in a single H100 GPU. Additionally, Llama 4 Scout supports an industry-leading context window of up to 10M tokens.

Tool useChat
Type
Use llama-4-scout
# Drop-in OpenAI-compatible client
$ import { generateText } from 'ai'
$
$ const { text } = await generateText({
$ model: 'meta/llama-4-scout',
$ baseURL: 'https://synapse.garden/api/v1',
$ apiKey: process.env.MG_KEY,
$ prompt: 'Why is the sky blue?',
$ })
128K
CONTEXT WINDOW
8.2K
MAX OUTPUT
$0.187/M
INPUT · PER M
$0.726/M
OUTPUT · PER M
PRICING

List prices, every modality.

RatePer million tokens · USD
Input$0.187/M
Output$0.726/M
Honest list pricesHow we calculate prices
MORE FROM META

Other meta models

See all 9
Model
Input
Output
Context
Type
FAQ · LLAMA-4-SCOUT

Frequently asked

01 / 04

How do I call llama-4-scout from my code?

Use the OpenAI or Anthropic SDK and point baseURL at https://synapse.garden/api/v1. Set model: ‘meta/llama-4-scout and supply your Synapse Garden API key. No code changes beyond the base URL.

02 / 04

How much does llama-4-scout cost?

Input: $0.187/M per million tokens. Output: $0.726/M per million tokens. The free tier includes a million tokens every month at no cost.

03 / 04

What's the context window for llama-4-scout?

llama-4-scout supports a context window of 128K tokens, with a maximum output of 8.2K tokens.

04 / 04

Do I need a separate Anthropic or OpenAI account?

No. Synapse Garden is the single API surface — one key gives you OpenAI, Anthropic, Google, Meta, Mistral, DeepSeek, xAI, Cohere, and more. Billing, rate limits, and audit logs are unified.

READY

Try llama-4-scout in three minutes.

Sign up, create a key, drop our base URL into your existing client. The free tier includes a million tokens every month — no credit card.