examples/provider-replicate/llama4-scout/README.md
You can run this example with:
npx promptfoo@latest init --example provider-replicate/llama4-scout
cd provider-replicate/llama4-scout
This example demonstrates how to use Replicate to run the new Llama 4 Scout model, a cutting-edge 17 billion parameter model with 16 experts using mixture-of-experts architecture.
Llama 4 Scout is part of the Llama 4 collection of natively multimodal AI models. Key features:
This example requires the following environment variable:
REPLICATE_API_TOKEN - Your Replicate API key (get one at https://replicate.com/account/api-tokens)You can set this in a .env file or directly in your environment:
export REPLICATE_API_TOKEN=your_api_token_here
This example:
promptfoo eval
promptfoo view
The example demonstrates key Replicate configuration options for Llama 4:
temperature: Controls randomness (0.0 = deterministic, 1.0 = very random)max_tokens: Maximum number of tokens to generatetop_p: Nucleus sampling threshold for token selectionThe example includes tests for:
You can modify this example to: