Back to Promptfoo

azure/llama (Azure Llama Models)

examples/azure/llama/README.md

0.121.91.6 KB
Original Source

azure/llama (Azure Llama Models)

This example demonstrates how to use Meta Llama models on Azure AI Foundry with promptfoo.

You can run this example with:

bash
npx promptfoo@latest init --example azure/llama
cd azure/llama

Setup

  1. Deploy Llama models in Azure AI Foundry
  2. Set your environment variables:
bash
export AZURE_API_KEY=your-api-key
export AZURE_API_HOST=your-deployment.services.ai.azure.com

Available Llama Models

ModelDescription
Llama-4-Maverick-17B-128E-Instruct-FP8Llama 4 Maverick (128 experts, FP8)
Llama-4-Scout-17B-16E-InstructLlama 4 Scout (16 experts)
Llama-3.3-70B-InstructLlama 3.3 70B
Meta-Llama-3.1-405B-InstructLlama 3.1 405B
Meta-Llama-3.1-70B-InstructLlama 3.1 70B
Meta-Llama-3.1-8B-InstructLlama 3.1 8B

Running the Example

bash
npx promptfoo@latest eval
npx promptfoo@latest view

Configuration

The example compares Llama 4 Maverick and Llama 4 Scout on code generation tasks. This helps evaluate the trade-off between model capacity (expert count), speed, and quality.

Documentation