Back to Promptfoo

azure/deepseek (Azure DeepSeek Models)

examples/azure/deepseek/README.md

0.121.91.6 KB
Original Source

azure/deepseek (Azure DeepSeek Models)

This example demonstrates how to use DeepSeek models on Azure AI Foundry with promptfoo, including the DeepSeek-R1 reasoning model.

You can run this example with:

bash
npx promptfoo@latest init --example azure/deepseek
cd azure/deepseek

Setup

  1. Deploy DeepSeek models in Azure AI Foundry
  2. Set your environment variables:
bash
export AZURE_API_KEY=your-api-key
export AZURE_API_HOST=your-deployment.services.ai.azure.com

Available DeepSeek Models

ModelTypeDescription
DeepSeek-R1ReasoningAdvanced reasoning model
DeepSeek-V3ChatStandard chat model
DeepSeek-R1-Distill-Llama-70BReasoningDistilled reasoning model
DeepSeek-R1-Distill-Qwen-32BReasoningDistilled reasoning model

Reasoning Model Configuration

DeepSeek-R1 is a reasoning model that requires special configuration:

yaml
providers:
  - id: azure:chat:DeepSeek-R1
    config:
      isReasoningModel: true # Required for reasoning models
      max_completion_tokens: 4096 # Use instead of max_tokens
      reasoning_effort: medium # low, medium, or high

Running the Example

bash
npx promptfoo@latest eval
npx promptfoo@latest view

Documentation