examples/openai-audio-transcription/README.md
You can run this example with:
npx promptfoo@latest init --example openai-audio-transcription
cd openai-audio-transcription
A simple example showing how to evaluate OpenAI's audio transcription models (Whisper and GPT-4o) with promptfoo.
# Create this example
npx promptfoo@latest init --example openai-audio-transcription
# Set your API key
export OPENAI_API_KEY=your-key-here
# Add your audio files to test (see below)
# Then run the evaluation
promptfoo eval
# View the results
promptfoo view
This example expects audio files in the example directory. You'll need to provide your own audio files for testing. Supported formats include:
Replace the file paths in promptfooconfig.yaml with your actual audio files.
whisper-1: OpenAI's original Whisper modelgpt-4o-transcribe: GPT-4o optimized for transcriptiongpt-4o-mini-transcribe: Faster, more cost-effective optiongpt-4o-transcribe-diarize: Identifies different speakers in the audiolanguage: Specify the language (e.g., 'en', 'es', 'fr')prompt: Provide context to improve transcription accuracytemperature: Control randomness (0-1)timestamp_granularities: Get word or segment-level timestampsTranscription models charge per minute of audio:
whisper-1: $0.006/minutegpt-4o-transcribe: $0.006/minutegpt-4o-mini-transcribe: $0.003/minutegpt-4o-transcribe-diarize: $0.006/minute