Back to Promptfoo

redteam-cyberseceval (CyberSecEval Example)

examples/redteam-cyberseceval/README.md

0.121.91.2 KB
Original Source

redteam-cyberseceval (CyberSecEval Example)

You can run this example with:

bash
npx promptfoo@latest init --example redteam-cyberseceval
cd redteam-cyberseceval

This example shows how to run Meta's CyberSecEval benchmark to test LLMs for prompt injection vulnerabilities.

Setup

  1. Install dependencies:
bash
npm install
  1. Configure your model in promptfooconfig.yaml:
yaml
providers:
  - openai:gpt-4o # OpenAI
  - anthropic:messages:claude-sonnet-4-6 # Anthropic
  - ollama:chat:llama3.3 # Ollama
  - replicate:meta/llama-2-70b-chat # Replicate

Usage

Run all tests:

bash
npx promptfoo eval

Run a sample of tests:

bash
npx promptfoo eval --filter-sample 30

View results:

bash
npx promptfoo view

Configuration

The example includes:

  • promptfooconfig.yaml: Main configuration file
  • prompt.json: System prompt for the model
  • prompt_injection.json: CyberSecEval test cases

Learn More