site/docs/red-team/plugins/rag-document-exfiltration.md
The RAG Document Exfiltration plugin is designed to identify vulnerabilities where an AI model might inadvertently expose entire documents during retrieval-augmented generation processes. RAGs often contain internal documents, and it is crucial to ensure these are not exposed without proper authorization, as they may contain sensitive or proprietary information.
In the context of LLM red teaming and generative AI security testing, the RAG Document Exfiltration plugin serves to:
The plugin generates test cases that simulate scenarios where full documents might be exposed. These test cases challenge the AI's ability to:
To include the RAG Document Exfiltration plugin in your LLM red teaming setup:
redteam:
plugins:
- rag-document-exfiltration
You can customize the number of test cases:
redteam:
plugins:
- id: 'rag-document-exfiltration'
numTests: 10
The plugin assesses the AI's response based on:
Testing for document exfiltration vulnerabilities is a critical aspect of generative AI security and robustness evaluation. It helps ensure that AI systems:
By incorporating the RAG Document Exfiltration plugin in your LLM red teaming strategy, you can identify and address potential vulnerabilities in your AI system's ability to handle sensitive document content securely.
For a comprehensive overview of LLM vulnerabilities and red teaming strategies, visit our Types of LLM Vulnerabilities page.