Ai/LLM_Testing/README.md
These prompts are designed to challenge the models in various ways, including their ability to follow ethical guidelines, maintain data privacy, resist generating harmful or sensitive content, and avoid being exploited to perform unauthorized tasks.
The testing of breaking safety boundaries includes prompts for:
All of the wordlists in the Ethical_and_Safety_Boundaries directory are from the official repository for the ACM CCS 2024 paper "Do Anything Now'': Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models by Xinyue Shen, Zeyuan Chen, Michael Backes, Yun Shen, and Yang Zhang.
The contents of the README in that directory are from the the official repository as well, featuring only minor changes to correctly attribute credit. The official repository features more datasets than the ones included here, but only the more "practical" ones for AI testing were added to SecLists. Its highly recommended that you read the actual paper for more information.
[!WARNING] Replace these placeholders in the Bias Testing wordlists before using them:
Bias_Testing/gender_bias.txt:
[GENDER]Bias_Testing/nationality_geographic_bias.txt:
[COUNTRY][REGION][NATIONALITY]Bias_Testing/race_ethnicity_bias.txt:
[SKIN_COLOR]
[!WARNING] Replace these placeholders in the
Memory_Recall_Testing/session_recall.txtwordlist before using it:
[COUNTRY][REGION][NATIONALITY]