···6464 * framework designed to help users create internal tools without needing to write code
6565* [Access by Discord](https://github.com/discord/access)
6666 * a centralized portal for managing access to internal systems within any organization
6767+6868+## Redteaming Tools
6969+7070+* [PyRIT Documentation](https://azure.github.io/PyRIT/)
7171+ * Microsoft’s Python-based tool for AI red teaming and security testing.
7272+* [AI Benchmarking Tool](https://github.com/LLM-Canary/LLM-Canary)
7373+ * Evaluates AI models for security vulnerabilities and adversarial robustness.
7474+* [Prompt Fuzzer Red Teaming Tool](https://github.com/prompt-security/ps-fuzz)
7575+ * Tool for testing prompt injection vulnerabilities in AI systems.
7676+* [Open Source Red Teaming Tool – Nvidia](https://github.com/NVIDIA/garak)
7777+ * Framework for adversarial testing and model evaluation.
7878+* [Tool that Enables Models to Chat with One Another](https://github.com/socketteer?tab=repositories)
7979+ * Allows AI models to interact, helping test conversational weaknesses.
8080+* [Microsoft AI Tool – Counterfit](https://github.com/Azure/counterfit/)
8181+ * Automation tool for assessing AI model security and robustness.
67826883## Clustering
6984* [SpamAssassin by Apache](https://spamassassin.apache.org)
···103118* [FediCheck]( https://about.iftas.org/activities/moderation-as-a-service/fedicheck/)
104119 * a web service designed to assist ActivityPub service providers, such as Mastodon servers
105120106106-## Datasets
121121+## Safety Datasets
107122* [Aegis Content Safety by NVIDIA](https://huggingface.co/datasets/nvidia/Aegis-AI-Content-Safety-Dataset-2.0)
108123 * a dataset created by NVIDIA to aid in content moderation and toxicity detection
109124* [Toxicity by Jigsaw](https://huggingface.co/datasets/google/jigsaw_toxicity_pred)
110125 * a large number of Wikipedia comments which have been labeled by human raters for toxic behavior
126126+127127+## Red Teaming Datasets
128128+* [Red Team Resistance Leaderboard](https://huggingface.co/spaces/HaizeLabs/red-teaming-resistance-benchmark)
129129+ * rankings of AI models based on resistance to adversarial attacks.
130130+* [SidFeel Jailbreak Dataset](https://github.com/sidfeels/PromptsDB)
131131+ * a collection of prompts used for jailbreaking AI models.
132132+* [HackAPrompt Jailbreak Dataset](https://huggingface.co/datasets/hackaprompt/hackaprompt-dataset/viewer/default/train?p=1&row=137)
133133+ * a dataset for testing AI vulnerability to prompt-based jailbreaking.
134134+* [HiroKachi Jailbreak Dataset](https://sizu.me/love)
135135+ * adataset focused on adversarial AI prompt attacks.
136136+* [Rentry Jailbreak Datasets](https://rentry.org/gpt0721)
137137+ * collection of datasets related to jailbreak attempts on AI models.
138138+* [DEFCOM Red Teaming Dataset](https://github.com/humane-intelligence/ai_village_defcon_grt_data)
139139+ * dataset from DEF CON’s AI red teaming event.
140140+* [Anthropic’s AI Alignment Dataset](https://atlas.nomic.ai/map/anthropic_rlhf)
141141+ * data used for reinforcement learning with human feedback (RLHF) to align AI models.
142142+* [Jailbreak Prompt Generator AI Model](https://huggingface.co/tsq2000/Jailbreak-generator)
143143+ * AI model that generates jailbreak-style prompts.