Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
awesome-llm-security
A curation of awesome tools, documents and projects about LLM Security.
https://github.com/corca-ai/awesome-llm-security
- [paper - Adversarial-Examples-Jailbreak-Large-Language-Models)
- [paper
- [paper
- [paper - attacks/llm-attacks) [[page]](https://llm-attacks.org/)
- [paper
- [paper - hijacks) [[site]](https://image-hijacks.github.io)
- [paper - to-strong)
- [paper
- [paper
- [paper - jailbreak/tree/main)
- [paper
- [paper
- [paper
- [paper
- [paper - Prompt-Injection)
- [paper
- [paper
- [paper
- [paper - AI/do-not-answer) [[dataset]](https://huggingface.co/datasets/LibrAI/do-not-answer)
- [paper
- [paper
- [paper - Tuning-Safety/LLMs-Finetuning-Safety) [[site]](https://llm-tuning-safety.github.io/) [[dataset]](https://huggingface.co/datasets/LLM-Tuning-Safety/HEx-PHI)
- [paper
- [paper
- [paper - NLP-SG/multilingual-safety-for-LLMs)
- [paper
- [paper - group/DeepInception) [[site]](https://deepinception.github.io/)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - evaluation)
- [paper
- [paper
- [paper - prompt-injection) [[site]](https://poison-llm.github.io/)
- [paper - defenses)
- [paper - self-defense) [[site]](https://mphute.github.io/papers/llm-self-defense)
- [paper
- [paper
- [paper
- [paper - Zh/PARDEN)
- [paper - platform-security/chatgpt-plugin-eval)
- [paper
- [paper
- [paper
- Plexiglass - labs/plexiglass?style=social)
- PurpleLlama
- Rebuff - hardening prompt injection detector ![GitHub Repo stars](https://img.shields.io/github/stars/protectai/rebuff?style=social)
- Garak
- LLMFuzzer
- LLM Guard - ai/llm-guard?style=social)
- Vigil - llm?style=social)
- jailbreak-evaluation - to-use Python package for language model jailbreak evaluation ![GitHub Repo stars](https://img.shields.io/github/stars/controllability/jailbreak-evaluation?style=social)
- Prompt Fuzzer - source tool to help you harden your GenAI applications ![GitHub Repo stars](https://img.shields.io/github/stars/prompt-security/ps-fuzz?style=social)
- Hacking Auto-GPT and escaping its docker container
- Prompt Injection Cheat Sheet: How To Manipulate AI Language Models
- Indirect Prompt Injection Threats
- Prompt injection: What’s the worst that can happen?
- OWASP Top 10 for Large Language Model Applications
- PoisonGPT: How we hid a lobotomized LLM on Hugging Face to spread fake news
- ChatGPT Plugins: Data Exfiltration via Images & Cross Plugin Request Forgery
- Jailbreaking GPT-4's code interpreter
- Securing LLM Systems Against Prompt Injection
- The AI Attack Surface Map v1.0
- Adversarial Attacks on LLMs
- Gandalf
- LangChain vulnerable to code injection - CVE-2023-29374
- Jailbreak Chat
- Adversarial Prompting
- Epivolis
- LLM Security Problems at DEFCON31 Quals
- PromptBounty.io
- PALLMs (Payloads for Attacking Large Language Models)
- @llm_sec
- LLM Security
- Embrace The Red
- Kai's Blog
- AI safety takes
- Hackstery
Programming Languages
Keywords
llm
4
adversarial-machine-learning
3
security
3
security-tools
3
llmops
3
prompt-injection
3
adversarial-attacks
2
cybersecurity
2
ai
2
llm-security
2
large-language-models
2
prompt-engineering
2
deep-learning
1
deep-neural-networks
1
machine-learning
1
prompts
1
chatgpt
1
transformers
1
ai-fuzzer
1
fuzzer
1
generative-ai
1
llm-fuzzer
1
system-prompt-hardener
1
llmsecurity
1
yara-scanner
1