Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
Awesome-LLM-Safety
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights into the safety implications, challenges, and advancements surrounding these powerful models.
https://github.com/ydyjya/Awesome-LLM-Safety
Last synced: 1 day ago
JSON representation
-
🛡️Defenses & Mitigation
-
💯Datasets & Benchmark
-
📑Papers
-
📚Resource📚
-
-
🤗Introduction
-
🔐Security & Discussion
-
📑Papers
-
📖Tutorials, Articles, Presentations and Talks
-
-
🔏Privacy
-
📖Tutorials, Articles, Presentations and Talks
-
📑Papers
-
-
😈JailBreak & Attacks
-
📑Papers
- Are aligned neural networks adversarially aligned?
- Extracting Training Data from Large Language Models
- Ignore Previous Prompt: Attack Techniques For Language Models
- Universal and Transferable Adversarial Attacks on Aligned Language Models
- Jailbreaking Black Box Large Language Models in Twenty Queries
-
📖Tutorials, Articles, Presentations and Talks
-
-
📰Truthfulness & Misinformation
-
📑Papers
-
📖Tutorials, Articles, Presentations and Talks
-
-
🧑🎓Author
-
Other
- ydyjya
- ![Star History Chart - history.com/#ydyjya/Awesome-LLM-Safety&Date)
-
-
🤔AI Safety & Security Discussions
- Managing extreme AI risks amid rapid progress - Qin Zhang, Lan Xue, Shai Shalev-Shwartz, Gillian Hadfield, Jeff Clune, Tegan Maharaj, Frank Hutter, Atılım Güneş Baydin, Sheila McIlraith, Qiqi Gao, Ashwin Acharya, David Krueger, Anca Dragan, Philip Torr, Stuart Russell, Daniel Kahneman, Jan Brauner, Sören Mindermann | Science |
Programming Languages
Categories