Awesome LLM Safety Papers 📚 This repository contains list of papers according to our survey: Large Language Model Safety: A Holistic Survey Dan Shi1*, Tianhao Shen1*, Yufei Huang1, Zhigen Li1,2, Yongqi Leng1, Renren Jin1, Chuang Liu1, Xinwei Wu1, Zishan Guo1, Linhao Yu1, Ling ...
If a resource is relevant to multiple subcategories, we place it under each applicable section. For instance, the "Awesome-LLM-Safety" repository will be listed under each subcategory to which it pertains🤩!. ✔️ Perfect for Majority ...
Lessons Learned from Cybersecurity | **LLM Safety Fine-Tuning**&**Cybersecurity Lessons**&**Jailbreak Defense** | | 25.01 | Squirrel Ai Learning | arxiv | [LLM-Virus: Evolutionary Jailbreak Attack on Large Language Models](https://arxiv.org/abs/2501.00055) | **LLM Safety**&**Jailbreak...
Awesome-MLLM-Safety A collection (won't be updated) of papers related to safety of Multimodal Large Language Models (MLLMs). We follow the definition of safety from the paperSafety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions: ...
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights into the safety implications, challenges, and advancements
A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.). - ThuCCSLab/Awesome-LM-SSP