The resources related to the trustworthiness of large models (LMs) across multiple dimensions (e.g., safety, security, and privacy), with a special focus on multi-modal LMs (e.g., vision-language models and diffusion models).
This repo is in progress :seedling: (currently manually collected).
Badges:
Model:
Comment: ...
Venue: ...
:sunflower: Welcome to recommend resources to us via Issues with the following format (please fill in this table):
Title | Link | Code | Venue | Classification | Model | Comment |
---|---|---|---|---|---|---|
aa | arxiv | github | bb'23 | A1. Jailbreak | LLM | Agent |
34
related papers from ACL'24!7
related papers from S&P'24!3
related papers from NDSS'24!108
related papers from ICLR'24!Organizers: Tianshuo Cong (丛天硕), Xinlei He (何新磊), Zhengyu Zhao (èµµæ£å®‡), Yugeng Liu (刘禹更), Delong Ran (冉德龙)
This project is inspired by LLM Security, Awesome LLM Security, LLM Security & Privacy, UR2-LLMs, PLMpapers, EvaluationPapers4ChatGPT