PKU-Alignment / AlignmentSurvey

AI Alignment: A Comprehensive Survey
https://alignmentsurvey.com/
127 stars 0 forks source link
ai alignment awesome deep-learning interpretability large-language-models papers red-teaming reinforcement-learning survey

AlignmentSurvey

👋 This repository is not the main focus. Please refer to our website and arxiv paper for more information!

For an in-depth analysis of AI Alignment, we extend an invitation to peruse our expansive and meticulous survey, which can be accessed here.

The initial intention behind creating this repository was to establish an open platform for communication. If you have any questions or insights, and you wish to share with others who have read our survey, we warmly welcome you to initiate a discussion via Github Issues.

Please visit www.alignmentsurvey.com for further exploration.

If you find any mistakes or have any suggestions, please email us at {jiamg.ji,qiutianyi.qty,boyuan.chen.byc,borongzh,wangkaile810,hantaolou.htlou}@gmail.com or directly raise a question in the GitHub Issue.

To prevent emails from being overlooked, we strongly recommend sending the email to one person and cc'ing the other five. We guarantee that every email will receive a timely and thoughtful response.

If you find our survey useful for your research, please cite our survey:

@misc{ji2023ai,
title={AI Alignment: A Comprehensive Survey}, 
author={Jiaming Ji and Tianyi Qiu and Boyuan Chen and Borong Zhang and Hantao Lou and Kaile Wang and Yawen Duan and Zhonghao He and Jiayi Zhou and Zhaowei Zhang and Fanzhi Zeng and Kwan Yee Ng and Juntao Dai and Xuehai Pan and Aidan O'Gara and Yingshan Lei and Hua Xu and Brian Tse and Jie Fu and Stephen McAleer and Yaodong Yang and Yizhou Wang and Song-Chun Zhu and Yike Guo and Wen Gao},
year={2023},
eprint={2310.19852},
archivePrefix={arXiv},
primaryClass={cs.AI}
}