PAIR Lab: PKU Alignment and Interaction Research Lab
PAIR Lab: PKU Alignment and Interaction Research Lab
Open-Source Projects
People
News
Publications
Resources
Contact
Jiayi Zhou
Latest
Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedback
Omnisafe: An infrastructure for accelerating safe reinforcement learning research
Language models resist alignment: Evidence from data compression
Safety Gymnasium: A Unified Safe Reinforcement Learning Benchmark
Cite
×