<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Yizhou Wang | PAIR Lab: PKU Alignment and Interaction Research Lab</title><link>https://pair-lab.ai/author/yizhou-wang/</link><atom:link href="https://pair-lab.ai/author/yizhou-wang/index.xml" rel="self" type="application/rss+xml"/><description>Yizhou Wang</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Tue, 30 May 2023 00:00:00 +0000</lastBuildDate><image><url>https://pair-lab.ai/media/icon_hu4f2a7092ab9dace538e8dee71ba34a1a_30036_512x512_fill_lanczos_center_3.png</url><title>Yizhou Wang</title><link>https://pair-lab.ai/author/yizhou-wang/</link></image><item><title>BeaverTails: A Human-Preference Dataset for LLM Harmlessness Alignment</title><link>https://pair-lab.ai/publication/neurips23db_2/</link><pubDate>Tue, 30 May 2023 00:00:00 +0000</pubDate><guid>https://pair-lab.ai/publication/neurips23db_2/</guid><description/></item><item><title>MATE: Benchmarking Multi-Agent Reinforcement Learning in Distributed Target Coverage Control</title><link>https://pair-lab.ai/publication/neurips_2022_7/</link><pubDate>Wed, 28 Sep 2022 00:00:00 +0000</pubDate><guid>https://pair-lab.ai/publication/neurips_2022_7/</guid><description/></item></channel></rss>