Search Results for author: Zhaowei Zhang

Found 10 papers, 4 papers with code

Incentive Compatibility for AI Alignment in Sociotechnical Systems: Positions and Prospects

no code implementations20 Feb 2024 Zhaowei Zhang, Fengshuo Bai, Mingzhi Wang, Haoyang Ye, Chengdong Ma, Yaodong Yang

The burgeoning integration of artificial intelligence (AI) into human society brings forth significant implications for societal governance and safety.

CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents

1 code implementation19 Jan 2024 Siyuan Qi, Shuo Chen, Yexin Li, Xiangyu Kong, Junqi Wang, Bangcheng Yang, Pring Wong, Yifan Zhong, Xiaoyuan Zhang, Zhaowei Zhang, Nian Liu, Wei Wang, Yaodong Yang, Song-Chun Zhu

Within CivRealm, we provide interfaces for two typical agent types: tensor-based agents that focus on learning, and language-based agents that emphasize reasoning.

Decision Making

AI Alignment: A Comprehensive Survey

no code implementations30 Oct 2023 Jiaming Ji, Tianyi Qiu, Boyuan Chen, Borong Zhang, Hantao Lou, Kaile Wang, Yawen Duan, Zhonghao He, Jiayi Zhou, Zhaowei Zhang, Fanzhi Zeng, Kwan Yee Ng, Juntao Dai, Xuehai Pan, Aidan O'Gara, Yingshan Lei, Hua Xu, Brian Tse, Jie Fu, Stephen Mcaleer, Yaodong Yang, Yizhou Wang, Song-Chun Zhu, Yike Guo, Wen Gao

The former aims to make AI systems aligned via alignment training, while the latter aims to gain evidence about the systems' alignment and govern them appropriately to avoid exacerbating misalignment risks.

Measuring Value Understanding in Language Models through Discriminator-Critique Gap

no code implementations30 Sep 2023 Zhaowei Zhang, Fengshuo Bai, Jun Gao, Yaodong Yang

We argue that truly understanding values in LLMs requires considering both "know what" and "know why".

Heterogeneous Value Alignment Evaluation for Large Language Models

2 code implementations26 May 2023 Zhaowei Zhang, Ceyao Zhang, Nian Liu, Siyuan Qi, Ziqi Rong, Song-Chun Zhu, Shuguang Cui, Yaodong Yang

We conduct evaluations with new auto-metric \textit{value rationality} to represent the ability of LLMs to align with specific values.

Attribute

STAS: Spatial-Temporal Return Decomposition for Multi-agent Reinforcement Learning

1 code implementation15 Apr 2023 Sirui Chen, Zhaowei Zhang, Yaodong Yang, Yali Du

It first decomposes the global return back to each time step, then utilizes the Shapley Value to redistribute the individual payoff from the decomposed global reward.

Multi-agent Reinforcement Learning reinforcement-learning

Contextual Transformer for Offline Meta Reinforcement Learning

no code implementations15 Nov 2022 Runji Lin, Ye Li, Xidong Feng, Zhaowei Zhang, Xian Hong Wu Fung, Haifeng Zhang, Jun Wang, Yali Du, Yaodong Yang

Firstly, we propose prompt tuning for offline RL, where a context vector sequence is concatenated with the input to guide the conditional policy generation.

D4RL Meta Reinforcement Learning +4

Continuous Decomposition of Granularity for Neural Paraphrase Generation

1 code implementation COLING 2022 Xiaodong Gu, Zhaowei Zhang, Sang-Woo Lee, Kang Min Yoo, Jung-Woo Ha

While Transformers have had significant success in paragraph generation, they treat sentences as linear sequences of tokens and often neglect their hierarchical information.

Paraphrase Generation Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.