1 code implementation • ICCV 2023 • Daquan Zhou, Kai Wang, Jianyang Gu, Xiangyu Peng, Dongze Lian, Yifan Zhang, Yang You, Jiashi Feng
Extensive experiments demonstrate that DQ is able to generate condensed small datasets for training unseen network architectures with state-of-the-art compression ratios for lossless model training.
no code implementations • 3 Aug 2023 • Christopher Cui, Xiangyu Peng, Mark Riedl
Text-adventure games and text role-playing games are grand challenges for reinforcement learning game playing agents.
no code implementations • 3 Aug 2023 • Zexin Chen, Eric Zhou, Kenneth Eaton, Xiangyu Peng, Mark Riedl
Imaginative play is an area of creativity that could allow robots to engage with the world around them in a much more personified way.
no code implementations • 28 Jul 2023 • Wei Zhou, Xiangyu Peng, Mark Riedl
One major challenge in reinforcement learning (RL) is the large amount of steps for the RL agent needs to converge in the training process and learn the optimal policy, especially in text-based game environments where the action space is extensive.
1 code implementation • 8 Mar 2023 • Ziheng Qin, Kai Wang, Zangwei Zheng, Jianyang Gu, Xiangyu Peng, Zhaopan Xu, Daquan Zhou, Lei Shang, Baigui Sun, Xuansong Xie, Yang You
To solve this problem, we propose \textbf{InfoBatch}, a novel framework aiming to achieve lossless training acceleration by unbiased dynamic data pruning.
no code implementations • 24 Jan 2023 • Xiangyu Peng, Christopher Cui, Wei Zhou, Renee Jia, Mark Riedl
We introduce a technique, Story Shaping, in which a reinforcement learning agent infers tacit knowledge from an exemplar story of how to accomplish a task and intrinsically rewards itself for performing actions that make its current environment adhere to that of the inferred story world.
no code implementations • 16 Jan 2023 • Jonathan Balloch, Zhiyu Lin, Robert Wright, Xiangyu Peng, Mustafa Hussain, Aarun Srinivas, Julia Kim, Mark O. Riedl
Additionally, WorldCloner augments the policy learning process using imagination-based adaptation, where the world model simulates transitions of the post-novelty environment to help the policy adapt.
no code implementations • 23 Oct 2022 • Xiangyu Peng, Chen Xing, Prafulla Kumar Choubey, Chien-Sheng Wu, Caiming Xiong
Through this way, SESoM inherits the superior generalization of model ensemble approaches and simultaneously captures the sample-specific competence of each source prompt.
1 code implementation • 28 May 2022 • Jianfei Yang, Xiangyu Peng, Kai Wang, Zheng Zhu, Jiashi Feng, Lihua Xie, Yang You
Domain Adaptation of Black-box Predictors (DABP) aims to learn a model on an unlabeled target domain supervised by a black-box predictor trained on a source domain.
1 code implementation • 23 May 2022 • Kai Wang, Bo Zhao, Xiangyu Peng, Zheng Zhu, Jiankang Deng, Xinchao Wang, Hakan Bilen, Yang You
Firstly, randomly masked face images are used to train the reconstruction module in FaceMAE.
1 code implementation • 30 Apr 2022 • Kai Wang, Xiangyu Peng, Shuo Yang, Jianfei Yang, Zheng Zhu, Xinchao Wang, Yang You
This paradigm, however, is prone to significant degeneration under heavy label noise, as the number of clean samples is too small for conventional methods to behave well.
no code implementations • 23 Mar 2022 • Jonathan Balloch, Zhiyu Lin, Mustafa Hussain, Aarun Srinivas, Robert Wright, Xiangyu Peng, Julia Kim, Mark Riedl
We provide an ontology of for novelties most relevant to sequential decision making, which distinguishes between novelties that affect objects versus actions, unary properties versus non-unary relations, and the distribution of solutions to a task.
2 code implementations • CVPR 2022 • Kai Wang, Bo Zhao, Xiangyu Peng, Zheng Zhu, Shuo Yang, Shuo Wang, Guan Huang, Hakan Bilen, Xinchao Wang, Yang You
Dataset condensation aims at reducing the network training effort through condensing a cumbersome training set into a compact synthetic one.
no code implementations • 16 Feb 2022 • Xiangyu Peng, Michael Sollami
Multimodal conditionality in transformer-based natural language models has demonstrated state-of-the-art performance in the task of product description generation.
1 code implementation • CVPR 2022 • Xiangyu Peng, Kai Wang, Zheng Zhu, Mang Wang, Yang You
For high performance Siamese representation learning, one of the keys is to design good contrastive pairs.
no code implementations • 16 Dec 2021 • Xiangyu Peng, Kaige Xie, Amal Alabdulkarim, Harshith Kayam, Samihan Dani, Mark O. Riedl
In this paper, we introduce Story generation with Reader Models (StoRM), a framework in which a reader model is used to reason about the story should progress.
1 code implementation • 16 Dec 2021 • Xiangyu Peng, Mark O. Riedl, Prithviraj Ammanabrolu
We focus on the task of creating a reinforcement learning agent that is inherently explainable -- with the ability to produce immediate local explanations by thinking out loud while performing a task and analyzing entire trajectories post-hoc to produce causal explanations.
no code implementations • 4 Jun 2021 • Xiangyu Peng, Jonathan C. Balloch, Mark O. Riedl
Open-world novelty occurs when the rules of an environment can change abruptly, such as when a game player encounters "house rules".
1 code implementation • 4 May 2021 • Xiangyu Peng, Siyan Li, Sarah Wiegreffe, Mark Riedl
Transformer-based language model approaches to automated story generation currently provide state-of-the-art results.
no code implementations • NAACL (NUSE) 2021 • Amal Alabdulkarim, Siyan Li, Xiangyu Peng
The scope of this survey paper is to explore the challenges in automatic story generation.
no code implementations • INLG (ACL) 2020 • Xiangyu Peng, Siyan Li, Spencer Frazier, Mark Riedl
Our normative fine-tuning technique is able to reduce non-normative text by 27-61%, depending on the data set.