Search Results for author: Xiangyu Peng

Found 21 papers, 9 papers with code

Dataset Quantization

1 code implementation ICCV 2023 Daquan Zhou, Kai Wang, Jianyang Gu, Xiangyu Peng, Dongze Lian, Yifan Zhang, Yang You, Jiashi Feng

Extensive experiments demonstrate that DQ is able to generate condensed small datasets for training unseen network architectures with state-of-the-art compression ratios for lossless model training.

object-detection Object Detection +2

Thespian: Multi-Character Text Role-Playing Game Agents

no code implementations3 Aug 2023 Christopher Cui, Xiangyu Peng, Mark Riedl

Text-adventure games and text role-playing games are grand challenges for reinforcement learning game playing agents.

Few-Shot Learning

Ambient Adventures: Teaching ChatGPT on Developing Complex Stories

no code implementations3 Aug 2023 Zexin Chen, Eric Zhou, Kenneth Eaton, Xiangyu Peng, Mark Riedl

Imaginative play is an area of creativity that could allow robots to engage with the world around them in a much more personified way.

Story Generation

Dialogue Shaping: Empowering Agents through NPC Interaction

no code implementations28 Jul 2023 Wei Zhou, Xiangyu Peng, Mark Riedl

One major challenge in reinforcement learning (RL) is the large amount of steps for the RL agent needs to converge in the training process and learn the optimal policy, especially in text-based game environments where the action space is extensive.

Knowledge Graphs reinforcement-learning +1

InfoBatch: Lossless Training Speed Up by Unbiased Dynamic Data Pruning

1 code implementation8 Mar 2023 Ziheng Qin, Kai Wang, Zangwei Zheng, Jianyang Gu, Xiangyu Peng, Zhaopan Xu, Daquan Zhou, Lei Shang, Baigui Sun, Xuansong Xie, Yang You

To solve this problem, we propose \textbf{InfoBatch}, a novel framework aiming to achieve lossless training acceleration by unbiased dynamic data pruning.

Semantic Segmentation

Story Shaping: Teaching Agents Human-like Behavior with Stories

no code implementations24 Jan 2023 Xiangyu Peng, Christopher Cui, Wei Zhou, Renee Jia, Mark Riedl

We introduce a technique, Story Shaping, in which a reinforcement learning agent infers tacit knowledge from an exemplar story of how to accomplish a task and intrinsically rewards itself for performing actions that make its current environment adhere to that of the inferred story world.

reinforcement-learning Reinforcement Learning (RL) +1

Neuro-Symbolic World Models for Adapting to Open World Novelty

no code implementations16 Jan 2023 Jonathan Balloch, Zhiyu Lin, Robert Wright, Xiangyu Peng, Mustafa Hussain, Aarun Srinivas, Julia Kim, Mark O. Riedl

Additionally, WorldCloner augments the policy learning process using imagination-based adaptation, where the world model simulates transitions of the post-novelty environment to help the policy adapt.

Decision Making reinforcement-learning +1

Model ensemble instead of prompt fusion: a sample-specific knowledge transfer method for few-shot prompt tuning

no code implementations23 Oct 2022 Xiangyu Peng, Chen Xing, Prafulla Kumar Choubey, Chien-Sheng Wu, Caiming Xiong

Through this way, SESoM inherits the superior generalization of model ensemble approaches and simultaneously captures the sample-specific competence of each source prompt.

Transfer Learning

Divide to Adapt: Mitigating Confirmation Bias for Domain Adaptation of Black-Box Predictors

1 code implementation28 May 2022 Jianfei Yang, Xiangyu Peng, Kai Wang, Zheng Zhu, Jiashi Feng, Lihua Xie, Yang You

Domain Adaptation of Black-box Predictors (DABP) aims to learn a model on an unlabeled target domain supervised by a black-box predictor trained on a source domain.

Domain Adaptation Knowledge Distillation

Reliable Label Correction is a Good Booster When Learning with Extremely Noisy Labels

1 code implementation30 Apr 2022 Kai Wang, Xiangyu Peng, Shuo Yang, Jianfei Yang, Zheng Zhu, Xinchao Wang, Yang You

This paradigm, however, is prone to significant degeneration under heavy label noise, as the number of clean samples is too small for conventional methods to behave well.

Learning with noisy labels

NovGrid: A Flexible Grid World for Evaluating Agent Response to Novelty

no code implementations23 Mar 2022 Jonathan Balloch, Zhiyu Lin, Mustafa Hussain, Aarun Srinivas, Robert Wright, Xiangyu Peng, Julia Kim, Mark Riedl

We provide an ontology of for novelties most relevant to sequential decision making, which distinguishes between novelties that affect objects versus actions, unary properties versus non-unary relations, and the distribution of solutions to a task.

Decision Making reinforcement-learning +1

CAFE: Learning to Condense Dataset by Aligning Features

2 code implementations CVPR 2022 Kai Wang, Bo Zhao, Xiangyu Peng, Zheng Zhu, Shuo Yang, Shuo Wang, Guan Huang, Hakan Bilen, Xinchao Wang, Yang You

Dataset condensation aims at reducing the network training effort through condensing a cumbersome training set into a compact synthetic one.

Dataset Condensation

XFBoost: Improving Text Generation with Controllable Decoders

no code implementations16 Feb 2022 Xiangyu Peng, Michael Sollami

Multimodal conditionality in transformer-based natural language models has demonstrated state-of-the-art performance in the task of product description generation.

Descriptive Language Modelling +1

Guiding Neural Story Generation with Reader Models

no code implementations16 Dec 2021 Xiangyu Peng, Kaige Xie, Amal Alabdulkarim, Harshith Kayam, Samihan Dani, Mark O. Riedl

In this paper, we introduce Story generation with Reader Models (StoRM), a framework in which a reader model is used to reason about the story should progress.

Story Generation

Inherently Explainable Reinforcement Learning in Natural Language

1 code implementation16 Dec 2021 Xiangyu Peng, Mark O. Riedl, Prithviraj Ammanabrolu

We focus on the task of creating a reinforcement learning agent that is inherently explainable -- with the ability to produce immediate local explanations by thinking out loud while performing a task and analyzing entire trajectories post-hoc to produce causal explanations.

Graph Attention reinforcement-learning +1

Detecting and Adapting to Novelty in Games

no code implementations4 Jun 2021 Xiangyu Peng, Jonathan C. Balloch, Mark O. Riedl

Open-world novelty occurs when the rules of an environment can change abruptly, such as when a game player encounters "house rules".

Knowledge Graphs Model-based Reinforcement Learning +2

Inferring the Reader: Guiding Automated Story Generation with Commonsense Reasoning

1 code implementation4 May 2021 Xiangyu Peng, Siyan Li, Sarah Wiegreffe, Mark Riedl

Transformer-based language model approaches to automated story generation currently provide state-of-the-art results.

Language Modelling Story Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.