Search Results for author: Anji Liu

Found 28 papers, 14 papers with code

MineStudio: A Streamlined Package for Minecraft AI Agent Development

1 code implementation24 Dec 2024 Shaofei Cai, Zhancun Mu, Kaichen He, Bowei Zhang, Xinyue Zheng, Anji Liu, Yitao Liang

Minecraft has emerged as a valuable testbed for embodied intelligence and sequential decision-making research, yet the development and validation of novel agents remains hindered by significant engineering challenges.

AI Agent Decision Making +2

GROOT-2: Weakly Supervised Multi-Modal Instruction Following Agents

no code implementations7 Dec 2024 Shaofei Cai, Bowei Zhang, ZiHao Wang, Haowei Lin, Xiaojian Ma, Anji Liu, Yitao Liang

Developing agents that can follow multimodal instructions remains a fundamental challenge in robotics and AI.

Instruction Following

ROCKET-1: Mastering Open-World Interaction with Visual-Temporal Context Prompting

1 code implementation23 Oct 2024 Shaofei Cai, ZiHao Wang, Kewei Lian, Zhancun Mu, Xiaojian Ma, Anji Liu, Yitao Liang

Using this approach, we train ROCKET-1, a low-level policy that predicts actions based on concatenated visual observations and segmentation masks, supported by real-time object tracking from SAM-2.

Decision Making Minecraft +3

Discrete Copula Diffusion

no code implementations2 Oct 2024 Anji Liu, Oliver Broadrick, Mathias Niepert, Guy Van Den Broeck

When we apply this approach to autoregressive copula models, the combined model outperforms both models individually in unconditional and conditional text generation.

Conditional Text Generation Denoising

OmniJARVIS: Unified Vision-Language-Action Tokenization Enables Open-World Instruction Following Agents

no code implementations27 Jun 2024 ZiHao Wang, Shaofei Cai, Zhancun Mu, Haowei Lin, Ceyao Zhang, Xuejie Liu, Qing Li, Anji Liu, Xiaojian Ma, Yitao Liang

First, we introduce a self-supervised approach to learn a behavior encoder that produces discretized tokens for behavior trajectories $\tau = \{o_0, a_0, \dots\}$ and an imitation learning policy decoder conditioned on these tokens.

Decoder Imitation Learning +2

Efficient Adaptation in Mixed-Motive Environments via Hierarchical Opponent Modeling and Planning

no code implementations12 Jun 2024 Yizhe Huang, Anji Liu, Fanqi Kong, Yaodong Yang, Song-Chun Zhu, Xue Feng

To address these issues, we propose Hierarchical Opponent modeling and Planning (HOP), a novel multi-agent decision-making algorithm that enables few-shot adaptation to unseen policies in mixed-motive environments.

Decision Making Multi-agent Reinforcement Learning

Learning to Discretize Denoising Diffusion ODEs

1 code implementation24 May 2024 Vinh Tong, Trung-Dung Hoang, Anji Liu, Guy Van Den Broeck, Mathias Niepert

We achieve FIDs of 2. 38 (10 NFE), and 2. 27 (10 NFE) on unconditional CIFAR10 and AFHQv2 in 5-10 minutes of training.

Denoising Image Generation +1

Smart Help: Strategic Opponent Modeling for Proactive and Adaptive Robot Assistance in Households

no code implementations CVPR 2024 Zhihao Cao, Zidong Wang, Siwen Xie, Anji Liu, Lifeng Fan

Our findings illustrate the potential of AI-imbued assistive robots in improving the well-being of vulnerable groups.

RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generation

1 code implementation8 Mar 2024 ZiHao Wang, Anji Liu, Haowei Lin, Jiaqi Li, Xiaojian Ma, Yitao Liang

We explore how iterative revising a chain of thoughts with the help of information retrieval significantly improves large language models' reasoning and generation ability in long-horizon generation tasks, while hugely mitigating hallucination.

Code Generation Hallucination +4

Image Inpainting via Tractable Steering of Diffusion Models

1 code implementation28 Nov 2023 Anji Liu, Mathias Niepert, Guy Van Den Broeck

Further, with the help of an image encoder and decoder, our method can readily accept semantic constraints on specific regions of the image, which opens up the potential for more controlled image generation tasks.

Decoder Denoising +1

JARVIS-1: Open-World Multi-task Agents with Memory-Augmented Multimodal Language Models

1 code implementation10 Nov 2023 ZiHao Wang, Shaofei Cai, Anji Liu, Yonggang Jin, Jinbing Hou, Bowei Zhang, Haowei Lin, Zhaofeng He, Zilong Zheng, Yaodong Yang, Xiaojian Ma, Yitao Liang

Achieving human-like planning and control with multimodal observations in an open world is a key milestone for more functional generalist agents.

Minecraft

A Tractable Inference Perspective of Offline RL

no code implementations31 Oct 2023 Xuejie Liu, Anji Liu, Guy Van Den Broeck, Yitao Liang

A popular paradigm for offline Reinforcement Learning (RL) tasks is to first fit the offline trajectories to a sequence model, and then prompt the model for actions that lead to high expected return.

Offline RL Reinforcement Learning (RL)

GROOT: Learning to Follow Instructions by Watching Gameplay Videos

no code implementations12 Oct 2023 Shaofei Cai, Bowei Zhang, ZiHao Wang, Xiaojian Ma, Anji Liu, Yitao Liang

We propose to follow reference videos as instructions, which offer expressive goal specifications while eliminating the need for expensive text-gameplay annotations.

Decoder Instruction Following +1

Understanding the Distillation Process from Deep Generative Models to Tractable Probabilistic Circuits

no code implementations16 Feb 2023 Xuejie Liu, Anji Liu, Guy Van Den Broeck, Yitao Liang

In this paper, we theoretically and empirically discover that the performance of a PC can exceed that of its teacher model.

Open-World Multi-Task Control Through Goal-Aware Representation Learning and Adaptive Horizon Prediction

2 code implementations CVPR 2023 Shaofei Cai, ZiHao Wang, Xiaojian Ma, Anji Liu, Yitao Liang

We study the problem of learning goal-conditioned policies in Minecraft, a popular, widely accessible yet challenging open-ended environment for developing human-level multi-task agents.

Diversity Minecraft +2

Sparse Probabilistic Circuits via Pruning and Growing

1 code implementation22 Nov 2022 Meihua Dang, Anji Liu, Guy Van Den Broeck

The growing operation increases model capacity by increasing the size of the latent space.

Model Compression

Efficient Meta Reinforcement Learning for Preference-based Fast Adaptation

1 code implementation20 Nov 2022 Zhizhou Ren, Anji Liu, Yitao Liang, Jian Peng, Jianzhu Ma

To bridge this gap, we study the problem of few-shot adaptation in the context of human-in-the-loop reinforcement learning.

Meta Reinforcement Learning reinforcement-learning +2

Scaling Up Probabilistic Circuits by Latent Variable Distillation

no code implementations10 Oct 2022 Anji Liu, Honghua Zhang, Guy Van Den Broeck

We propose to overcome such bottleneck by latent variable distillation: we leverage the less tractable but more expressive deep generative models to provide extra supervision over the latent variables of PCs.

Language Modeling Language Modelling

Lossless Compression with Probabilistic Circuits

1 code implementation ICLR 2022 Anji Liu, Stephan Mandt, Guy Van Den Broeck

To overcome such problems, we establish a new class of tractable lossless compression models that permit efficient encoding and decoding: Probabilistic Circuits (PCs).

Data Compression Image Generation

Tractable Regularization of Probabilistic Circuits

no code implementations NeurIPS 2021 Anji Liu, Guy Van Den Broeck

Instead, we re-think regularization for PCs and propose two intuitive techniques, data softening and entropy regularization, that both take advantage of PCs' tractability and still have an efficient implementation as a computation graph.

Density Estimation

A Compositional Atlas of Tractable Circuit Operations for Probabilistic Inference

1 code implementation NeurIPS 2021 Antonio Vergari, YooJung Choi, Anji Liu, Stefano Teso, Guy Van Den Broeck

Circuit representations are becoming the lingua franca to express and reason about tractable generative and discriminative models.

On Effective Parallelization of Monte Carlo Tree Search

no code implementations15 Jun 2020 Anji Liu, Yitao Liang, Ji Liu, Guy Van Den Broeck, Jianshu Chen

Second, and more importantly, we demonstrate how the proposed necessary conditions can be adopted to design more effective parallel MCTS algorithms.

Atari Games

Off-Policy Deep Reinforcement Learning with Analogous Disentangled Exploration

1 code implementation25 Feb 2020 Anji Liu, Yitao Liang, Guy Van Den Broeck

Off-policy reinforcement learning (RL) is concerned with learning a rewarding policy by executing another policy that gathers samples of experience.

continuous-control Continuous Control +3

Watch the Unobserved: A Simple Approach to Parallelizing Monte Carlo Tree Search

4 code implementations ICLR 2020 Anji Liu, Jianshu Chen, Mingze Yu, Yu Zhai, Xuewen Zhou, Ji Liu

Monte Carlo Tree Search (MCTS) algorithms have achieved great success on many challenging benchmarks (e. g., Computer Go).

Cannot find the paper you are looking for? You can Submit a new open access paper.