1 code implementation • 25 May 2023 • Yachen Kang, Diyuan Shi, Jinxin Liu, Li He, Donglin Wang
Instead, the agent is provided with pre-existing offline trajectories and human preferences between pairs of trajectories to extract the dynamics and task information, respectively.
no code implementations • 23 May 2023 • Sibo Gai, Donglin Wang, Li He
In this paper, we formulate a new setting, continual offline reinforcement learning (CORL), where an agent learns a sequence of offline reinforcement learning tasks and pursues good performance on all learned tasks with a small replay buffer without exploring any of the environments of all the sequential tasks.
no code implementations • 27 Mar 2023 • Siteng Huang, Biao Gong, Yutong Feng, Yiliang Lv, Donglin Wang
Recent compositional zero-shot learning (CZSL) methods adapt pre-trained vision-language models (VLMs) by constructing trainable prompts only for composed state-object pairs.
no code implementations • 12 Mar 2023 • Min Zhang, Zifeng Zhuang, Zhitao Wang, Donglin Wang, Wenbin Li
OOD exacerbates inconsistencies in magnitudes and directions of task gradients, which brings challenges for GBML to optimize the meta-knowledge by minimizing the sum of task gradients in each minibatch.
1 code implementation • 22 Feb 2023 • Zifeng Zhuang, Kun Lei, Jinxin Liu, Donglin Wang, Yilang Guo
Offline reinforcement learning (RL) is a challenging setting where existing off-policy actor-critic methods perform poorly due to the overestimation of out-of-distribution state-action pairs.
no code implementations • 11 Jan 2023 • Ting Wang, Zongkai Wu, Feiyu Yao, Donglin Wang
First, we propose an Environment Representation Graph (ERG) through object detection to express the environment in semantic level.
1 code implementation • CVPR 2023 • Siteng Huang, Biao Gong, Yulin Pan, Jianwen Jiang, Yiliang Lv, Yuyuan Li, Donglin Wang
Many recent studies leverage the pre-trained CLIP for text-video cross-modal retrieval by tuning the backbone with additional heavy modules, which not only brings huge computational burdens with much more parameters, but also leads to the knowledge forgetting from upstream models.
no code implementations • 15 Sep 2022 • Ziqi Zhang, Yile Wang, Yue Zhang, Donglin Wang
Experimental results show that our RL pre-trained models can give close performance compared with the models using the LM training objective, showing that there exist common useful features across these two modalities.
no code implementations • 13 Sep 2022 • Feng Zhao, Ziqi Zhang, Donglin Wang
This is the first study that we are aware of that looks into dynamic KSG for skill retrieval and learning.
1 code implementation • 22 Aug 2022 • Siteng Huang, Qiyao Wei, Donglin Wang
Compositional zero-shot learning (CZSL) refers to recognizing unseen compositions of known visual primitives, which is an essential ability for artificial intelligence systems to learn and understand the world.
1 code implementation • 14 Jul 2022 • Min Zhang, Siteng Huang, Wenbin Li, Donglin Wang
To solve this problem, we present a plug-in Hierarchical Tree Structure-aware (HTS) method, which not only learns the relationship of FSL and pretext tasks, but more importantly, can adaptively select and aggregate feature representations generated by pretext tasks to maximize the performance of FSL tasks.
no code implementations • ICLR 2022 • Jinxin Liu, Hongyin Zhang, Donglin Wang
Specifically, DARA emphasizes learning from those source transition pairs that are adaptive for the target environment and mitigates the offline dynamics shift by characterizing state-action-next-state pairs instead of the typical state-action distribution sketched by prior offline RL methods.
no code implementations • 2 Mar 2022 • Qingfeng Yao, Jilong Wang, Shuyu Yang, Cong Wang, Hongyin Zhang, Qifeng Zhang, Donglin Wang
The deep learning model extracts key points during animal motion from videos.
no code implementations • 2 Mar 2022 • Qingfeng Yao, Jilong Wan, Shuyu Yang, Cong Wang, Linghan Meng, Qifeng Zhang, Donglin Wang
Due to their ability to adapt to different terrains, quadruped robots have drawn much attention in the research field of robot learning.
1 code implementation • 13 Nov 2021 • Xintao Xiang, Tiancheng Huang, Donglin Wang
In this paper, we propose Learning to Evolve on Dynamic Graphs (LEDG) - a novel algorithm that jointly learns graph information and time information.
no code implementations • 12 Nov 2021 • Donglin Wang, Qiuheng Zhou, Sanket Partani, Anjie Qiu, Hans D. Schotten
Nowadays mobile communication is growing fast in the 5G communication industry.
no code implementations • NeurIPS 2021 • Jinxin Liu, Hao Shen, Donglin Wang, Yachen Kang, Qiangxing Tian
Unsupervised reinforcement learning aims to acquire skills without prior goal representations, where an agent automatically explores an open-ended environment to represent goals and learn the goal-conditioned policy.
no code implementations • 21 Oct 2021 • Yachen Kang, Jinxin Liu, Xin Cao, Donglin Wang
To achieve this, the widely used GAN-inspired IRL method is adopted, and its discriminator, recognizing policy-generating trajectories, is modified with the quantification of dynamics difference.
no code implementations • 15 Oct 2021 • Ryan Jacobs, Mingren Shen, YuHan Liu, Wei Hao, Xiaoshan Li, Ruoyu He, Jacob RC Greaves, Donglin Wang, Zeming Xie, Zitong Huang, Chao Wang, Kevin G. Field, Dane Morgan
In this work, we perform semantic segmentation of multiple defect types in electron microscopy images of irradiated FeCrAl alloys using a deep learning Mask Regional Convolutional Neural Network (Mask R-CNN) model.
no code implementations • 29 Sep 2021 • Siteng Huang, Qiyao Wei, Donglin Wang
To narrow the considerable gap between artificial and human intelligence, we propose a new task, namely reference-limited compositional learning (RLCL), which reproduces three core challenges to mimic human perception: compositional learning, few-shot, and few referential compositions.
1 code implementation • 7 Jun 2021 • Xin Yang, Ning Zhang, Donglin Wang
Fourth, we generate three corresponding masks based on the 20 selected ROIs from group ICA, the 20 ROIs selected from dictionary learning, and the 40 combined ROIs selected from both.
no code implementations • 27 Apr 2021 • Shiqi Chen, Zhengyu Chen, Donglin Wang
Meta Reinforcement Learning (MRL) enables an agent to learn from a limited number of past trajectories and extrapolate to a new task.
no code implementations • CVPR 2021 • Zhengyu Chen, Jixie Ge, Heshen Zhan, Siteng Huang, Donglin Wang
While few-shot learning (FSL) aims for rapid generalization to new concepts with little supervision, self-supervised learning (SSL) constructs supervisory signals directly computed from unlabeled data.
no code implementations • 11 Apr 2021 • Jinxin Liu, Donglin Wang, Qiangxing Tian, Zhengyu Chen
It is of significance for an agent to learn a widely applicable and general-purpose policy that can achieve diverse goals including images and text descriptions.
no code implementations • 10 Dec 2020 • Ting Wang, Zongkai Wu, Donglin Wang
In the training phase, we first locate the generalization problem to the visual perception module, and then compare two meta-learning algorithms for better generalization in seen and unseen environments.
no code implementations • 10 Dec 2020 • Tiancheng Huang, Ke Xu, Donglin Wang
Domain adaptation using graph-structured networks learns label-discriminative and network-invariant node embeddings by sharing graph parameters.
1 code implementation • 10 Sep 2020 • Siteng Huang, Min Zhang, Yachen Kang, Donglin Wang
However, these approaches only augment the representations of samples with available semantics while ignoring the query set, which loses the potential for the improvement and may lead to a shift between the modalities combination and the pure-visual representation.
no code implementations • 25 Sep 2019 • Qiangxing Tian, Jinxin Liu, Donglin Wang
By maximizing an information theoretic objective, a few recent methods empower the agent to explore the environment and learn useful skills without supervision.