Search Results for author: Xueqian Wang

Found 45 papers, 17 papers with code

Heterogeneous Federated Learning with Splited Language Model

no code implementations24 Mar 2024 Yifan Shi, Yuhui Zhang, Ziyue Huang, Xiaofeng Yang, Li Shen, Wei Chen, Xueqian Wang

Federated Split Learning (FSL) is a promising distributed learning paradigm in practice, which gathers the strengths of both Federated Learning (FL) and Split Learning (SL) paradigms, to ensure model privacy while diminishing the resource overhead of each client, especially on large transformer models in a resource-constrained environment, e. g., Internet of Things (IoT).

Federated Learning Language Modelling

Global-guided Focal Neural Radiance Field for Large-scale Scene Rendering

no code implementations19 Mar 2024 Mingqi Shao, Feng Xiong, Hang Zhang, Shuang Yang, Mu Xu, Wei Bian, Xueqian Wang

The global stage obtains a continuous representation of the entire scene while the focal stage decomposes the scene into multiple blocks and further processes them with distinct sub-encoders.

GAgent: An Adaptive Rigid-Soft Gripping Agent with Vision Language Models for Complex Lighting Environments

no code implementations16 Mar 2024 Zhuowei Li, Miao Zhang, Xiaotian Lin, Meng Yin, Shuai Lu, Xueqian Wang

This paper introduces GAgent: an Gripping Agent designed for open-world environments that provides advanced cognitive abilities via VLM agents and flexible grasping abilities with variable stiffness soft grippers.

Language Modelling

PaddingFlow: Improving Normalizing Flows with Padding-Dimensional Noise

1 code implementation13 Mar 2024 Qinglong Meng, Chongkun Xia, Xueqian Wang

To implement PaddingFlow, only the dimension of normalizing flows needs to be modified.

 Ranked #1 on Density Estimation on MNIST (MMD-L2 metric)

Density Estimation

Offline Goal-Conditioned Reinforcement Learning for Safety-Critical Tasks with Recovery Policy

1 code implementation4 Mar 2024 Chenyang Cao, Zichen Yan, Renhao Lu, Junbo Tan, Xueqian Wang

Offline goal-conditioned reinforcement learning (GCRL) aims at solving goal-reaching tasks with sparse rewards from an offline dataset.

Step-On-Feet Tuning: Scaling Self-Alignment of LLMs via Bootstrapping

no code implementations12 Feb 2024 Haoyu Wang, Guozheng Ma, Ziqiao Meng, Zeyu Qin, Li Shen, Zhong Zhang, Bingzhe Wu, Liu Liu, Yatao Bian, Tingyang Xu, Xueqian Wang, Peilin Zhao

To further exploit the capabilities of bootstrapping, we investigate and adjust the training order of data, which yields improved performance of the model.

In-Context Learning

PPNet: A Two-Stage Neural Network for End-to-end Path Planning

1 code implementation18 Jan 2024 Qinglong Meng, Chongkun Xia, Xueqian Wang, Songping Mai, Bin Liang

The results show that PPNet can find a near-optimal solution in 15. 3ms, which is much shorter than the state-of-the-art path planners.

Replay-enhanced Continual Reinforcement Learning

no code implementations20 Nov 2023 Tiantian Zhang, Kevin Zehua Shen, Zichuan Lin, Bo Yuan, Xueqian Wang, Xiu Li, Deheng Ye

On the other hand, offline learning on replayed tasks while learning a new task may induce a distributional shift between the dataset and the learned policy on old tasks, resulting in forgetting.

Continual Learning reinforcement-learning

Learning visual-based deformable object rearrangement with local graph neural networks

1 code implementation16 Oct 2023 Yuhong Deng, Xueqian Wang, Lipeng Chen

Our method reaches much higher success rates on a variety of deformable rearrangement tasks (96. 3% on average) than state-of-the-art method in simulation experiments.

Multi-Task Learning

Revisiting Plasticity in Visual Reinforcement Learning: Data, Modules and Training Stages

1 code implementation11 Oct 2023 Guozheng Ma, Lu Li, Sen Zhang, Zixuan Liu, Zhen Wang, Yixin Chen, Li Shen, Xueqian Wang, DaCheng Tao

Plasticity, the ability of a neural network to evolve with new data, is crucial for high-performance and sample-efficient visual reinforcement learning (VRL).

Data Augmentation reinforcement-learning

DiffCPS: Diffusion Model based Constrained Policy Search for Offline Reinforcement Learning

1 code implementation9 Oct 2023 Longxiang He, Li Shen, Linrui Zhang, Junbo Tan, Xueqian Wang

Constrained policy search (CPS) is a fundamental problem in offline reinforcement learning, which is generally solved by advantage weighted regression (AWR).

D4RL Offline RL +1

Efficient Federated Prompt Tuning for Black-box Large Pre-trained Models

no code implementations4 Oct 2023 Zihao Lin, Yan Sun, Yifan Shi, Xueqian Wang, Lifu Huang, Li Shen, DaCheng Tao

With the blowout development of pre-trained models (PTMs), the efficient tuning of these models for diverse downstream applications has emerged as a pivotal research concern.

DFWLayer: Differentiable Frank-Wolfe Optimization Layer

1 code implementation21 Aug 2023 Zixuan Liu, Liu Liu, Xueqian Wang, Peilin Zhao

Differentiable optimization has received a significant amount of attention due to its foundational role in the domain of machine learning based on neural networks.

Towards More Suitable Personalization in Federated Learning via Decentralized Partial Model Training

no code implementations24 May 2023 Yifan Shi, Yingqi Liu, Yan Sun, Zihao Lin, Li Shen, Xueqian Wang, DaCheng Tao

Personalized federated learning (PFL) aims to produce the greatest personalized model for each client to face an insurmountable problem--data heterogeneity in real FL systems.

Personalized Federated Learning

Towards the Flatter Landscape and Better Generalization in Federated Learning under Client-level Differential Privacy

1 code implementation1 May 2023 Yifan Shi, Kang Wei, Li Shen, Yingqi Liu, Xueqian Wang, Bo Yuan, DaCheng Tao

To defend the inference attacks and mitigate the sensitive information leakages in Federated Learning (FL), client-level Differentially Private FL (DPFL) is the de-facto standard for privacy protection by clipping local updates and adding random noise.

Federated Learning

COMIC: An Unsupervised Change Detection Method for Heterogeneous Remote Sensing Images Based on Copula Mixtures and Cycle-Consistent Adversarial Networks

no code implementations3 Apr 2023 Chengxi Li, Gang Li, Zhuoyue Wang, Xueqian Wang, Pramod K. Varshney

For this problem, an unsupervised change detection method has been proposed recently based on the image translation technique of Cycle-Consistent Adversarial Networks (CycleGANs), where one image is translated from its original modality to the modality of the other image so that the difference map can be obtained by performing arithmetical subtraction.

Change Detection Translation

Make Landscape Flatter in Differentially Private Federated Learning

1 code implementation CVPR 2023 Yifan Shi, Yingqi Liu, Kang Wei, Li Shen, Xueqian Wang, DaCheng Tao

Specifically, DP-FedSAM integrates Sharpness Aware Minimization (SAM) optimizer to generate local flatness models with better stability and weight perturbation robustness, which results in the small norm of local updates and robustness to DP noise, thereby improving the performance.

Federated Learning

Graph-Transporter: A Graph-based Learning Method for Goal-Conditioned Deformable Object Rearranging Task

no code implementations21 Feb 2023 Yuhong Deng, Chongkun Xia, Xueqian Wang, Lipeng Chen

Rearranging deformable objects is a long-standing challenge in robotic manipulation for the high dimensionality of configuration space and the complex dynamics of deformable objects.

Object

Deep Reinforcement Learning Based on Local GNN for Goal-conditioned Deformable Object Rearranging

no code implementations21 Feb 2023 Yuhong Deng, Chongkun Xia, Xueqian Wang, Lipeng Chen

Some research has been attempting to design a general framework to obtain more advanced manipulation capabilities for deformable rearranging tasks, with lots of progress achieved in simulation.

Improving the Model Consistency of Decentralized Federated Learning

no code implementations8 Feb 2023 Yifan Shi, Li Shen, Kang Wei, Yan Sun, Bo Yuan, Xueqian Wang, DaCheng Tao

To mitigate the privacy leakages and communication burdens of Federated Learning (FL), decentralized FL (DFL) discards the central server and each client only communicates with its neighbors in a decentralized communication network.

Federated Learning

Foldsformer: Learning Sequential Multi-Step Cloth Manipulation With Space-Time Attention

1 code implementation8 Jan 2023 Kai Mo, Chongkun Xia, Xueqian Wang, Yuhong Deng, Xuehai Gao, Bin Liang

Foldformer can complete multi-step cloth manipulation tasks even when configurations of the cloth (e. g., size and pose) vary from configurations in the general demonstrations.

Pre-Trained Image Encoder for Generalizable Visual Reinforcement Learning

no code implementations17 Dec 2022 Zhecheng Yuan, Zhengrong Xue, Bo Yuan, Xueqian Wang, Yi Wu, Yang Gao, Huazhe Xu

Hence, we propose Pre-trained Image Encoder for Generalizable visual reinforcement learning (PIE-G), a simple yet effective framework that can generalize to the unseen visual scenarios in a zero-shot manner.

reinforcement-learning Reinforcement Learning (RL)

Evaluating Model-free Reinforcement Learning toward Safety-critical Tasks

no code implementations12 Dec 2022 Linrui Zhang, Qin Zhang, Li Shen, Bo Yuan, Xueqian Wang, DaCheng Tao

Despite a large number of reinforcement learning (RL) methods focusing on safety-critical tasks, there is still a lack of high-quality evaluation of those algorithms that adheres to safety constraints at each decision step under complex and unknown dynamics.

Autonomous Driving reinforcement-learning +2

Visual-tactile Fusion for Transparent Object Grasping in Complex Backgrounds

no code implementations30 Nov 2022 Shoujie Li, Haixin Yu, Wenbo Ding, Houde Liu, Linqi Ye, Chongkun Xia, Xueqian Wang, Xiao-Ping Zhang

Here, a visual-tactile fusion framework for transparent object grasping under complex backgrounds and variant light conditions is proposed, including the grasping position detection, tactile calibration, and visual-tactile fusion based classification.

Classification Position +1

MCTNet: A Multi-Scale CNN-Transformer Network for Change Detection in Optical Remote Sensing Images

no code implementations14 Oct 2022 Weiming Li, Lihui Xue, Xueqian Wang, Gang Li

For the task of change detection (CD) in remote sensing images, deep convolution neural networks (CNNs)-based methods have recently aggregated transformer modules to improve the capability of global feature extraction.

Change Detection

A Comprehensive Survey of Data Augmentation in Visual Reinforcement Learning

1 code implementation10 Oct 2022 Guozheng Ma, Zhen Wang, Zhecheng Yuan, Xueqian Wang, Bo Yuan, DaCheng Tao

Visual reinforcement learning (RL), which makes decisions directly from high-dimensional visual inputs, has demonstrated significant potential in various domains.

Data Augmentation reinforcement-learning +1

USEEK: Unsupervised SE(3)-Equivariant 3D Keypoints for Generalizable Manipulation

no code implementations28 Sep 2022 Zhengrong Xue, Zhecheng Yuan, Jiashun Wang, Xueqian Wang, Yang Gao, Huazhe Xu

Can a robot manipulate intra-category unseen objects in arbitrary poses with the help of a mere demonstration of grasping pose on a single object instance?

Keypoint Detection Object

Dynamics-Adaptive Continual Reinforcement Learning via Progressive Contextualization

no code implementations1 Sep 2022 Tiantian Zhang, Zichuan Lin, Yuxing Wang, Deheng Ye, Qiang Fu, Wei Yang, Xueqian Wang, Bin Liang, Bo Yuan, Xiu Li

A key challenge of continual reinforcement learning (CRL) in dynamic environments is to promptly adapt the RL agent's behavior as the environment changes over its lifetime, while minimizing the catastrophic forgetting of the learned information.

Bayesian Inference Knowledge Distillation +3

Polarimetric Inverse Rendering for Transparent Shapes Reconstruction

1 code implementation25 Aug 2022 Mingqi Shao, Chongkun Xia, Dongxu Duan, Xueqian Wang

We build a polarization dataset for multi-view transparent shapes reconstruction to verify our method.

Inverse Rendering Transparent objects

SafeRL-Kit: Evaluating Efficient Reinforcement Learning Methods for Safe Autonomous Driving

1 code implementation17 Jun 2022 Linrui Zhang, Qin Zhang, Li Shen, Bo Yuan, Xueqian Wang

Safe reinforcement learning (RL) has achieved significant success on risk-sensitive tasks and shown promise in autonomous driving (AD) as well.

Autonomous Driving reinforcement-learning +2

Penalized Proximal Policy Optimization for Safe Reinforcement Learning

no code implementations24 May 2022 Linrui Zhang, Li Shen, Long Yang, Shixiang Chen, Bo Yuan, Xueqian Wang, DaCheng Tao

Safe reinforcement learning aims to learn the optimal policy while satisfying safety constraints, which is essential in real-world applications.

reinforcement-learning Reinforcement Learning (RL) +1

Transparent Shape from a Single View Polarization Image

1 code implementation ICCV 2023 Mingqi Shao, Chongkun Xia, Zhendong Yang, Junnan Huang, Xueqian Wang

To train and test our method, we construct a dataset for transparent shape from polarization with paired polarization images and ground-truth normal maps.

Data-Driven Robust Control for Discrete Linear Time-Invariant Systems: A Descriptor System Approach

no code implementations14 Mar 2022 Jiabao He, Xuan Zhang, Feng Xu, Junbo Tan, Xueqian Wang

Given the recent surge of interest in data-driven control, this paper proposes a two-step method to study robust data-driven control for a parameter-unknown linear time-invariant (LTI) system that is affected by energy-bounded noises.

A Surrogate-Assisted Controller for Expensive Evolutionary Reinforcement Learning

no code implementations1 Jan 2022 Yuxing Wang, Tiantian Zhang, Yongzhe Chang, Bin Liang, Xueqian Wang, Bo Yuan

The integration of Reinforcement Learning (RL) and Evolutionary Algorithms (EAs) aims at simultaneously exploiting the sample efficiency as well as the diversity and robustness of the two paradigms.

Continuous Control Evolutionary Algorithms +3

Probability Density Estimation Based Imitation Learning

no code implementations13 Dec 2021 Yang Liu, Yongzhe Chang, Shilei Jiang, Xueqian Wang, Bin Liang, Bo Yuan

In general, IL methods can be categorized into Behavioral Cloning (BC) and Inverse Reinforcement Learning (IRL).

Density Estimation Imitation Learning

Data-Driven Controllability Analysis and Stabilization for Linear Descriptor Systems

no code implementations7 Dec 2021 Jiabao He, Xuan Zhang, Feng Xu, Junbo Tan, Xueqian Wang

For a parameter-unknown linear descriptor system, this paper proposes data-driven methods to testify the system's type and controllability and then to stabilize it.

Value Penalized Q-Learning for Recommender Systems

no code implementations15 Oct 2021 Chengqian Gao, Ke Xu, Kuangqi Zhou, Lanqing Li, Xueqian Wang, Bo Yuan, Peilin Zhao

To alleviate the action distribution shift problem in extracting RL policy from static trajectories, we propose Value Penalized Q-learning (VPQ), an uncertainty-based offline RL algorithm.

Offline RL Q-Learning +2

Catastrophic Interference in Reinforcement Learning: A Solution Based on Context Division and Knowledge Distillation

1 code implementation1 Sep 2021 Tiantian Zhang, Xueqian Wang, Bin Liang, Bo Yuan

In this paper, we present IQ, i. e., interference-aware deep Q-learning, to mitigate catastrophic interference in single-task deep reinforcement learning.

General Reinforcement Learning Knowledge Distillation +5

Cannot find the paper you are looking for? You can Submit a new open access paper.