Search Results for author: Qiang He

Found 19 papers, 7 papers with code

The Case for Controls: Identifying outbreak risk factors through case-control comparisons

no code implementations3 Nov 2023 Nina H. Fefferman, Michael J. Blum, Lydia Bourouiba, Nathaniel L. Gibson, Qiang He, Debra L. Miller, Monica Papes, Dana K. Pasquale, Connor Verheyen, Sadie J. Ryan

Investigations of infectious disease outbreaks often focus on identifying place- and context-dependent factors responsible for emergence and spread, resulting in phenomenological narratives ill-suited to developing generalizable predictive and preventive measures.

Decision Making

Self-supervised Noise2noise Method Utilizing Corrupted Images with a Modular Network for LDCT Denoising

1 code implementation13 Aug 2023 Yuting Zhu, Qiang He, YuDong Yao, Yueyang Teng

Note that we use LDCT images based on the noisy-as-clean strategy for corruption instead of NDCT images.

Image Denoising

Eigensubspace of Temporal-Difference Dynamics and How It Improves Value Approximation in Reinforcement Learning

no code implementations29 Jun 2023 Qiang He, Tianyi Zhou, Meng Fang, Setareh Maghsudi

In ERC, we propose a regularizer that guides the approximation error tending towards the 1-eigensubspace, resulting in a more efficient and stable path of value approximation.

Reinforcement Learning (RL)

PGPointNovo: an efficient neural network-based tool for parallel de novo peptide sequencing

1 code implementation Bioinformatics Advances 2023 Xiaofang Xu, Chunde Yang, Qiang He, Kunxian Shu, Yuan Xinpu, Zhiguang Chen, Yunping Zhu, Tao Chen

De novo peptide sequencing for tandem mass spectrometry data is not only a key technology for novel peptide identification, but also a precedent task for many downstream tasks, such as vaccine and antibody studies.

de novo peptide sequencing Efficient Neural Network

Quadratic Graph Attention Network (Q-GAT) for Robust Construction of Gene Regulatory Networks

1 code implementation24 Mar 2023 HUI ZHANG, Xuexin An, Qiang He, YuDong Yao, Yudong Zhang, Feng-Lei Fan, Yueyang Teng

The former informs that nonlinear aggregation of quadratic neurons can amplify useful signals and suppress unwanted noise, thereby facilitating robustness, while the latter reveals that Q-GAT can leverage more features in prediction thanks to the dual attention mechanism, which endows Q-GAT with the ability to confront adversarial perturbation.

Graph Attention

Centralized Cooperative Exploration Policy for Continuous Control Tasks

1 code implementation6 Jan 2023 Chao Li, Chen Gong, Qiang He, Xinwen Hou, Yu Liu

To explicitly encourage exploration in continuous control tasks, we propose CCEP (Centralized Cooperative Exploration Policy), which utilizes underestimation and overestimation of value functions to maintain the capacity of exploration.

Continuous Control

Adaptive Weighted Nonnegative Matrix Factorization for Robust Feature Representation

1 code implementation7 Jun 2022 Tingting Shen, Junhang Li, Can Tong, Qiang He, Chen Li, YuDong Yao, Yueyang Teng

Nonnegative matrix factorization (NMF) has been widely used to dimensionality reduction in machine learning.

Dimensionality Reduction

Frustratingly Easy Regularization on Representation Can Boost Deep Reinforcement Learning

no code implementations CVPR 2023 Qiang He, Huangyuan Su, Jieyu Zhang, Xinwen Hou

In this work, we demonstrate that the learned representation of the $Q$-network and its target $Q$-network should, in theory, satisfy a favorable distinguishable representation property.

Continuous Control reinforcement-learning +2

Improving the Level of Autism Discrimination through GraphRNN Link Prediction

no code implementations19 Feb 2022 Haonan Sun, Qiang He, Shouliang Qi, YuDong Yao, Yueyang Teng

This paper is based on the latter technique, which learns the edge distribution of real brain network through GraphRNN, and generates the synthetic data which has incentive effect on the discriminant model.

feature selection Link Prediction

An Entropy Weighted Nonnegative Matrix Factorization Algorithm for Feature Representation

1 code implementation27 Nov 2021 Jiao Wei, Can Tong, Bingxue Wu, Qiang He, Shouliang Qi, YuDong Yao, Yueyang Teng

Nonnegative matrix factorization (NMF) has been widely used to learn low-dimensional representations of data.

Attribute

The $f$-Divergence Reinforcement Learning Framework

no code implementations24 Sep 2021 Chen Gong, Qiang He, Yunpeng Bai, Zhou Yang, Xiaoyu Chen, Xinwen Hou, Xianjie Zhang, Yu Liu, Guoliang Fan

In FRL, the policy evaluation and policy improvement phases are simultaneously performed by minimizing the $f$-divergence between the learning policy and sampling policy, which is distinct from conventional DRL algorithms that aim to maximize the expected cumulative rewards.

Decision Making Mathematical Proofs +2

LDC-VAE: A Latent Distribution Consistency Approach to Variational AutoEncoders

no code implementations22 Sep 2021 Xiaoyu Chen, Chen Gong, Qiang He, Xinwen Hou, Yu Liu

Variational autoencoders (VAEs), as an important aspect of generative models, have received a lot of research interests and reached many successful applications.

Image Generation

MEPG: A Minimalist Ensemble Policy Gradient Framework for Deep Reinforcement Learning

no code implementations22 Sep 2021 Qiang He, Huangyuan Su, Chen Gong, Xinwen Hou

During the training of a reinforcement learning (RL) agent, the distribution of training data is non-stationary as the agent's behavior changes over time.

Gaussian Processes Q-Learning +2

Session-based Social and Dependency-aware Software Recommendation

no code implementations10 Mar 2021 Dengcheng Yan, Tianyi Tang, Wenxin Xie, Yiwen Zhang, Qiang He

With the increase of complexity of modern software, social collaborative coding and reuse of open source software packages become more and more popular, which thus greatly enhances the development efficiency and software quality.

Graph Attention Recommendation Systems

POPO: Pessimistic Offline Policy Optimization

1 code implementation26 Dec 2020 Qiang He, Xinwen Hou

Offline reinforcement learning (RL), also known as batch RL, aims to optimize policy from a large pre-recorded dataset without interaction with the environment.

Offline RL Q-Learning +1

WD3: Taming the Estimation Bias in Deep Reinforcement Learning

no code implementations18 Jun 2020 Qiang He, Xinwen Hou

To obtain a more precise estimation for value function, we unify these two opposites and propose a novel algorithm \underline{W}eighted \underline{D}elayed \underline{D}eep \underline{D}eterministic Policy Gradient (WD3), which can eliminate the estimation bias and further improve the performance by weighting a pair of critics.

Continuous Control OpenAI Gym +2

PA-Cache: Evolving Learning-Based Popularity-Aware Content Caching in Edge Networks

no code implementations20 Feb 2020 Qilin Fan, Xiuhua Li, Jian Li, Qiang He, Kai Wang, Junhao Wen

Compared to the conventional content delivery networks, caches in edge networks with smaller sizes usually have to accommodate more bursty requests.

Decision Making

Cannot find the paper you are looking for? You can Submit a new open access paper.