Search Results for author: Qiang Liu

Found 257 papers, 91 papers with code

VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts

2 code implementations3 Nov 2021 Hangbo Bao, Wenhui Wang, Li Dong, Qiang Liu, Owais Khan Mohammed, Kriti Aggarwal, Subhojit Som, Furu Wei

We present a unified Vision-Language pretrained Model (VLMo) that jointly learns a dual encoder and a fusion encoder with a modular Transformer network.

Image Retrieval Retrieval +3

Deep Graph Contrastive Representation Learning

3 code implementations7 Jun 2020 Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, Liang Wang

Moreover, our unsupervised method even surpasses its supervised counterparts on transductive tasks, demonstrating its great potential in real-world applications.

Attribute Contrastive Learning +2

Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm

13 code implementations NeurIPS 2016 Qiang Liu, Dilin Wang

We propose a general purpose variational inference algorithm that forms a natural counterpart of gradient descent for optimization.

Bayesian Inference Variational Inference

A Convolutional Click Prediction Model

4 code implementations1 Jan 2015 Qiang Liu, Feng Yu, Shu Wu, Liang Wang

The explosion in online advertisement urges to better estimate the click prediction of ads.

Yi: Open Foundation Models by 01.AI

1 code implementation7 Mar 2024 01. AI, :, Alex Young, Bei Chen, Chao Li, Chengen Huang, Ge Zhang, Guanwei Zhang, Heng Li, Jiangcheng Zhu, Jianqun Chen, Jing Chang, Kaidong Yu, Peng Liu, Qiang Liu, Shawn Yue, Senbin Yang, Shiming Yang, Tao Yu, Wen Xie, Wenhao Huang, Xiaohui Hu, Xiaoyi Ren, Xinyao Niu, Pengcheng Nie, Yuchi Xu, Yudong Liu, Yue Wang, Yuxuan Cai, Zhenyu Gu, Zhiyuan Liu, Zonghong Dai

The Yi model family is based on 6B and 34B pretrained language models, then we extend them to chat models, 200K long context models, depth-upscaled models, and vision-language models.

Attribute Chatbot +2

Planning-oriented Autonomous Driving

1 code implementation CVPR 2023 Yihan Hu, Jiazhi Yang, Li Chen, Keyu Li, Chonghao Sima, Xizhou Zhu, Siqi Chai, Senyao Du, Tianwei Lin, Wenhai Wang, Lewei Lu, Xiaosong Jia, Qiang Liu, Jifeng Dai, Yu Qiao, Hongyang Li

Oriented at this, we revisit the key components within perception and prediction, and prioritize the tasks such that all these tasks contribute to planning.

Autonomous Driving Philosophy

Monkey: Image Resolution and Text Label Are Important Things for Large Multi-modal Models

1 code implementation11 Nov 2023 Zhang Li, Biao Yang, Qiang Liu, Zhiyin Ma, Shuo Zhang, Jingxu Yang, Yabo Sun, Yuliang Liu, Xiang Bai

Additionally, experiments on 18 datasets further demonstrate that Monkey surpasses existing LMMs in many tasks like Image Captioning and various Visual Question Answering formats.

Image Captioning Question Answering +2

InstaFlow: One Step is Enough for High-Quality Diffusion-Based Text-to-Image Generation

1 code implementation12 Sep 2023 Xingchao Liu, Xiwen Zhang, Jianzhu Ma, Jian Peng, Qiang Liu

Leveraging our new pipeline, we create, to the best of our knowledge, the first one-step diffusion-based text-to-image generator with SD-level image quality, achieving an FID (Frechet Inception Distance) of $23. 3$ on MS COCO 2017-5k, surpassing the previous state-of-the-art technique, progressive distillation, by a significant margin ($37. 2$ $\rightarrow$ $23. 3$ in FID).

Text-to-Image Generation

An Empirical Study of Graph Contrastive Learning

2 code implementations2 Sep 2021 Yanqiao Zhu, Yichen Xu, Qiang Liu, Shu Wu

We envision this work to provide useful empirical evidence of effective GCL algorithms and offer several insights for future research.

Graph Classification Management +1

Disentangled Self-Attentive Neural Networks for Click-Through Rate Prediction

2 code implementations11 Jan 2021 Yichen Xu, Yanqiao Zhu, Feng Yu, Qiang Liu, Shu Wu

To better model complex feature interaction, in this paper we propose a novel DisentanglEd Self-atTentIve NEtwork (DESTINE) framework for CTR prediction that explicitly decouples the computation of unary feature importance from pairwise interaction.

Click-Through Rate Prediction Computational Efficiency +1

Deep Interaction Machine: A Simple but Effective Model for High-order Feature Interactions

1 code implementation1 Jan 2020 Feng Yu, Zhaocheng Liu, Qiang Liu, Haoli Zhang, Shu Wu, Liang Wang

IM is an efficient and exact implementation of high-order FM, whose time complexity linearly grows with the order of interactions and the number of feature fields.

Click-Through Rate Prediction Feature Engineering

Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow

3 code implementations7 Sep 2022 Xingchao Liu, Chengyue Gong, Qiang Liu

The idea of rectified flow is to learn the ODE to follow the straight paths connecting the points drawn from \pi_0 and \pi_1 as much as possible.

Domain Adaptation Image-to-Image Translation +1

Rectified Flow: A Marginal Preserving Approach to Optimal Transport

1 code implementation29 Sep 2022 Qiang Liu

We present a flow-based approach to the optimal transport (OT) problem between two continuous distributions $\pi_0,\pi_1$ on $\mathbb{R}^d$, of minimizing a transport cost $\mathbb{E}[c(X_1-X_0)]$ in the set of couplings $(X_0, X_1)$ whose marginal distributions on $X_0, X_1$ equals $\pi_0,\pi_1$, respectively, where $c$ is a cost function.

valid

LLM+P: Empowering Large Language Models with Optimal Planning Proficiency

1 code implementation22 Apr 2023 Bo Liu, Yuqian Jiang, Xiaohan Zhang, Qiang Liu, Shiqi Zhang, Joydeep Biswas, Peter Stone

LLM+P takes in a natural language description of a planning problem, then returns a correct (or optimal) plan for solving that problem in natural language.

Zero-shot Generalization

Are we really making much progress? Revisiting, benchmarking, and refining heterogeneous graph neural networks

1 code implementation30 Dec 2021 Qingsong Lv, Ming Ding, Qiang Liu, Yuxiang Chen, Wenzheng Feng, Siming He, Chang Zhou, Jianguo Jiang, Yuxiao Dong, Jie Tang

Heterogeneous graph neural networks (HGNNs) have been blossoming in recent years, but the unique data processing and evaluation setups used by each work obstruct a full understanding of their advancements.

Benchmarking

FuseDream: Training-Free Text-to-Image Generation with Improved CLIP+GAN Space Optimization

1 code implementation2 Dec 2021 Xingchao Liu, Chengyue Gong, Lemeng Wu, Shujian Zhang, Hao Su, Qiang Liu

We approach text-to-image generation by combining the power of the retrained CLIP representation with an off-the-shelf image generator (GANs), optimizing in the latent space of GAN to find images that achieve maximum CLIP score with the given input text.

counterfactual Navigate +1

Conflict-Averse Gradient Descent for Multi-task Learning

3 code implementations NeurIPS 2021 Bo Liu, Xingchao Liu, Xiaojie Jin, Peter Stone, Qiang Liu

The goal of multi-task learning is to enable more efficient learning than single task learning by sharing model structures for a diverse set of tasks.

Multi-Task Learning

Graph Contrastive Learning with Adaptive Augmentation

1 code implementation27 Oct 2020 Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, Liang Wang

On the node attribute level, we corrupt node features by adding more noise to unimportant node features, to enforce the model to recognize underlying semantic information.

Attribute Contrastive Learning +3

Efficient and Accurate Quantized Image Super-Resolution on Mobile NPUs, Mobile AI & AIM 2022 challenge: Report

2 code implementations7 Nov 2022 Andrey Ignatov, Radu Timofte, Maurizio Denna, Abdel Younes, Ganzorig Gankhuyag, Jingang Huh, Myeong Kyun Kim, Kihwan Yoon, Hyeon-Cheol Moon, Seungho Lee, Yoonsik Choe, Jinwoo Jeong, Sungjei Kim, Maciej Smyl, Tomasz Latkowski, Pawel Kubik, Michal Sokolski, Yujie Ma, Jiahao Chao, Zhou Zhou, Hongfan Gao, Zhengfeng Yang, Zhenbing Zeng, Zhengyang Zhuge, Chenghua Li, Dan Zhu, Mengdi Sun, Ran Duan, Yan Gao, Lingshun Kong, Long Sun, Xiang Li, Xingdong Zhang, Jiawei Zhang, Yaqi Wu, Jinshan Pan, Gaocheng Yu, Jin Zhang, Feng Zhang, Zhe Ma, Hongbin Wang, Hojin Cho, Steve Kim, Huaen Li, Yanbo Ma, Ziwei Luo, Youwei Li, Lei Yu, Zhihong Wen, Qi Wu, Haoqiang Fan, Shuaicheng Liu, Lize Zhang, Zhikai Zong, Jeremy Kwon, Junxi Zhang, Mengyuan Li, Nianxiang Fu, Guanchen Ding, Han Zhu, Zhenzhong Chen, Gen Li, Yuanfan Zhang, Lei Sun, Dafeng Zhang, Neo Yang, Fitz Liu, Jerry Zhao, Mustafa Ayazoglu, Bahri Batuhan Bilecen, Shota Hirose, Kasidis Arunruangsirilert, Luo Ao, Ho Chun Leung, Andrew Wei, Jie Liu, Qiang Liu, Dahai Yu, Ao Li, Lei Luo, Ce Zhu, Seongmin Hong, Dongwon Park, Joonhee Lee, Byeong Hyun Lee, Seunggyu Lee, Se Young Chun, Ruiyuan He, Xuhao Jiang, Haihang Ruan, Xinjian Zhang, Jing Liu, Garas Gendy, Nabil Sabor, Jingchao Hou, Guanghui He

While numerous solutions have been proposed for this problem in the past, they are usually not compatible with low-power mobile NPUs having many computational and memory constraints.

Image Super-Resolution

AlphaNet: Improved Training of Supernets with Alpha-Divergence

2 code implementations16 Feb 2021 Dilin Wang, Chengyue Gong, Meng Li, Qiang Liu, Vikas Chandra

Weight-sharing NAS builds a supernet that assembles all the architectures as its sub-networks and jointly trains the supernet with the sub-networks.

Image Classification Neural Architecture Search

Vision Transformers with Patch Diversification

1 code implementation26 Apr 2021 Chengyue Gong, Dilin Wang, Meng Li, Vikas Chandra, Qiang Liu

To alleviate this problem, in this work, we introduce novel loss functions in vision transformer training to explicitly encourage diversity across patch representations for more discriminative feature extraction.

Image Classification Semantic Segmentation

NASViT: Neural Architecture Search for Efficient Vision Transformers with Gradient Conflict aware Supernet Training

1 code implementation ICLR 2022 Chengyue Gong, Dilin Wang, Meng Li, Xinlei Chen, Zhicheng Yan, Yuandong Tian, Qiang Liu, Vikas Chandra

In this work, we observe that the poor performance is due to a gradient conflict issue: the gradients of different sub-networks conflict with that of the supernet more severely in ViTs than CNNs, which leads to early saturation in training and inferior convergence.

Data Augmentation Image Classification +2

Dynamic Graph Neural Networks for Sequential Recommendation

1 code implementation15 Apr 2021 Mengqi Zhang, Shu Wu, Xueli Yu, Qiang Liu, Liang Wang

We propose a new method named Dynamic Graph Neural Network for Sequential Recommendation (DGSR), which connects different user sequences through a dynamic graph structure, exploring the interactive behavior of users and items with time and order information.

Graph Attention Link Prediction +1

VCNet and Functional Targeted Regularization For Learning Causal Effects of Continuous Treatments

1 code implementation14 Mar 2021 Lizhen Nie, Mao Ye, Qiang Liu, Dan Nicolae

Motivated by the rising abundance of observational data with continuous treatments, we investigate the problem of estimating the average dose-response curve (ADRF).

FAMO: Fast Adaptive Multitask Optimization

1 code implementation NeurIPS 2023 Bo Liu, Yihao Feng, Peter Stone, Qiang Liu

One of the grand enduring goals of AI is to create generalist agents that can learn multiple different tasks from diverse data via multitask learning (MTL).

Computational Efficiency

TAGNN: Target Attentive Graph Neural Networks for Session-based Recommendation

1 code implementation6 May 2020 Feng Yu, Yanqiao Zhu, Qiang Liu, Shu Wu, Liang Wang, Tieniu Tan

However, these methods compress a session into one fixed representation vector without considering the target items to be predicted.

Session-Based Recommendations

Sparsely Annotated Semantic Segmentation With Adaptive Gaussian Mixtures

1 code implementation CVPR 2023 Linshan Wu, Zhun Zhong, Leyuan Fang, Xingxin He, Qiang Liu, Jiayi Ma, Hao Chen

Our AGMM can effectively endow reliable supervision for unlabeled pixels based on the distributions of labeled and unlabeled pixels.

Contrastive Learning Semantic Segmentation

Improving Neural Language Modeling via Adversarial Training

1 code implementation10 Jun 2019 Dilin Wang, Chengyue Gong, Qiang Liu

Theoretically, we show that our adversarial mechanism effectively encourages the diversity of the embedding vectors, helping to increase the robustness of models.

Language Modelling Machine Translation +1

Evidence-aware Fake News Detection with Graph Neural Networks

1 code implementation18 Jan 2022 Weizhi Xu, Junfei Wu, Qiang Liu, Shu Wu, Liang Wang

In this paper, we focus on the evidence-based fake news detection, where several evidences are utilized to probe the veracity of news (i. e., a claim).

Fake News Detection Graph structure learning

Metric Residual Networks for Sample Efficient Goal-Conditioned Reinforcement Learning

2 code implementations17 Aug 2022 Bo Liu, Yihao Feng, Qiang Liu, Peter Stone

Furthermore, we introduce the metric residual network (MRN) that deliberately decomposes the action-value function Q(s, a, g) into the negated summation of a metric plus a residual asymmetric component.

reinforcement-learning Reinforcement Learning (RL)

Mining Latent Structures for Multimedia Recommendation

1 code implementation19 Apr 2021 Jinghao Zhang, Yanqiao Zhu, Qiang Liu, Shu Wu, Shuhui Wang, Liang Wang

To be specific, in the proposed LATTICE model, we devise a novel modality-aware structure learning layer, which learns item-item structures for each modality and aggregates multiple modalities to obtain latent item graphs.

Collaborative Filtering Multimedia recommendation +1

Action-depedent Control Variates for Policy Optimization via Stein's Identity

2 code implementations30 Oct 2017 Hao Liu, Yihao Feng, Yi Mao, Dengyong Zhou, Jian Peng, Qiang Liu

Policy gradient methods have achieved remarkable successes in solving challenging reinforcement learning problems.

Policy Gradient Methods reinforcement-learning +1

Statistical Adaptive Stochastic Gradient Methods

1 code implementation25 Feb 2020 Pengchuan Zhang, Hunter Lang, Qiang Liu, Lin Xiao

We propose a statistical adaptive procedure called SALSA for automatically scheduling the learning rate (step size) in stochastic gradient methods.

Scheduling

Firefly Neural Architecture Descent: a General Approach for Growing Neural Networks

1 code implementation NeurIPS 2020 Lemeng Wu, Bo Liu, Peter Stone, Qiang Liu

We propose firefly neural architecture descent, a general framework for progressively and dynamically growing neural networks to jointly optimize the networks' parameters and architectures.

Continual Learning Image Classification +1

A Langevin-like Sampler for Discrete Distributions

1 code implementation20 Jun 2022 Ruqi Zhang, Xingchao Liu, Qiang Liu

We propose discrete Langevin proposal (DLP), a simple and scalable gradient-based proposal for sampling complex high-dimensional discrete distributions.

Efficient Exploration Text Generation

Any equation is a forest: Symbolic genetic algorithm for discovering open-form partial differential equations (SGA-PDE)

2 code implementations9 Jun 2021 Yuntian Chen, Yingtao Luo, Qiang Liu, Hao Xu, Dongxiao Zhang

Partial differential equations (PDEs) are concise and understandable representations of domain knowledge, which are essential for deepening our understanding of physical processes and predicting future responses.

A Computational Framework for Solving Wasserstein Lagrangian Flows

1 code implementation16 Oct 2023 Kirill Neklyudov, Rob Brekelmans, Alexander Tong, Lazar Atanackovic, Qiang Liu, Alireza Makhzani

The dynamical formulation of the optimal transport can be extended through various choices of the underlying geometry ($\textit{kinetic energy}$), and the regularization of density paths ($\textit{potential energy}$).

Learning to Draw Samples: With Application to Amortized MLE for Generative Adversarial Learning

1 code implementation6 Nov 2016 Dilin Wang, Qiang Liu

We propose a simple algorithm to train stochastic neural networks to draw samples from given target distributions for probabilistic inference.

Ranked #19 on Conditional Image Generation on CIFAR-10 (Inception score metric)

Conditional Image Generation

Certified Monotonic Neural Networks

1 code implementation NeurIPS 2020 Xingchao Liu, Xing Han, Na Zhang, Qiang Liu

In this work, we propose to certify the monotonicity of the general piece-wise linear neural networks by solving a mixed integer linear programming problem. This provides a new general approach for learning monotonic neural networks with arbitrary model structures.

Fairness

Exploration via Hindsight Goal Generation

1 code implementation NeurIPS 2019 Zhizhou Ren, Kefan Dong, Yuan Zhou, Qiang Liu, Jian Peng

Goal-oriented reinforcement learning has recently been a practical framework for robotic manipulation tasks, in which an agent is required to reach a certain goal defined by a function on the state space.

reinforcement-learning Reinforcement Learning (RL)

Fast Point Cloud Generation with Straight Flows

1 code implementation CVPR 2023 Lemeng Wu, Dilin Wang, Chengyue Gong, Xingchao Liu, Yunyang Xiong, Rakesh Ranjan, Raghuraman Krishnamoorthi, Vikas Chandra, Qiang Liu

We perform evaluations on multiple 3D tasks and find that our PSF performs comparably to the standard diffusion model, outperforming other efficient 3D point cloud generation methods.

Point Cloud Completion

SAFER: A Structure-free Approach for Certified Robustness to Adversarial Word Substitutions

1 code implementation ACL 2020 Mao Ye, Chengyue Gong, Qiang Liu

For security reasons, it is of critical importance to develop models with certified robustness that can provably guarantee that the prediction is can not be altered by any possible synonymous word substitution.

text-classification Text Classification

Latent Structure Mining with Contrastive Modality Fusion for Multimedia Recommendation

1 code implementation1 Nov 2021 Jinghao Zhang, Yanqiao Zhu, Qiang Liu, Mengqi Zhang, Shu Wu, Liang Wang

Although having access to multiple modalities might allow us to capture rich information, we argue that the simple coarse-grained fusion by linear combination or concatenation in previous work is insufficient to fully understand content information and item relationships. To this end, we propose a latent structure MIning with ContRastive mOdality fusion method (MICRO for brevity).

Collaborative Filtering Multimedia recommendation

Good Subnetworks Provably Exist: Pruning via Greedy Forward Selection

1 code implementation3 Mar 2020 Mao Ye, Chengyue Gong, Lizhen Nie, Denny Zhou, Adam Klivans, Qiang Liu

This differs from the existing methods based on backward elimination, which remove redundant neurons from the large network.

Network Pruning

Greedy Optimization Provably Wins the Lottery: Logarithmic Number of Winning Tickets is Enough

1 code implementation NeurIPS 2020 Mao Ye, Lemeng Wu, Qiang Liu

Despite the great success of deep learning, recent works show that large deep neural networks are often highly redundant and can be significantly reduced in size.

Uncertainty-aware Surrogate Models for Airfoil Flow Simulations with Denoising Diffusion Probabilistic Models

1 code implementation8 Dec 2023 Qiang Liu, Nils Thuerey

Leveraging neural networks as surrogate models for turbulence simulation is a topic of growing interest.

Denoising

Variational Inference with Tail-adaptive f-Divergence

1 code implementation NeurIPS 2018 Dilin Wang, Hao liu, Qiang Liu

Variational inference with {\alpha}-divergences has been widely used in modern probabilistic machine learning.

Variational Inference

Coach-Player Multi-Agent Reinforcement Learning for Dynamic Team Composition

1 code implementation18 May 2021 Bo Liu, Qiang Liu, Peter Stone, Animesh Garg, Yuke Zhu, Animashree Anandkumar

Specifically, we 1) adopt the attention mechanism for both the coach and the players; 2) propose a variational objective to regularize learning; and 3) design an adaptive communication method to let the coach decide when to communicate with the players.

Multi-agent Reinforcement Learning reinforcement-learning +3

Learning Belief Representations for Imitation Learning in POMDPs

1 code implementation22 Jun 2019 Tanmay Gangwani, Joel Lehman, Qiang Liu, Jian Peng

We consider the problem of imitation learning from expert demonstrations in partially observable Markov decision processes (POMDPs).

Continuous Control Imitation Learning +1

Profiling Pareto Front With Multi-Objective Stein Variational Gradient Descent

1 code implementation NeurIPS 2021 Xingchao Liu, Xin Tong, Qiang Liu

Finding diverse and representative Pareto solutions from the Pareto front is a key challenge in multi-objective optimization (MOO).

Profiling Pareto Front With Multi-Objective Stein Variational Gradient Descent

1 code implementation NeurIPS 2021 Xingchao Liu, Xin Tong, Qiang Liu

Finding diverse and representative Pareto solutions from the Pareto front is a key challenge in multi-objective optimization (MOO).

Energy-Aware Neural Architecture Optimization with Fast Splitting Steepest Descent

1 code implementation ICLR 2020 Dilin Wang, Meng Li, Lemeng Wu, Vikas Chandra, Qiang Liu

Designing energy-efficient networks is of critical importance for enabling state-of-the-art deep learning in mobile and edge settings where the computation and energy budgets are highly limited.

Breaking the Curse of Horizon: Infinite-Horizon Off-Policy Estimation

2 code implementations NeurIPS 2018 Qiang Liu, Lihong Li, Ziyang Tang, Dengyong Zhou

We consider the off-policy estimation problem of estimating the expected reward of a target policy using samples collected by a different behavior policy.

Splitting Steepest Descent for Growing Neural Architectures

1 code implementation NeurIPS 2019 Qiang Liu, Lemeng Wu, Dilin Wang

We develop a progressive training approach for neural networks which adaptively grows the network structure by splitting existing neurons to multiple off-springs.

Adversarial Contrastive Learning for Evidence-aware Fake News Detection with Graph Neural Networks

1 code implementation11 Oct 2022 Junfei Wu, Weizhi Xu, Qiang Liu, Shu Wu, Liang Wang

Comprehensive experiments have demonstrated the superiority of GETRAL over the state-of-the-arts and validated the efficacy of semantic mining with graph structure and contrastive learning.

Contrastive Learning Fake News Detection +2

Continual Learning and Private Unlearning

1 code implementation24 Mar 2022 Bo Liu, Qiang Liu, Peter Stone

As intelligent agents become autonomous over longer periods of time, they may eventually become lifelong counterparts to specific people.

Continual Learning

Stein Variational Gradient Descent With Matrix-Valued Kernels

1 code implementation NeurIPS 2019 Dilin Wang, Ziyang Tang, Chandrajit Bajaj, Qiang Liu

Stein variational gradient descent (SVGD) is a particle-based inference algorithm that leverages gradient information for efficient approximate inference.

Bayesian Inference

Logical Closed Loop: Uncovering Object Hallucinations in Large Vision-Language Models

1 code implementation18 Feb 2024 Junfei Wu, Qiang Liu, Ding Wang, Jinghao Zhang, Shu Wu, Liang Wang, Tieniu Tan

In this work, we adopt the intuition that the LVLM tends to respond logically consistently for existent objects but inconsistently for hallucinated objects.

Hallucination Object

Relation-aware Heterogeneous Graph for User Profiling

1 code implementation14 Oct 2021 Qilong Yan, Yufeng Zhang, Qiang Liu, Shu Wu, Liang Wang

User profiling has long been an important problem that investigates user interests in many real applications.

Node Classification Relation

Atlas: Automate Online Service Configuration in Network Slicing

1 code implementation30 Oct 2022 Qiang Liu, Nakjung Choi, Tao Han

First, we design a learning-based simulator to reduce the sim-to-real discrepancy, which is accomplished by a new parameter searching method based on Bayesian optimization.

Bayesian Optimization Safe Exploration +1

Sequence Modeling of Temporal Credit Assignment for Episodic Reinforcement Learning

1 code implementation31 May 2019 Yang Liu, Yunan Luo, Yuanyi Zhong, Xi Chen, Qiang Liu, Jian Peng

Recent advances in deep reinforcement learning algorithms have shown great potential and success for solving many challenging real-world problems, including Go game and robotic applications.

reinforcement-learning Reinforcement Learning (RL)

Split Localized Conformal Prediction

1 code implementation27 Jun 2022 Xing Han, Ziyang Tang, Joydeep Ghosh, Qiang Liu

The modified score inherits the spirit of split conformal methods, which is simple and efficient and can scale to high dimensional settings.

Conformal Prediction Density Estimation +2

Network Pruning via Feature Shift Minimization

1 code implementation6 Jul 2022 Yuanzhi Duan, Yue Zhou, Peng He, Qiang Liu, Shukai Duan, Xiaofang Hu

In this paper, we propose a novel Feature Shift Minimization (FSM) method to compress CNN models, which evaluates the feature shift by converging the information of both features and filters.

Network Pruning

EX-FEVER: A Dataset for Multi-hop Explainable Fact Verification

1 code implementation15 Oct 2023 Huanhuan Ma, Weizhi Xu, Yifan Wei, Liuji Chen, Qiang Liu, Shu Wu, Liang Wang

Each instance is accompanied by a veracity label and an explanation that outlines the reasoning path supporting the veracity classification.

Claim Verification Explanation Generation +3

Debiasing Convolutional Neural Networks via Meta Orthogonalization

1 code implementation15 Nov 2020 Kurtis Evan David, Qiang Liu, Ruth Fong

While deep learning models often achieve strong task performance, their successes are hampered by their inability to disentangle spurious correlations from causative factors, such as when they use protected attributes (e. g., race, gender, etc.)

Word Embeddings

Sampling in Constrained Domains with Orthogonal-Space Variational Gradient Descent

1 code implementation12 Oct 2022 Ruqi Zhang, Qiang Liu, Xin T. Tong

Sampling methods, as important inference and learning techniques, are typically designed for unconstrained domains.

Fairness

MaxUp: A Simple Way to Improve Generalization of Neural Network Training

1 code implementation20 Feb 2020 Chengyue Gong, Tongzheng Ren, Mao Ye, Qiang Liu

The idea is to generate a set of augmented data with some random perturbations or transforms and minimize the maximum, or worst case loss over the augmented data.

Few-Shot Image Classification General Classification +1

Sampling with Mollified Interaction Energy Descent

2 code implementations24 Oct 2022 Lingxiao Li, Qiang Liu, Anna Korba, Mikhail Yurochkin, Justin Solomon

These energies rely on mollifier functions -- smooth approximations of the Dirac delta originated from PDE theory.

AdaFlow: Imitation Learning with Variance-Adaptive Flow-Based Policies

1 code implementation6 Feb 2024 Xixi Hu, Bo Liu, Xingchao Liu, Qiang Liu

To address this challenge, we propose AdaFlow, an imitation learning framework based on flow-based generative modeling.

Imitation Learning

Text-Guided Molecule Generation with Diffusion Language Model

1 code implementation20 Feb 2024 Haisong Gong, Qiang Liu, Shu Wu, Liang Wang

In this work, we propose the Text-Guided Molecule Generation with Diffusion Language Model (TGM-DLM), a novel approach that leverages diffusion models to address the limitations of autoregressive methods.

Language Modelling Text-based de novo Molecule Generation +1

Network Compression via Central Filter

1 code implementation10 Dec 2021 Yuanzhi Duan, Xiaofang Hu, Yue Zhou, Qiang Liu, Shukai Duan

In this paper, by exploring the similarities between feature maps, we propose a novel filter pruning method, Central Filter (CF), which suggests that a filter is approximately equal to a set of other filters after appropriate adjustments.

Network Pruning

Heterogeneous Graph Reasoning for Fact Checking over Texts and Tables

1 code implementation20 Feb 2024 Haisong Gong, Weizhi Xu, Shu Wu, Qiang Liu, Liang Wang

To address this, we propose a novel word-level Heterogeneous-graph-based model for Fact Checking over unstructured and structured information, namely HeterFC.

Fact Checking Language Modelling +1

Attributed Multi-Relational Attention Network for Fact-checking URL Recommendation

1 code implementation7 Jan 2020 Di You, Nguyen Vo, Kyumin Lee, Qiang Liu

To combat fake news, researchers mostly focused on detecting fake news and journalists built and maintained fact-checking sites (e. g., Snopes. com and Politifact. com).

Fact Checking Graph Attention +1

Stein Self-Repulsive Dynamics: Benefits From Past Samples

1 code implementation NeurIPS 2020 Mao Ye, Tongzheng Ren, Qiang Liu

Our idea is to introduce Stein variational gradient as a repulsive force to push the samples of Langevin dynamics away from the past trajectories.

Sampling with Trusthworthy Constraints: A Variational Gradient Framework

1 code implementation NeurIPS 2021 Xingchao Liu, Xin Tong, Qiang Liu

In this work, we propose a family of constrained sampling algorithms which generalize Langevin Dynamics (LD) and Stein Variational Gradient Descent (SVGD) to incorporate a moment constraint specified by a general nonlinear function.

Bayesian Inference Fairness

Sampling with Trusthworthy Constraints: A Variational Gradient Framework

1 code implementation NeurIPS 2021 Xingchao Liu, Xin Tong, Qiang Liu

In this work, we propose a family of constrained sampling algorithms which generalize Langevin Dynamics (LD) and Stein Variational Gradient Descent (SVGD) to incorporate a moment constraint specified by a general nonlinear function.

Bayesian Inference Fairness

Deep Stable Representation Learning on Electronic Health Records

1 code implementation3 Sep 2022 Yingtao Luo, Zhaocheng Liu, Qiang Liu

The unstable correlation between procedures and diagnoses existed in the training distribution can cause spurious correlation between historical EHR and future diagnosis.

Disease Prediction Representation Learning

Stein Variational Message Passing for Continuous Graphical Models

no code implementations ICML 2018 Dilin Wang, Zhe Zeng, Qiang Liu

We propose a novel distributed inference algorithm for continuous graphical models, by extending Stein variational gradient descent (SVGD) to leverage the Markov dependency structure of the distribution of interest.

Stein Variational Gradient Descent Without Gradient

no code implementations ICML 2018 Jun Han, Qiang Liu

Stein variational gradient decent (SVGD) has been shown to be a powerful approximate inference algorithm for complex distributions.

Learning Self-Imitating Diverse Policies

no code implementations ICLR 2019 Tanmay Gangwani, Qiang Liu, Jian Peng

Improving the efficiency of RL algorithms in real-world problems with sparse or episodic rewards is therefore a pressing need.

Continuous Control Imitation Learning +2

Stochastic Variance Reduction for Policy Gradient Estimation

no code implementations17 Oct 2017 Tianbing Xu, Qiang Liu, Jian Peng

Recent advances in policy gradient methods and deep learning have demonstrated their applicability for complex reinforcement learning problems.

Continuous Control Policy Gradient Methods +2

Learning to Explore with Meta-Policy Gradient

no code implementations13 Mar 2018 Tianbing Xu, Qiang Liu, Liang Zhao, Jian Peng

The performance of off-policy learning, including deep Q-learning and deep deterministic policy gradient (DDPG), critically depends on the choice of the exploration policy.

Q-Learning Reinforcement Learning (RL)

BEBP: An Poisoning Method Against Machine Learning Based IDSs

no code implementations11 Mar 2018 Pan Li, Qiang Liu, Wentao Zhao, Dongxu Wang, Siqi Wang

In this paper, we adopt the Edge Pattern Detection (EPD) algorithm to design a novel poisoning method that attack against several machine learning algorithms used in IDSs.

BIG-bench Machine Learning Intrusion Detection

On the Discrimination-Generalization Tradeoff in GANs

no code implementations ICLR 2018 Pengchuan Zhang, Qiang Liu, Dengyong Zhou, Tao Xu, Xiaodong He

When evaluated with neural distance, our bounds show that generalization is guaranteed as long as the discriminator set is small enough, regardless of the size of the generator or hypothesis set.

Generalization Bounds

Adaptive Scan Gibbs Sampler for Large Scale Inference Problems

no code implementations27 Jan 2018 Vadim Smolyakov, Qiang Liu, John W. Fisher III

For large scale on-line inference problems the update strategy is critical for performance.

Stein Variational Gradient Descent as Gradient Flow

no code implementations NeurIPS 2017 Qiang Liu

Stein variational gradient descent (SVGD) is a deterministic sampling algorithm that iteratively transports a set of particles to approximate given distributions, based on an efficient gradient-based update that guarantees to optimally decrease the KL divergence within a function space.

Two Methods For Wild Variational Inference

no code implementations30 Nov 2016 Qiang Liu, Yihao Feng

Variational inference provides a powerful tool for approximate probabilistic in- ference on complex, structured models.

Variational Inference Vocal Bursts Valence Prediction

Learning to Draw Samples with Amortized Stein Variational Gradient Descent

no code implementations20 Jul 2017 Yihao Feng, Dilin Wang, Qiang Liu

We propose a simple algorithm to train stochastic neural networks to draw samples from given target distributions for probabilistic inference.

Bayesian Inference

Efficient Localized Inference for Large Graphical Models

no code implementations28 Oct 2017 Jinglin Chen, Jian Peng, Qiang Liu

We propose a new localized inference algorithm for answering marginalization queries in large graphical models with the correlation decay property.

Learning Infinite RBMs with Frank-Wolfe

no code implementations NeurIPS 2016 Wei Ping, Qiang Liu, Alexander Ihler

In this work, we propose an infinite restricted Boltzmann machine~(RBM), whose maximum likelihood estimation~(MLE) corresponds to a constrained convex optimization.

Energy-efficient Amortized Inference with Cascaded Deep Classifiers

no code implementations10 Oct 2017 Jiaqi Guan, Yang Liu, Qiang Liu, Jian Peng

Deep neural networks have been remarkable successful in various AI tasks but often cast high computation and energy cost for energy-constrained applications such as mobile sensing.

Image Classification

Stein Variational Adaptive Importance Sampling

no code implementations18 Apr 2017 Jun Han, Qiang Liu

We propose a novel adaptive importance sampling algorithm which incorporates Stein variational gradient decent algorithm (SVGD) with importance sampling (IS).

Learning Deep Energy Models: Contrastive Divergence vs. Amortized MLE

no code implementations4 Jul 2017 Qiang Liu, Dilin Wang

We propose a number of new algorithms for learning deep energy models and demonstrate their properties.

Approximate Inference with Amortised MCMC

no code implementations27 Feb 2017 Yingzhen Li, Richard E. Turner, Qiang Liu

We propose a novel approximate inference algorithm that approximates a target distribution by amortising the dynamics of a user-selected MCMC sampler.

Stein Variational Policy Gradient

no code implementations7 Apr 2017 Yang Liu, Prajit Ramachandran, Qiang Liu, Jian Peng

Policy gradient methods have been successfully applied to many complex reinforcement learning problems.

Bayesian Inference Continuous Control +3

Bootstrap Model Aggregation for Distributed Statistical Learning

no code implementations NeurIPS 2016 Jun Han, Qiang Liu

In distributed, or privacy-preserving learning, we are often given a set of probabilistic models estimated from different local repositories, and asked to combine them into a single model that gives efficient statistical estimation.

Privacy Preserving

ICE: Information Credibility Evaluation on Social Media via Representation Learning

no code implementations29 Sep 2016 Qiang Liu, Shu Wu, Feng Yu, Liang Wang, Tieniu Tan

In this paper, we propose a novel representation learning method, Information Credibility Evaluation (ICE), to learn representations of information credibility on social media.

Feature Engineering Representation Learning

Black-box Importance Sampling

no code implementations17 Oct 2016 Qiang Liu, Jason D. Lee

Importance sampling is widely used in machine learning and statistics, but its power is limited by the restriction of using simple proposals for which the importance weights can be tractably calculated.

BIG-bench Machine Learning

Local Perturb-and-MAP for Structured Prediction

no code implementations24 May 2016 Gedas Bertasius, Qiang Liu, Lorenzo Torresani, Jianbo Shi

In this work, we present a new Local Perturb-and-MAP (locPMAP) framework that replaces the global optimization with a local optimization by exploiting our observed connection between locPMAP and the pseudolikelihood of the original CRF model.

Combinatorial Optimization Structured Prediction

Context-aware Sequential Recommendation

no code implementations19 Sep 2016 Qiang Liu, Shu Wu, Diyi Wang, Zhaokang Li, Liang Wang

Recently, Recurrent Neural Networks (RNN) based methods have been successfully applied in several sequential modeling tasks.

Sequential Recommendation

A Kernelized Stein Discrepancy for Goodness-of-fit Tests and Model Evaluation

no code implementations10 Feb 2016 Qiang Liu, Jason D. Lee, Michael. I. Jordan

We derive a new discrepancy statistic for measuring differences between two probability distributions based on combining Stein's identity with the reproducing kernel Hilbert space theory.

Decomposition Bounds for Marginal MAP

no code implementations NeurIPS 2015 Wei Ping, Qiang Liu, Alexander Ihler

Marginal MAP inference involves making MAP predictions in systems defined with latent variables or missing information.

Communication-efficient sparse regression: a one-shot approach

no code implementations14 Mar 2015 Jason D. Lee, Yuekai Sun, Qiang Liu, Jonathan E. Taylor

We devise a one-shot approach to distributed sparse regression in the high-dimensional setting.

regression

Distributed Estimation, Information Loss and Exponential Families

no code implementations NeurIPS 2014 Qiang Liu, Alexander Ihler

Distributed learning of probabilistic models from multiple data repositories with minimum communication is increasingly important.

Marginal Structured SVM with Hidden Variables

no code implementations4 Sep 2014 Wei Ping, Qiang Liu, Alexander Ihler

In this work, we propose the marginal structured SVM (MSSVM) for structured prediction with hidden variables.

Structured Prediction

Variational Algorithms for Marginal MAP

no code implementations26 Feb 2013 Qiang Liu, Alexander Ihler

The marginal maximum a posteriori probability (MAP) estimation problem, which calculates the mode of the marginal posterior distribution of a subset of variables with the remaining variables marginalized, is an important inference problem in many models, such as those with hidden variables or uncertain parameters.

Regularization Matters: Generalization and Optimization of Neural Nets v.s. their Induced Kernel

no code implementations NeurIPS 2019 Colin Wei, Jason D. Lee, Qiang Liu, Tengyu Ma

We prove that for infinite-width two-layer nets, noisy gradient descent optimizes the regularized neural net loss to a global minimum in polynomial iterations.

Stein Variational Gradient Descent as Moment Matching

no code implementations NeurIPS 2018 Qiang Liu, Dilin Wang

Stein variational gradient descent (SVGD) is a non-parametric inference algorithm that evolves a set of particles to fit a given distribution of interest.

Probabilistic Variational Bounds for Graphical Models

no code implementations NeurIPS 2015 Qiang Liu, John W. Fisher III, Alexander T. Ihler

We propose a simple Monte Carlo based inference method that augments convex variational bounds by adding importance sampling (IS).

Scoring Workers in Crowdsourcing: How Many Control Questions are Enough?

no code implementations NeurIPS 2013 Qiang Liu, Alexander T. Ihler, Mark Steyvers

We study the problem of estimating continuous quantities, such as prices, probabilities, and point spreads, using a crowdsourcing approach.

Variational Planning for Graph-based MDPs

no code implementations NeurIPS 2013 Qiang Cheng, Qiang Liu, Feng Chen, Alexander T. Ihler

The KL divergence is optimized using the belief propagation algorithm, with complexity exponential in only the cluster size of the graph.

Decision Making

Goodness-of-fit Testing for Discrete Distributions via Stein Discrepancy

no code implementations ICML 2018 Jiasen Yang, Qiang Liu, Vinayak Rao, Jennifer Neville

Recent work has combined Stein’s method with reproducing kernel Hilbert space theory to develop nonparametric goodness-of-fit tests for un-normalized probability distributions.

Learning to Explore via Meta-Policy Gradient

no code implementations ICML 2018 Tianbing Xu, Qiang Liu, Liang Zhao, Jian Peng

The performance of off-policy learning, including deep Q-learning and deep deterministic policy gradient (DDPG), critically depends on the choice of the exploration policy.

Continuous Control Q-Learning +2

On the Margin Theory of Feedforward Neural Networks

no code implementations ICLR 2019 Colin Wei, Jason Lee, Qiang Liu, Tengyu Ma

We establish: 1) for multi-layer feedforward relu networks, the global minimizer of a weakly-regularized cross-entropy loss has the maximum normalized margin among all networks, 2) as a result, increasing the over-parametrization improves the normalized margin and generalization error bounds for deep networks.

A Kernel Loss for Solving the Bellman Equation

1 code implementation NeurIPS 2019 Yihao Feng, Lihong Li, Qiang Liu

Value function learning plays a central role in many state-of-the-art reinforcement-learning algorithms.

Q-Learning Reinforcement Learning (RL)

Training Robust Deep Neural Networks via Adversarial Noise Propagation

no code implementations19 Sep 2019 Aishan Liu, Xianglong Liu, Chongzhi Zhang, Hang Yu, Qiang Liu, DaCheng Tao

Various adversarial defense methods have accordingly been developed to improve adversarial robustness for deep models.

Adversarial Defense Adversarial Robustness

Network localization is unalterable by infections in bursts

1 code implementation11 Oct 2018 Qiang Liu, Piet Van Mieghem

To shed light on the disease localization phenomenon, we study a bursty susceptible-infected-susceptible (SIS) model and analyze the model under the mean-field approximation.

Physics and Society Social and Information Networks

Learning Preferences and Demands in Visual Recommendation

no code implementations11 Nov 2019 Qiang Liu, Shu Wu, Liang Wang

For modeling users' demands on different categories of items, the problem can be formulated as recommendation with contextual and sequential information.

Recommendation Systems

Detection and Classification of Astronomical Targets with Deep Neural Networks in Wide Field Small Aperture Telescopes

no code implementations21 Feb 2020 Peng Jia, Qiang Liu, Yongyang Sun

To increase the generalization ability of our framework, we use both simulated and real observation images to train the neural network.

General Classification Transfer Learning

Black-Box Certification with Randomized Smoothing: A Functional Optimization Based Framework

no code implementations NeurIPS 2020 Dinghuai Zhang, Mao Ye, Chengyue Gong, Zhanxing Zhu, Qiang Liu

Randomized classifiers have been shown to provide a promising approach for achieving certified robustness against adversarial attacks in deep learning.

Post-training Quantization with Multiple Points: Mixed Precision without Mixed Precision

no code implementations20 Feb 2020 Xingchao Liu, Mao Ye, Dengyong Zhou, Qiang Liu

We propose multipoint quantization, a quantization method that approximates a full-precision weight vector using a linear combination of multiple vectors of low-bit numbers; this is in contrast to typical quantization methods that approximate each weight using a single low precision number.

object-detection Object Detection +1

Stein Variational Inference for Discrete Distributions

no code implementations1 Mar 2020 Jun Han, Fan Ding, Xianglong Liu, Lorenzo Torresani, Jian Peng, Qiang Liu

In addition, such transform can be straightforwardly employed in gradient-free kernelized Stein discrepancy to perform goodness-of-fit (GOF) test on discrete distributions.

Variational Inference

Steepest Descent Neural Architecture Optimization: Escaping Local Optimum with Signed Neural Splitting

no code implementations23 Mar 2020 Lemeng Wu, Mao Ye, Qi Lei, Jason D. Lee, Qiang Liu

Recently, Liu et al.[19] proposed a splitting steepest descent (S2D) method that jointly optimizes the neural parameters and architectures based on progressively growing network structures by splitting neurons into multiple copies in a steepest descent fashion.

Dimension Independent Generalization Error by Stochastic Gradient Descent

no code implementations25 Mar 2020 Xi Chen, Qiang Liu, Xin T. Tong

One classical canon of statistics is that large models are prone to overfitting, and model selection procedures are necessary for high dimensional data.

Model Selection regression

Black-box Off-policy Estimation for Infinite-Horizon Reinforcement Learning

no code implementations ICLR 2020 Ali Mousavi, Lihong Li, Qiang Liu, Denny Zhou

Off-policy estimation for long-horizon problems is important in many real-life applications such as healthcare and robotics, where high-fidelity simulators may not be available and on-policy evaluation is expensive or impossible.

reinforcement-learning Reinforcement Learning (RL)

An Empirical Study on Feature Discretization

no code implementations27 Apr 2020 Qiang Liu, Zhaocheng Liu, Haoli Zhang

When dealing with continuous numeric features, we usually adopt feature discretization.

TFNet: Multi-Semantic Feature Interaction for CTR Prediction

no code implementations29 Jun 2020 Shu Wu, Feng Yu, Xueli Yu, Qiang Liu, Liang Wang, Tieniu Tan, Jie Shao, Fan Huang

The CTR (Click-Through Rate) prediction plays a central role in the domain of computational advertising and recommender systems.

Click-Through Rate Prediction Recommendation Systems

Go Wide, Then Narrow: Efficient Training of Deep Thin Networks

no code implementations ICML 2020 Denny Zhou, Mao Ye, Chen Chen, Tianjian Meng, Mingxing Tan, Xiaodan Song, Quoc Le, Qiang Liu, Dale Schuurmans

This is achieved by layerwise imitation, that is, forcing the thin network to mimic the intermediate outputs of the wide network from layer to layer.

Computational Efficiency Model Compression

Simplification of Graph Convolutional Networks: A Matrix Factorization-based Perspective

no code implementations17 Jul 2020 Qiang Liu, Haoli Zhang, Zhaocheng Liu

Moreover, we have also conducted experiments on a typical task of graph embedding, i. e., community detection, and the proposed UCMF model outperforms several representative graph embedding models.

Community Detection Distributed Computing +2

Deep Active Learning by Model Interpretability

no code implementations23 Jul 2020 Qiang Liu, Zhaocheng Liu, Xiaofang Zhu, Yeliang Xiu

In this paper, inspired by piece-wise linear interpretability in DNN, we introduce the linearly separable regions of samples to the problem of active learning, and propose a novel Deep Active learning approach by Model Interpretability (DAMI).

Active Learning Clustering +1

Disentangled Item Representation for Recommender Systems

no code implementations17 Aug 2020 Zeyu Cui, Feng Yu, Shu Wu, Qiang Liu, Liang Wang

In this way, the items are represented at the attribute level, which can provide fine-grained information of items in recommendation.

Attribute Recommendation Systems

Accountable Off-Policy Evaluation With Kernel Bellman Statistics

no code implementations15 Aug 2020 Yihao Feng, Tongzheng Ren, Ziyang Tang, Qiang Liu

We consider off-policy evaluation (OPE), which evaluates the performance of a new policy from observed data collected from previous experiments, without requiring the execution of the new policy.

Medical Diagnosis Off-policy evaluation +1

DeepSlicing: Deep Reinforcement Learning Assisted Resource Allocation for Network Slicing

no code implementations17 Aug 2020 Qiang Liu, Tao Han, Ning Zhang, Ye Wang

Network slicing enables multiple virtual networks run on the same physical infrastructure to support various use cases in 5G and beyond.

reinforcement-learning Reinforcement Learning (RL)

DNN2LR: Interpretation-inspired Feature Crossing for Real-world Tabular Data

no code implementations22 Aug 2020 Zhaocheng Liu, Qiang Liu, Haoli Zhang, Yuntian Chen

Simple classifiers, e. g., Logistic Regression (LR), are globally interpretable, but not powerful enough to model complex nonlinear interactions among features in tabular data.

A Chance-Constrained Generative Framework for Sequence Optimization

no code implementations ICML 2020 Xianggen Liu, Jian Peng, Qiang Liu, Sen Song

Deep generative modeling has achieved many successes for continuous data generation, such as producing realistic images and controlling their properties (e. g., styles).

valid

Accountable Off-Policy Evaluation via a Kernelized Bellman Statistics

no code implementations ICML 2020 Yihao Feng, Tongzheng Ren, Ziyang Tang, Qiang Liu

In this work, we investigate the statistical properties of the kernel loss, which allows us to find a feasible set that contains the true value function with high probability.

Off-policy evaluation

Network Pruning by Greedy Subnetwork Selection

no code implementations ICML 2020 Mao Ye, Chengyue Gong, Lizhen Nie, Denny Zhou, Adam Klivans, Qiang Liu

Theoretically, we show that the small networks pruned using our method achieve provably lower loss than small networks trained from scratch with the same size.

Network Pruning

Modeling Dyadic Conversations for Personality Inference

no code implementations26 Sep 2020 Qiang Liu

We adjust the formulation of each layer of a conventional GRU with sequence to sequence learning and personal information of both sides of the conversation.

Speeding up Deep Learning Training by Sharing Weights and Then Unsharing

no code implementations1 Jan 2021 Shuo Yang, Le Hou, Xiaodan Song, Qiang Liu, Denny Zhou

It has been widely observed that increasing deep learning model sizes often leads to significant performance improvements on a variety of natural language processing and computer vision tasks.

A Coach-Player Framework for Dynamic Team Composition

no code implementations1 Jan 2021 Bo Liu, Qiang Liu, Peter Stone, Animesh Garg, Yuke Zhu, Anima Anandkumar

The performance of our method is comparable or even better than the setting where all players have a full view of the environment, but no coach.

Zero-shot Generalization

Varying Coefficient Neural Network with Functional Targeted Regularization for Estimating Continuous Treatment Effects

no code implementations ICLR 2021 Lizhen Nie, Mao Ye, Qiang Liu, Dan Nicolae

With the rising abundance of observational data with continuous treatments, we investigate the problem of estimating average dose-response curve (ADRF).

Fast Training of Contrastive Learning with Intermediate Contrastive Loss

no code implementations1 Jan 2021 Chengyue Gong, Xingchao Liu, Qiang Liu

We apply our method to recently-proposed MOCO, SimCLR, SwAV and notice that we can reduce the computational cost with little loss on the performance of ImageNet linear classification and other downstream tasks.

Contrastive Learning

An efficient representation of chronological events in medical texts

no code implementations EMNLP (Louhi) 2020 Andrey Kormilitzin, Nemanja Vaci, Qiang Liu, Hao Ni, Goran Nenadic, Alejo Nevado-Holgado

In this work we addressed the problem of capturing sequential information contained in longitudinal electronic health records (EHRs).

Adaptive Dense-to-Sparse Paradigm for Pruning Online Recommendation System with Non-Stationary Data

no code implementations16 Oct 2020 Mao Ye, Dhruv Choudhary, Jiecao Yu, Ellie Wen, Zeliang Chen, Jiyan Yang, Jongsoo Park, Qiang Liu, Arun Kejariwal

To the best of our knowledge, this is the first work to provide in-depth analysis and discussion of applying pruning to online recommendation systems with non-stationary data distribution.

Recommendation Systems

Population Gradients improve performance across data-sets and architectures in object classification

no code implementations23 Oct 2020 Yurika Sakai, Andrey Kormilitzin, Qiang Liu, Alejo Nevado-Holgado

The most successful methods such as ReLU transfer functions, batch normalization, Xavier initialization, dropout, learning rate decay, or dynamic optimizers, have become standards in the field due, particularly, to their ability to increase the performance of Neural Networks (NNs) significantly and in almost all situations.

General Classification

Off-Policy Interval Estimation with Lipschitz Value Iteration

no code implementations NeurIPS 2020 Ziyang Tang, Yihao Feng, Na Zhang, Jian Peng, Qiang Liu

Off-policy evaluation provides an essential tool for evaluating the effects of different policies or treatments using only observed data.

Decision Making Medical Diagnosis +1

When Contrastive Learning Meets Active Learning: A Novel Graph Active Learning Paradigm with Self-Supervision

no code implementations30 Oct 2020 Yanqiao Zhu, Weizhi Xu, Qiang Liu, Shu Wu

To this end, we present a minimax selection scheme that explicitly harnesses neighborhood information and discover homophilous subgraphs to facilitate active selection.

Active Learning Contrastive Learning +2

Smart obervation method with wide field small aperture telescopes for real time transient detection

no code implementations20 Nov 2020 Peng Jia, Qiang Liu, Yongyang Sun, Yitian Zheng, Wenbo Liu, Yifei Zhao

The ARGUS uses a deep learning based astronomical detection algorithm implemented in embedded devices in each WFSATs to detect astronomical targets.

Ensemble Learning

AlphaMatch: Improving Consistency for Semi-supervised Learning with Alpha-divergence

no code implementations CVPR 2021 Chengyue Gong, Dilin Wang, Qiang Liu

Semi-supervised learning (SSL) is a key approach toward more data-efficient machine learning by jointly leverage both labeled and unlabeled data.

The Black-Scholes-Merton dual equation

no code implementations22 Dec 2019 Shuxin Guo, Qiang Liu

We derive the Black-Scholes-Merton dual equation, which has exactly the same form as the Black-Scholes-Merton equation.

Eudoxus: Characterizing and Accelerating Localization in Autonomous Machines

no code implementations2 Dec 2020 Yiming Gan, Yu Bo, Boyuan Tian, Leimeng Xu, Wei Hu, Shaoshan Liu, Qiang Liu, Yanjun Zhang, Jie Tang, Yuhao Zhu

We develop and commercialize autonomous machines, such as logistic robots and self-driving cars, around the globe.

Self-Driving Cars Hardware Architecture

LiveMap: Real-Time Dynamic Map in Automotive Edge Computing

no code implementations16 Dec 2020 Qiang Liu, Tao Han, Jiang, Xie, BaekGyu Kim

In this paper, we propose LiveMap, a real-time dynamic map, that detects, matches, and tracks objects on the road with crowdsourcing data from connected vehicles in sub-second.

Autonomous Driving Edge-computing +4

Coherent ray-wave structured light based on (helical) Ince-Gaussian modes

no code implementations4 Feb 2021 Zhaoyang Wang, Yijie Shen, Qiang Liu, Xing Fu

The topological evolution of classic eigenmodes including Hermite-Laguerre-Gaussian and (helical) InceGaussian modes is exploited to construct coherent state modes, which unifies the representations of travelingwave (TW) and standing-wave (SW) ray-wave structured light for the first time and realizes the TW-SW unified ray-wave geometric beam with topology of raytrajectories splitting effect, breaking the boundary of TW and SW structured light.

Optics

Centroid Transformers: Learning to Abstract with Attention

no code implementations17 Feb 2021 Lemeng Wu, Xingchao Liu, Qiang Liu

Self-attention, as the key block of transformers, is a powerful mechanism for extracting features from the inputs.

Abstractive Text Summarization Clustering +1

DNN2LR: Automatic Feature Crossing for Credit Scoring

no code implementations24 Feb 2021 Qiang Liu, Zhaocheng Liu, Haoli Zhang, Yuntian Chen, Jun Zhu

Accordingly, we can design an automatic feature crossing method to find feature interactions in DNN, and use them as cross features in LR.

Feature Engineering

A Survey on Graph Structure Learning: Progress and Opportunities

no code implementations4 Mar 2021 Yanqiao Zhu, Weizhi Xu, Jinghao Zhang, Yuanqi Du, Jieyu Zhang, Qiang Liu, Carl Yang, Shu Wu

Specifically, we first formulate a general pipeline of GSL and review state-of-the-art methods classified by the way of modeling graph structures, followed by applications of GSL across domains.

Graph structure learning

Non-asymptotic Confidence Intervals of Off-policy Evaluation: Primal and Dual Bounds

no code implementations ICLR 2021 Yihao Feng, Ziyang Tang, Na Zhang, Qiang Liu

Off-policy evaluation (OPE) is the task of estimating the expected reward of a given policy based on offline data previously collected under different policies.

Off-policy evaluation Open-Ended Question Answering +1

DyGCN: Dynamic Graph Embedding with Graph Convolutional Network

no code implementations7 Apr 2021 Zeyu Cui, Zekun Li, Shu Wu, XiaoYu Zhang, Qiang Liu, Liang Wang, Mengmeng Ai

We naturally generalizes the embedding propagation scheme of GCN to dynamic setting in an efficient manner, which is to propagate the change along the graph to update node embeddings.

Dynamic graph embedding

Physics-Guided Discovery of Highly Nonlinear Parametric Partial Differential Equations

no code implementations2 Jun 2021 Yingtao Luo, Qiang Liu, Yuntian Chen, WenBo Hu, Tian Tian, Jun Zhu

Especially, the discovery of PDEs with highly nonlinear coefficients from low-quality data remains largely under-addressed.

Density Estimation Model Optimization

PNet -- A Deep Learning Based Photometry and Astrometry Bayesian Framework

no code implementations28 Jun 2021 Rui Sun, Peng Jia, Yongyang Sun, Zhimin Yang, Qiang Liu, Hongyan Wei

Time domain astronomy has emerged as a vibrant research field in recent years, focusing on celestial objects that exhibit variable magnitudes or positions.

Astronomy regression +1

MaxUp: Lightweight Adversarial Training With Data Augmentation Improves Neural Network Training

no code implementations CVPR 2021 Chengyue Gong, Tongzheng Ren, Mao Ye, Qiang Liu

The idea is to generate a set of augmented data with some random perturbations or transforms, and minimize the maximum, or worst case loss over the augmented data.

Data Augmentation Image Classification +1

Real-Time Anchor-Free Single-Stage 3D Detection with IoU-Awareness

no code implementations29 Jul 2021 Runzhou Ge, Zhuangzhuang Ding, Yihan Hu, Wenxin Shao, Li Huang, Kun Li, Qiang Liu

Extended from our last year's award-winning model AFDet, we have made a handful of modifications to the base model, to improve the accuracy and at the same time to greatly reduce the latency.

Data Augmentation

Deep Contrastive Multiview Network Embedding

no code implementations16 Aug 2021 Mengqi Zhang, Yanqiao Zhu, Qiang Liu, Shu Wu, Liang Wang

In our work, different views can be obtained based on the various relations among nodes.

Attribute Contrastive Learning +2

Deep Active Learning for Text Classification with Diverse Interpretations

no code implementations15 Aug 2021 Qiang Liu, Yanqiao Zhu, Zhaocheng Liu, Yufeng Zhang, Shu Wu

To train high-performing models with the minimal annotation cost, active learning is proposed to select and label the most informative samples, yet it is still challenging to measure informativeness of samples used in DNNs.

Active Learning Informativeness +3

Structure-Aware Hard Negative Mining for Heterogeneous Graph Contrastive Learning

no code implementations31 Aug 2021 Yanqiao Zhu, Yichen Xu, Hejie Cui, Carl Yang, Qiang Liu, Shu Wu

Recently, heterogeneous Graph Neural Networks (GNNs) have become a de facto model for analyzing HGs, while most of them rely on a relative large number of labeled data.

Contrastive Learning

Speeding up Deep Model Training by Sharing Weights and Then Unsharing

no code implementations8 Oct 2021 Shuo Yang, Le Hou, Xiaodan Song, Qiang Liu, Denny Zhou

Our approach exploits the special structure of BERT that contains a stack of repeated modules (i. e., transformer encoders).

Neural Energy Minimization for Molecular Conformation Optimization

no code implementations ICLR 2022 Jiaqi Guan, Wesley Wei Qian, Qiang Liu, Wei-Ying Ma, Jianzhu Ma, Jian Peng

Assuming different forms of the underlying potential energy function, we can not only reinterpret and unify many of the existing models but also derive new variants of SE(3)-equivariant neural networks in a principled manner.

Pareto Navigation Gradient Descent: a First Order Algorithm for Optimization in Pareto Set

no code implementations29 Sep 2021 Mao Ye, Qiang Liu

The notion of the Pareto set allows us to focus on the set of (often infinite number of) models that cannot be strictly improved.

Multi-Task Learning

Pareto Navigation Gradient Descent: a First-Order Algorithm for Optimization in Pareto Set

no code implementations17 Oct 2021 Mao Ye, Qiang Liu

The notion of the Pareto set allows us to focus on the set of (often infinite number of) models that cannot be strictly improved.

Multi-Task Learning

Centroid Approximation for Bootstrap: Improving Particle Quality at Inference

no code implementations17 Oct 2021 Mao Ye, Qiang Liu

In this work, we propose an efficient method to explicitly \emph{optimize} a small set of high quality ``centroid'' points to better approximate the ideal bootstrap distribution.

Uncertainty Quantification

OnSlicing: Online End-to-End Network Slicing with Reinforcement Learning

no code implementations2 Nov 2021 Qiang Liu, Nakjung Choi, Tao Han

As online learning is converged, OnSlicing reduces 12. 5% usage without any violations as compared to the state-of-the-art online DRL solution.

reinforcement-learning Reinforcement Learning (RL)

A Convolutional Click Prediction Model

no code implementations CIKM 2015 Qiang Liu, Feng Yu, Shu Wu, Liang Wang

The explosion in online advertisement urges to better estimate the click prediction of ads.

Deep Interaction Machine: A Simple but Effective Model for High-order Feature Interactions

no code implementations CIKM 2020 Feng Yu, Zhaocheng Liu, Qiang Liu, Haoli Zhang, Shu Wu, Liang Wang

IM is an efficient and exact implementation of high-order FM, whose time complexity linearly grows with the order of interactions and the number of feature fields.

Click-Through Rate Prediction Feature Engineering

argmax centroid

no code implementations NeurIPS 2021 Chengyue Gong, Mao Ye, Qiang Liu

We propose a general method to construct centroid approximation for the distribution of maximum points of a random function (a. k. a.

Domain Adaptation Few-Shot Image Classification +2

Automatic and Harmless Regularization with Constrained and Lexicographic Optimization: A Dynamic Barrier Approach

no code implementations NeurIPS 2021 Chengyue Gong, Xingchao Liu, Qiang Liu

In this work, we consider constrained optimization as a more principled approach for trading off two losses, with a special emphasis on lexicographic optimization, a degenerated limit of constrained optimization which optimizes a secondary loss inside the optimal set of the main loss.

Shrinkage-based Bias-Variance Trade-off for Deep Reinforcement Learning

no code implementations27 Sep 2018 Yihao Feng, Hao liu, Jian Peng, Qiang Liu

Deep reinforcement learning has achieved remarkable successes in solving various challenging artificial intelligence tasks.

Continuous Control reinforcement-learning +1

Automatically Learning Feature Crossing from Model Interpretation for Tabular Data

no code implementations25 Sep 2019 Zhaocheng Liu, Qiang Liu, Haoli Zhang

Automatically feature generation is a major topic of automated machine learning.

Filling the Soap Bubbles: Efficient Black-Box Adversarial Certification with Non-Gaussian Smoothing

no code implementations25 Sep 2019 Dinghuai Zhang*, Mao Ye*, Chengyue Gong*, Zhanxing Zhu, Qiang Liu

Randomized classifiers have been shown to provide a promising approach for achieving certified robustness against adversarial attacks in deep learning.

Statistical Adaptive Stochastic Optimization

no code implementations25 Sep 2019 Pengchuan Zhang, Hunter Lang, Qiang Liu, Lin Xiao

We investigate statistical methods for automatically scheduling the learning rate (step size) in stochastic optimization.

Scheduling Stochastic Optimization

Automatic and Harmless Regularization with Constrained and Lexicographic Optimization: A Dynamic Barrier Approach

no code implementations NeurIPS 2021 Chengyue Gong, Xingchao Liu, Qiang Liu

In this work, we consider constrained optimization as a more principled approach for trading off two losses, with a special emphasis on lexicographic optimization, a degenerated limit of constrained optimization which optimizes a secondary loss inside the optimal set of the main loss.

Cannot find the paper you are looking for? You can Submit a new open access paper.