1 code implementation • 25 May 2021 • Meng-Jiun Chiou, Chun-Yu Liao, Li-Wei Wang, Roger Zimmermann, Jiashi Feng
Detecting human-object interactions (HOI) is an important step toward a comprehensive visual understanding of machines.
Ranked #3 on
Human-Object Interaction Anticipation
on VidHOI
1 code implementation • NeurIPS 2020 • Jingtong Su, Yihang Chen, Tianle Cai, Tianhao Wu, Ruiqi Gao, Li-Wei Wang, Jason D. Lee
In this paper, we conduct sanity checks for the above beliefs on several recent unstructured pruning methods and surprisingly find that: (1) A set of methods which aims to find good subnetworks of the randomly-initialized network (which we call "initial tickets"), hardly exploits any information from the training data; (2) For the pruned networks obtained by these methods, randomly changing the preserved weights in each layer, while keeping the total number of preserved weights unchanged per layer, does not affect the final performance.
1 code implementation • 7 Sep 2020 • Tianle Cai, Shengjie Luo, Keyulu Xu, Di He, Tie-Yan Liu, Li-Wei Wang
We provide an explanation by showing that InstanceNorm serves as a preconditioner for GNNs, but such preconditioning effect is weaker with BatchNorm due to the heavy batch noise in graph datasets.
Ranked #25 on
Graph Property Prediction
on ogbg-molhiv
no code implementations • ICLR 2021 • Xiaoyu Chen, Jiachen Hu, Lihong Li, Li-Wei Wang
The regret of FMDP-BF is shown to be exponentially smaller than that of optimal algorithms designed for non-factored MDPs, and improves on the best previous result for FMDPs~\citep{osband2014near} by a factored of $\sqrt{H|\mathcal{S}_i|}$, where $|\mathcal{S}_i|$ is the cardinality of the factored state subspace and $H$ is the planning horizon.
1 code implementation • ECCV 2020 • Zhengyuan Yang, Tianlang Chen, Li-Wei Wang, Jiebo Luo
We improve one-stage visual grounding by addressing current limitations on grounding long and complex queries.
1 code implementation • ICLR 2019 • Tiange Luo, Tianle Cai, Mengxiao Zhang, Siyu Chen, Li-Wei Wang
We next investigate the adversarial examples which 'fool' a CNN with Random Mask.
no code implementations • 24 Jul 2020 • Yunzhen Feng, Runtian Zhai, Di He, Li-Wei Wang, Bin Dong
Our experiments show that TD can provide fine-grained information for varied downstream tasks, and for the models trained from different initializations, the learned features are not the same in terms of downstream-task predictions.
1 code implementation • ECCV 2020 • Yiwu Zhong, Li-Wei Wang, Jianshu Chen, Dong Yu, Yin Li
We address the challenging problem of image captioning by revisiting the representation of image scene graph.
1 code implementation • NeurIPS 2020 • Yihong Chen, Zheng Zhang, Yue Cao, Li-Wei Wang, Stephen Lin, Han Hu
Though RepPoints provides high performance, we find that its heavy reliance on regression for object localization leaves room for improvement.
Ranked #27 on
Object Detection
on COCO-O
1 code implementation • 10 Jun 2020 • Zhenhui Xu, Linyuan Gong, Guolin Ke, Di He, Shuxin Zheng, Li-Wei Wang, Jiang Bian, Tie-Yan Liu
Pre-trained contextual representations (e. g., BERT) have become the foundation to achieve state-of-the-art results on many NLP tasks.
no code implementations • 5 Jun 2020 • Ke Lin, Zhuoxin Gan, Li-Wei Wang
This report describes our model for VATEX Captioning Challenge 2020.
no code implementations • ICML 2020 • Xiaoyu Chen, Kai Zheng, Zixin Zhou, Yunchang Yang, Wei Chen, Li-Wei Wang
In this paper, we study Combinatorial Semi-Bandits (CSB) that is an extension of classic Multi-Armed Bandits (MAB) under Differential Privacy (DP) and stronger Local Differential Privacy (LDP) setting.
3 code implementations • NeurIPS 2020 • Kai Zheng, Tianle Cai, Weiran Huang, Zhenguo Li, Li-Wei Wang
We study locally differentially private (LDP) bandits learning in this paper.
1 code implementation • 1 Jun 2020 • Yu-Chin Chan, Faez Ahmed, Li-Wei Wang, Wei Chen
In answer, we posit that a smaller yet diverse set of unit cells leads to scalable search and unbiased learning.
no code implementations • CVPR 2020 • Aoxue Li, Weiran Huang, Xu Lan, Jiashi Feng, Zhenguo Li, Li-Wei Wang
Few-shot learning (FSL) has attracted increasing attention in recent years but remains challenging, due to the intrinsic difficulty in learning to generalize from a few examples.
Ranked #1 on
Few-Shot Image Classification
on ImageNet (1-shot)
no code implementations • 27 May 2020 • Dong Wang, Kexin Zhang, Jia Ding, Li-Wei Wang
In the clinical practice, Tanner and Whitehouse (TW2) method is a widely-used method for radiologists to perform BAA.
1 code implementation • ACL 2020 • Jie Lei, Li-Wei Wang, Yelong Shen, Dong Yu, Tamara L. Berg, Mohit Bansal
Generating multi-sentence descriptions for videos is one of the most challenging captioning tasks due to its high requirements for not only visual relevance but also discourse-based coherence across the sentences in the paragraph.
Ranked #5 on
Video Captioning
on ActivityNet Captions
2 code implementations • CVPR 2020 • Yihong Chen, Yue Cao, Han Hu, Li-Wei Wang
We argue that there are two important cues for humans to recognize objects in videos: the global semantic information and the local localization information.
Ranked #14 on
Video Object Detection
on ImageNet VID
no code implementations • CVPR 2020 • Dong Wang, Yuan Zhang, Kexin Zhang, Li-Wei Wang
Applying artificial intelligence techniques in medical imaging is one of the most promising areas in medicine.
1 code implementation • ICLR 2020 • Tiange Luo, Kaichun Mo, Zhiao Huang, Jiarui Xu, Siyu Hu, Li-Wei Wang, Hao Su
We address the problem of discovering 3D parts for objects in unseen categories.
9 code implementations • ICML 2020 • Ruibin Xiong, Yunchang Yang, Di He, Kai Zheng, Shuxin Zheng, Chen Xing, Huishuai Zhang, Yanyan Lan, Li-Wei Wang, Tie-Yan Liu
This motivates us to remove the warm-up stage for the training of Pre-LN Transformers.
no code implementations • 10 Feb 2020 • Wei Chen, Li-Wei Wang, Haoyu Zhao, Kai Zheng
In a special case where the reward function is linear and we have an exact oracle, we design a parameter-free algorithm that achieves nearly optimal regret both in the switching case and in the dynamic case without knowing the parameters in advance.
2 code implementations • ICLR 2020 • Runtian Zhai, Chen Dan, Di He, huan zhang, Boqing Gong, Pradeep Ravikumar, Cho-Jui Hsieh, Li-Wei Wang
Adversarial training is one of the most popular ways to learn robust models but is usually attack-dependent and time costly.
2 code implementations • ECCV 2020 • Ze Yang, Yinghao Xu, Han Xue, Zheng Zhang, Raquel Urtasun, Li-Wei Wang, Stephen Lin, Han Hu
We present a new object representation, called Dense RepPoints, that utilizes a large set of points to describe an object at multiple levels, including both box level and pixel level.
1 code implementation • 19 Nov 2019 • Tiange Luo, Tianle Cai, Mengxiao Zhang, Siyu Chen, Di He, Li-Wei Wang
Robustness of convolutional neural networks (CNNs) has gained in importance on account of adversarial examples, i. e., inputs added as well-designed perturbations that are imperceptible to humans but can cause the model to predict incorrectly.
no code implementations • 24 Oct 2019 • Sunyang Fu, David Chen, Huan He, Sijia Liu, Sungrim Moon, Kevin J Peterson, Feichen Shen, Li-Wei Wang, Yanshan Wang, Andrew Wen, Yiqing Zhao, Sunghwan Sohn, Hongfang Liu
Background Concept extraction, a subdomain of natural language processing (NLP) with a focus on extracting concepts of interest, has been adopted to computationally extract clinical information from text for a wide range of applications ranging from clinical decision support to care quality improvement.
no code implementations • 13 Oct 2019 • Chao-Tsung Huang, Yu-Chun Ding, Huan-Ching Wang, Chi-Wen Weng, Kai-Ping Lin, Li-Wei Wang, Li-De Chen
In this paper, we approach this goal by considering the inference flow, network model, instruction set, and processor design jointly to optimize hardware performance and image quality.
no code implementations • 27 Sep 2019 • Jinchen Xuan, Yunchang Yang, Ze Yang, Di He, Li-Wei Wang
Motivated by this observation, we discover two specific problems of GANs leading to anomalous generalization behaviour, which we refer to as the sample insufficiency and the pixel-wise combination.
1 code implementation • ICLR 2020 • Chuanbiao Song, Kun He, Jiadong Lin, Li-Wei Wang, John E. Hopcroft
We continue to propose a new approach called Robust Local Features for Adversarial Training (RLFAT), which first learns the robust local features by adversarial training on the RBS-transformed adversarial examples, and then transfers the robust local features into the training of normal adversarial examples.
1 code implementation • IJCNLP 2019 • Zhuohan Li, Zi Lin, Di He, Fei Tian, Tao Qin, Li-Wei Wang, Tie-Yan Liu
Due to the unparallelizable nature of the autoregressive factorization, AutoRegressive Translation (ART) models have to generate tokens sequentially during decoding and thus suffer from high inference latency.
no code implementations • NeurIPS 2019 • Rui Ray Zhang, Xingwu Liu, Yuyi Wang, Li-Wei Wang
We demonstrate that for many types of dependent data, the forest complexity is small and thus implies good concentration.
2 code implementations • ICCV 2019 • Zhengyuan Yang, Boqing Gong, Li-Wei Wang, Wenbing Huang, Dong Yu, Jiebo Luo
We propose a simple, fast, and accurate one-stage approach to visual grounding, inspired by the following insight.
3 code implementations • ICLR 2020 • Jiadong Lin, Chuanbiao Song, Kun He, Li-Wei Wang, John E. Hopcroft
While SIM is based on our discovery on the scale-invariant property of deep learning models, for which we leverage to optimize the adversarial perturbations over the scale copies of the input images so as to avoid "overfitting" on the white-box model being attacked and generate more transferable adversarial examples.
2 code implementations • ICCV 2019 • Tiange Luo, Aoxue Li, Tao Xiang, Weiran Huang, Li-Wei Wang
In this paper, we propose to tackle the challenging few-shot learning (FSL) problem by learning global class representations using both base and novel class training samples.
1 code implementation • ICLR 2019 • Jun Gao, Di He, Xu Tan, Tao Qin, Li-Wei Wang, Tie-Yan Liu
We study an interesting problem in training neural network-based models for natural language generation tasks, which we call the \emph{representation degeneration problem}.
no code implementations • 23 Jun 2019 • Jun Jiang, Shumao Pang, Xia Zhao, Li-Wei Wang, Andrew Wen, Hongfang Liu, Qianjin Feng
In order to train a generalizable model, a large volume of text must be collected.
no code implementations • NeurIPS 2019 • Ruiqi Gao, Tianle Cai, Haochuan Li, Li-Wei Wang, Cho-Jui Hsieh, Jason D. Lee
Neural networks are vulnerable to adversarial examples, i. e. inputs that are imperceptibly perturbed from natural data and yet incorrectly classified by the network.
2 code implementations • ICLR 2020 • Yiping Lu, Zhuohan Li, Di He, Zhiqing Sun, Bin Dong, Tao Qin, Li-Wei Wang, Tie-Yan Liu
In this paper, we provide a novel perspective towards understanding the architecture: we show that the Transformer can be mathematically interpreted as a numerical Ordinary Differential Equation (ODE) solver for a convection-diffusion equation in a multi-particle dynamic system.
1 code implementation • 3 Jun 2019 • Runtian Zhai, Tianle Cai, Di He, Chen Dan, Kun He, John Hopcroft, Li-Wei Wang
Neural network robustness has recently been highlighted by the existence of adversarial examples.
no code implementations • WS 2019 • Sijia Liu, Li-Wei Wang, Vipin Chaudhary, Hongfang Liu
Neural network models have shown promise in the temporal relation extraction task.
no code implementations • NeurIPS 2019 • Kai Zheng, Haipeng Luo, Ilias Diakonikolas, Li-Wei Wang
We propose the first reduction-based approach to obtaining long-term memory guarantees for online learning in the sense of Bousquet and Warmuth, 2002, by reducing the problem to achieving typical switching regret.
no code implementations • 28 May 2019 • Tianle Cai, Ruiqi Gao, Jikai Hou, Siyu Chen, Dong Wang, Di He, Zhihua Zhang, Li-Wei Wang
First-order methods such as stochastic gradient descent (SGD) are currently the standard algorithm for training deep neural networks.
1 code implementation • ACL 2019 • Kun Xu, Li-Wei Wang, Mo Yu, Yansong Feng, Yan Song, Zhiguo Wang, Dong Yu
Previous cross-lingual knowledge graph (KG) alignment studies rely on entity embeddings derived only from monolingual KG structural information, which may fail at matching entities that have different facts in two KGs.
no code implementations • ICLR 2019 • Zhuohan Li, Di He, Fei Tian, Tao Qin, Li-Wei Wang, Tie-Yan Liu
To improve the accuracy of NART models, in this paper, we propose to leverage the hints from a well-trained ART model to train the NART model.
6 code implementations • ICCV 2019 • Ze Yang, Shaohui Liu, Han Hu, Li-Wei Wang, Stephen Lin
They furthermore do not require the use of anchors to sample a space of bounding boxes.
Ranked #100 on
Object Detection
on COCO minival
no code implementations • 16 Apr 2019 • Xiaosen Wang, Kun He, Chuanbiao Song, Li-Wei Wang, John E. Hopcroft
In this way, AT-GAN can learn the distribution of adversarial examples that is very close to the distribution of real data.
no code implementations • ICLR 2020 • Yuanhao Wang, Jiachen Hu, Xiaoyu Chen, Li-Wei Wang
We study the problem of regret minimization for distributed bandits learning, in which $M$ agents work collaboratively to minimize their total regret under the coordination of a central server.
no code implementations • 27 Mar 2019 • Jun Gao, Xiao Li, Li-Wei Wang, Sanja Fidler, Stephen Lin
We present a method for compositing virtual objects into a photograph such that the object colors appear to have been processed by the photo's camera imaging pipeline.
no code implementations • ICLR 2020 • Kefan Dong, Yuanhao Wang, Xiaoyu Chen, Li-Wei Wang
A fundamental question in reinforcement learning is whether model-free algorithms are sample efficient.
no code implementations • 9 Nov 2018 • Simon S. Du, Jason D. Lee, Haochuan Li, Li-Wei Wang, Xiyu Zhai
Gradient descent finds a global minimum in training deep neural networks despite the objective function being non-convex.
no code implementations • 19 Oct 2018 • Aoxue Li, Zhiwu Lu, Jiechao Guan, Tao Xiang, Li-Wei Wang, Ji-Rong Wen
Inspired by the fact that an unseen class is not exactly `unseen' if it belongs to the same superclass as a seen class, we propose a novel inductive ZSL model that leverages superclasses as the bridge between seen and unseen classes to narrow the domain gap.
2 code implementations • ICLR 2019 • Chuanbiao Song, Kun He, Li-Wei Wang, John E. Hopcroft
Our intuition is to regard the adversarial training on FGSM adversary as a domain adaption task with limited number of target domain samples.
2 code implementations • NeurIPS 2018 • Chengyue Gong, Di He, Xu Tan, Tao Qin, Li-Wei Wang, Tie-Yan Liu
Continuous word representation (aka word embedding) is a basic building block in many neural network-based models used in natural language processing tasks.
Ranked #3 on
Machine Translation
on IWSLT2015 German-English
12 code implementations • ECCV 2018 • Ze Yang, Tiange Luo, Dong Wang, Zhiqiang Hu, Jun Gao, Li-Wei Wang
In consideration of intrinsic consistency between informativeness of the regions and their probability being ground-truth class, we design a novel training paradigm, which enables Navigator to detect most informative regions under the guidance from Teacher.
Ranked #45 on
Fine-Grained Image Classification
on FGVC Aircraft
5 code implementations • 28 Aug 2018 • Yanshan Wang, Naveed Afzal, Sunyang Fu, Li-Wei Wang, Feichen Shen, Majid Rastegar-Mojarad, Hongfang Liu
A subset of MedSTS (MedSTS_ann) containing 1, 068 sentence pairs was annotated by two medical experts with semantic similarity scores of 0-5 (low to high similarity).
no code implementations • ICML 2018 • Wenlong Mou, Yuchen Zhou, Jun Gao, Li-Wei Wang
We study the problem of generalization guarantees for dropout training.
1 code implementation • ICML 2018 • Zhuohan Li, Di He, Fei Tian, Wei Chen, Tao Qin, Li-Wei Wang, Tie-Yan Liu
Long Short-Term Memory (LSTM) is one of the most widely used recurrent structures in sequence modeling.
no code implementations • CVPR 2019 • Aditya Deshpande, Jyoti Aneja, Li-Wei Wang, Alexander Schwing, D. A. Forsyth
We achieve the trifecta: (1) High accuracy for the diverse captions as evaluated by standard captioning metrics and user studies; (2) Faster computation of diverse captions compared to beam search and diverse beam search; and (3) High diversity as evaluated by counting novel sentences, distinct n-grams and mutual overlap (i. e., mBleu-4) scores.
no code implementations • ECCV 2018 • Jiayuan Gu, Han Hu, Li-Wei Wang, Yichen Wei, Jifeng Dai
While most steps in the modern object detection methods are learnable, the region feature extraction step remains largely hand-crafted, featured by RoI pooling methods.
2 code implementations • 1 Feb 2018 • Yanshan Wang, Sijia Liu, Naveed Afzal, Majid Rastegar-Mojarad, Li-Wei Wang, Feichen Shen, Paul Kingsbury, Hongfang Liu
First, the word embeddings trained on clinical notes and biomedical publications can capture the semantics of medical terms better, and find more relevant similar medical terms, and are closer to human experts' judgments, compared to these trained on Wikipedia and news.
Information Retrieval
no code implementations • NeurIPS 2017 • Di He, Hanqing Lu, Yingce Xia, Tao Qin, Li-Wei Wang, Tie-Yan Liu
Inspired by the success and methodology of AlphaGo, in this paper we propose using a prediction network to improve beam search, which takes the source sentence $x$, the currently available decoding output $y_1,\cdots, y_{t-1}$ and a candidate word $w$ at step $t$ as inputs and predicts the long-term value (e. g., BLEU score) of the partial target sentence if it is completed by the NMT model.
1 code implementation • NeurIPS 2017 • Zhou Lu, Hongming Pu, Feicheng Wang, Zhiqiang Hu, Li-Wei Wang
That is, there are classes of deep networks which cannot be realized by any shallow network whose size is no more than an exponential bound.
no code implementations • 19 Jul 2017 • Wenlong Mou, Li-Wei Wang, Xiyu Zhai, Kai Zheng
This is the first algorithm-dependent result with reasonable dependence on aggregated step sizes for non-convex learning, and has important implications to statistical learning aspects of stochastic gradient methods in complicated models such as deep learning.
no code implementations • 4 Jul 2017 • Aoxue Li, Zhiwu Lu, Li-Wei Wang, Tao Xiang, Xinqi Li, Ji-Rong Wen
In this paper, to address the two issues, we propose a two-phase framework for recognizing images from unseen fine-grained classes, i. e. zero-shot fine-grained classification.
no code implementations • 14 Jun 2017 • Jia Ding, Aoxue Li, Zhiqiang Hu, Li-Wei Wang
Early detection of pulmonary cancer is the most promising way to enhance a patient's chance for survival.
no code implementations • ICML 2017 • Kai Zheng, Wenlong Mou, Li-Wei Wang
For learning with smooth generalized linear losses, we propose an approximate stochastic gradient oracle estimated from non-interactive LDP channel, using Chebyshev expansion.
no code implementations • 2 Apr 2017 • Kun He, Jingbo Wang, Haochuan Li, Yao Shu, Mengxiao Zhang, Man Zhu, Li-Wei Wang, John E. Hopcroft
Toward a deeper understanding on the inner work of deep neural networks, we investigate CNN (convolutional neural network) using DCN (deconvolutional network) and randomization technique, and gain new insights for the intrinsic property of this network architecture.
no code implementations • 29 Mar 2017 • Jiaqi Zhang, Kai Zheng, Wenlong Mou, Li-Wei Wang
For strongly convex and smooth objectives, we prove that gradient descent with output perturbation not only achieves nearly optimal utility, but also significantly improves the running time of previous state-of-the-art private optimization algorithms, for both $\epsilon$-DP and $(\epsilon, \delta)$-DP.
no code implementations • 18 Feb 2017 • Lunjia Hu, Ruihan Wu, Tianhong Li, Li-Wei Wang
The RTD of a concept class $\mathcal C \subseteq \{0, 1\}^n$, introduced by Zilles et al. (2011), is a combinatorial complexity measure characterized by the worst-case number of examples necessary to identify a concept in $\mathcal C$ according to the recursive teaching model.
no code implementations • 14 Dec 2016 • Wenlong Mou, Zhi Wang, Li-Wei Wang
In Valiant's neuroidal model, the hippocampus was described as a randomly connected graph, the computation on which maps input to a set of activated neuroids with stable size.
1 code implementation • NeurIPS 2016 • Yingce Xia, Di He, Tao Qin, Li-Wei Wang, Nenghai Yu, Tie-Yan Liu, Wei-Ying Ma
Based on the feedback signals generated during this process (e. g., the language-model likelihood of the output of a model, and the reconstruction error of the original sentence after the primal and dual translations), we can iteratively update the two models until convergence (e. g., using the policy gradient methods).
no code implementations • 17 Jun 2015 • Shizhao Sun, Wei Chen, Li-Wei Wang, Xiaoguang Liu, Tie-Yan Liu
First, we derive an upper bound for RA of DNN, and show that it increases with increasing depth.
2 code implementations • ICCV 2015 • Bryan A. Plummer, Li-Wei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, Svetlana Lazebnik
The Flickr30k dataset has become a standard benchmark for sentence-based image description.
Ranked #15 on
Phrase Grounding
on Flickr30k Entities Test
no code implementations • 18 Jan 2015 • Zhiwu Lu, Li-Wei Wang, Ji-Rong Wen
This paper presents a new framework for visual bag-of-words (BOW) refinement and reduction to overcome the drawbacks associated with the visual BOW model which has been widely used for image classification.
no code implementations • 18 Jan 2015 • Zhiwu Lu, Li-Wei Wang
This paper presents a graph-based learning approach to pairwise constraint propagation on multi-view data.
no code implementations • 3 Jun 2014 • Di He, Wei Chen, Li-Wei Wang, Tie-Yan Liu
Sponsored search is an important monetization channel for search engines, in which an auction mechanism is used to select the ads shown to users and determine the prices charged from advertisers.
no code implementations • 7 Mar 2014 • Zhiwu Lu, Zhen-Yong Fu, Tao Xiang, Li-Wei Wang, Ji-Rong Wen
By oversegmenting all the images into regions, we formulate noisily tagged image parsing as a weakly supervised sparse learning problem over all the regions, where the initial labels of each region are inferred from image-level labels.
no code implementations • 7 Mar 2014 • Yunchao Gong, Li-Wei Wang, Ruiqi Guo, Svetlana Lazebnik
Deep convolutional neural networks (CNN) have shown their promise as a universal representation for recognition.
no code implementations • 6 Jan 2014 • Chi Jin, Ziteng Wang, Junliang Huang, Yiqiao Zhong, Li-Wei Wang
We develop an $\epsilon$-differentially private mechanism for the class of $K$-smooth queries.
no code implementations • NeurIPS 2013 • Ziteng Wang, Kai Fan, Jia-Qi Zhang, Li-Wei Wang
Outputting the summary runs in time $O(n^{1+\frac{d}{2d+K}})$, and the evaluation algorithm for answering a query runs in time $\tilde O (n^{\frac{d+2+\frac{2d}{K}}{2d+K}} )$.
no code implementations • 24 Apr 2013 • Yining Wang, Li-Wei Wang, Yuanzhi Li, Di He, Tie-Yan Liu, Wei Chen
We show that NDCG with logarithmic discount has consistent distinguishability although it converges to the same limit for all ranking functions.
no code implementations • NeurIPS 2012 • Chi Jin, Li-Wei Wang
We show that our bound is strictly sharper than a previously well-known PAC-Bayes margin bound if the feature space is of finite dimension; and the two bounds tend to be equivalent as the dimension goes to infinity.