Search Results for author: Zhangyang Wang

Found 357 papers, 221 papers with code

On a Theory of Nonparametric Pairwise Similarity for Clustering: Connecting Clustering to Classification

no code implementations NeurIPS 2014 Yingzhen Yang, Feng Liang, Shuicheng Yan, Zhangyang Wang, Thomas S. Huang

Modeling the underlying data distribution by nonparametric kernel density estimation, the generalization error bounds for both unsupervised nonparametric classifiers are the sum of nonparametric pairwise similarity terms between the data points for the purpose of clustering.

Clustering Density Estimation +2

Learning Super-Resolution Jointly from External and Internal Examples

no code implementations3 Mar 2015 Zhangyang Wang, Yingzhen Yang, Zhaowen Wang, Shiyu Chang, Jianchao Yang, Thomas S. Huang

Single image super-resolution (SR) aims to estimate a high-resolution (HR) image from a lowresolution (LR) input.

Image Super-Resolution

Designing A Composite Dictionary Adaptively From Joint Examples

no code implementations12 Mar 2015 Zhangyang Wang, Yingzhen Yang, Jianchao Yang, Thomas S. Huang

We study the complementary behaviors of external and internal examples in image restoration, and are motivated to formulate a composite dictionary design framework.

Image Denoising Image Restoration +1

DeepFont: Identify Your Font from An Image

1 code implementation12 Jul 2015 Zhangyang Wang, Jianchao Yang, Hailin Jin, Eli Shechtman, Aseem Agarwala, Jonathan Brandt, Thomas S. Huang

As font is one of the core design concepts, automatic font identification and similar font suggestion from an image or photo has been on the wish list of many designers.

Domain Adaptation Font Recognition +1

Learning Deep $\ell_0$ Encoders

no code implementations1 Sep 2015 Zhangyang Wang, Qing Ling, Thomas S. Huang

We study the $\ell_0$ sparse approximation problem with the tool of deep learning, by proposing Deep $\ell_0$ Encoders.

Learning A Task-Specific Deep Architecture For Clustering

no code implementations1 Sep 2015 Zhangyang Wang, Shiyu Chang, Jiayu Zhou, Meng Wang, Thomas S. Huang

In this paper, we propose to emulate the sparse coding-based clustering pipeline in the context of deep learning, leading to a carefully crafted deep model benefiting from both.

Clustering

$\mathbf{D^3}$: Deep Dual-Domain Based Fast Restoration of JPEG-Compressed Images

no code implementations16 Jan 2016 Zhangyang Wang, Ding Liu, Shiyu Chang, Qing Ling, Yingzhen Yang, Thomas S. Huang

In this paper, we design a Deep Dual-Domain ($\mathbf{D^3}$) based fast restoration model to remove artifacts of JPEG compressed images.

Learning A Deep $\ell_\infty$ Encoder for Hashing

no code implementations6 Apr 2016 Zhangyang Wang, Yingzhen Yang, Shiyu Chang, Qing Ling, Thomas S. Huang

We investigate the $\ell_\infty$-constrained representation which demonstrates robustness to quantization errors, utilizing the tool of deep learning.

Quantization

D3: Deep Dual-Domain Based Fast Restoration of JPEG-Compressed Images

no code implementations CVPR 2016 Zhangyang Wang, Ding Liu, Shiyu Chang, Qing Ling, Yingzhen Yang, Thomas S. Huang

In this paper, we design a Deep Dual-Domain (D3) based fast restoration model to remove artifacts of JPEG compressed images.

UnitBox: An Advanced Object Detection Network

no code implementations4 Aug 2016 Jiahui Yu, Yuning Jiang, Zhangyang Wang, Zhimin Cao, Thomas Huang

In present object detection systems, the deep convolutional neural networks (CNNs) are utilized to predict bounding boxes of object candidates, and have gained performance advantages over the traditional region proposal methods.

Face Detection Object +3

Stacked Approximated Regression Machine: A Simple Deep Learning Approach

no code implementations14 Aug 2016 Zhangyang Wang, Shiyu Chang, Qing Ling, Shuai Huang, Xia Hu, Honghui Shi, Thomas S. Huang

With the agreement of my coauthors, I Zhangyang Wang would like to withdraw the manuscript "Stacked Approximated Regression Machine: A Simple Deep Learning Approach".

regression

Deep Double Sparsity Encoder: Learning to Sparsify Not Only Features But Also Parameters

no code implementations23 Aug 2016 Zhangyang Wang, Thomas S. Huang

This paper emphasizes the significance to jointly exploit the problem structure and the parameter structure, in the context of deep modeling.

Dictionary Learning

When Image Denoising Meets High-Level Vision Tasks: A Deep Learning Approach

2 code implementations14 Jun 2017 Ding Liu, Bihan Wen, Xianming Liu, Zhangyang Wang, Thomas S. Huang

Conventionally, image denoising and high-level vision tasks are handled separately in computer vision.

Image Denoising

An All-in-One Network for Dehazing and Beyond

2 code implementations20 Jul 2017 Boyi Li, Xiulian Peng, Zhangyang Wang, Jizheng Xu, Dan Feng

This paper proposes an image dehazing model built with a convolutional neural network (CNN), called All-in-One Dehazing Network (AOD-Net).

Image Dehazing object-detection +2

Robust Emotion Recognition from Low Quality and Low Bit Rate Video: A Deep Learning Approach

no code implementations10 Sep 2017 Bowen Cheng, Zhangyang Wang, Zhaobin Zhang, Zhu Li, Ding Liu, Jianchao Yang, Shuai Huang, Thomas S. Huang

Emotion recognition from facial expressions is tremendously useful, especially when coupled with smart devices and wireless multimedia applications.

Decoder Emotion Recognition +1

End-to-End United Video Dehazing and Detection

no code implementations12 Sep 2017 Boyi Li, Xiulian Peng, Zhangyang Wang, Jizheng Xu, Dan Feng

Furthermore, we build an End-to-End United Video Dehazing and Detection Network(EVDD-Net), which concatenates and jointly trains EVD-Net with a video object detection model.

Image Dehazing object-detection +1

AOD-Net: All-In-One Dehazing Network

1 code implementation ICCV 2017 Boyi Li, Xiulian Peng, Zhangyang Wang, Jizheng Xu, Dan Feng

This paper proposes an image dehazing model built with a convolutional neural network (CNN), called All-in-One Dehazing Network (AOD-Net).

Image Dehazing object-detection +2

Robust Video Super-Resolution With Learned Temporal Dynamics

no code implementations ICCV 2017 Ding Liu, Zhaowen Wang, Yuchen Fan, Xian-Ming Liu, Zhangyang Wang, Shiyu Chang, Thomas Huang

Second, we reduce the complexity of motion between neighboring frames using a spatial alignment network that is much more robust and efficient than competing alignment methods and can be jointly trained with the temporal adaptive network in an end-to-end manner.

Relation Video Super-Resolution

Predicting Depression Severity by Multi-Modal Feature Engineering and Fusion

no code implementations29 Nov 2017 Aven Samareh, Yan Jin, Zhangyang Wang, Xiangyu Chang, Shuai Huang

We present our preliminary work to determine if patient's vocal acoustic, linguistic, and facial patterns could predict clinical ratings of depression severity, namely Patient Health Questionnaire depression scale (PHQ-8).

Feature Engineering

Benchmarking Single Image Dehazing and Beyond

1 code implementation12 Dec 2017 Boyi Li, Wenqi Ren, Dengpan Fu, DaCheng Tao, Dan Feng, Wen-Jun Zeng, Zhangyang Wang

We present a comprehensive study and evaluation of existing single image dehazing algorithms, using a new large-scale benchmark consisting of both synthetic and real-world hazy images, called REalistic Single Image DEhazing (RESIDE).

Benchmarking Image Dehazing +1

Enhance Visual Recognition under Adverse Conditions via Deep Networks

no code implementations20 Dec 2017 Ding Liu, Bowen Cheng, Zhangyang Wang, Haichao Zhang, Thomas S. Huang

Visual recognition under adverse conditions is a very important and challenging problem of high practical value, due to the ubiquitous existence of quality distortions during image acquisition, transmission, or storage.

Data Augmentation Image Restoration +3

Subspace Network: Deep Multi-Task Censored Regression for Modeling Neurodegenerative Diseases

1 code implementation ICLR 2018 Mengying Sun, Inci M. Baytas, Liang Zhan, Zhangyang Wang, Jiayu Zhou

Over the past decade a wide spectrum of machine learning models have been developed to model the neurodegenerative diseases, associating biomarkers, especially non-intrusive neuroimaging markers, with key clinical scores measuring the cognitive status of patients.

Multi-Task Learning regression

Frank-Wolfe Network: An Interpretable Deep Structure for Non-Sparse Coding

1 code implementation28 Feb 2018 Dong Liu, Ke Sun, Zhangyang Wang, Runsheng Liu, Zheng-Jun Zha

We propose an interpretable deep structure namely Frank-Wolfe Network (F-W Net), whose architecture is inspired by unrolling and truncating the Frank-Wolfe algorithm for solving an $L_p$-norm constrained problem with $p\geq 1$.

Handwritten Digit Recognition Image Denoising +2

Learning Simple Thresholded Features with Sparse Support Recovery

no code implementations16 Apr 2018 Hongyu Xu, Zhangyang Wang, Haichuan Yang, Ding Liu, Ji Liu

The thresholded feature has recently emerged as an extremely efficient, yet rough empirical approximation, of the time-consuming sparse coding inference process.

Dictionary Learning

Deep $k$-Means: Re-Training and Parameter Sharing with Harder Cluster Assignments for Compressing Deep Convolutions

1 code implementation24 Jun 2018 Junru Wu, Yue Wang, Zhen-Yu Wu, Zhangyang Wang, Ashok Veeraraghavan, Yingyan Lin

The current trend of pushing CNNs deeper with convolutions has created a pressing demand to achieve higher compression gains on CNNs where convolutions dominate the computation and parameter amount (e. g., GoogLeNet, ResNet and Wide ResNet).

Clustering

Improved Techniques for Learning to Dehaze and Beyond: A Collective Study

1 code implementation30 Jun 2018 Yu Liu, Guanlong Zhao, Boyuan Gong, Yang Li, Ritu Raj, Niraj Goel, Satya Kesav, Sandeep Gottimukkala, Zhangyang Wang, Wenqi Ren, DaCheng Tao

Here we explore two related but important tasks based on the recently released REalistic Single Image DEhazing (RESIDE) benchmark dataset: (i) single image dehazing as a low-level image restoration problem; and (ii) high-level visual understanding (e. g., object detection) of hazy images.

Image Dehazing Image Restoration +4

Deep k-Means: Re-Training and Parameter Sharing with Harder Cluster Assignments for Compressing Deep Convolutions

1 code implementation ICML 2018 Junru Wu, Yue Wang, Zhen-Yu Wu, Zhangyang Wang, Ashok Veeraraghavan, Yingyan Lin

The current trend of pushing CNNs deeper with convolutions has created a pressing demand to achieve higher compression gains on CNNs where convolutions dominate the computation and parameter amount (e. g., GoogLeNet, ResNet and Wide ResNet).

Clustering

Towards Privacy-Preserving Visual Recognition via Adversarial Training: A Pilot Study

3 code implementations ECCV 2018 Zhen-Yu Wu, Zhangyang Wang, Zhaowen Wang, Hailin Jin

This paper aims to improve privacy-preserving visual recognition, an increasingly demanded feature in smart camera applications, by formulating a unique adversarial training framework.

Action Recognition Privacy Preserving +1

DADA: Deep Adversarial Data Augmentation for Extremely Low Data Regime Classification

2 code implementations29 Aug 2018 Xiaofeng Zhang, Zhangyang Wang, Dong Liu, Qing Ling

Given insufficient data, while many techniques have been developed to help combat overfitting, the challenge remains if one tries to train deep networks, especially in the ill-posed extremely low data regimes: only a small set of labeled data are available, and nothing -- including unlabeled data -- else.

Data Augmentation General Classification +2

Connecting Image Denoising and High-Level Vision Tasks via Deep Learning

1 code implementation6 Sep 2018 Ding Liu, Bihan Wen, Jianbo Jiao, Xian-Ming Liu, Zhangyang Wang, Thomas S. Huang

Second we propose a deep neural network solution that cascades two modules for image denoising and various high-level tasks, respectively, and use the joint loss for updating only the denoising network via back-propagation.

Image Denoising Vocal Bursts Intensity Prediction

EnergyNet: Energy-Efficient Dynamic Inference

no code implementations NIPS Workshop CDNNRIA 2018 Yue Wang, Tan Nguyen, Yang Zhao, Zhangyang Wang, Yingyan Lin, Richard Baraniuk

The prohibitive energy cost of running high-performance Convolutional Neural Networks (CNNs) has been limiting their deployment on resource-constrained platforms including mobile and wearable devices.

Can We Gain More from Orthogonality Regularizations in Training Deep CNNs?

1 code implementation NeurIPS 2018 Nitin Bansal, Xiaohan Chen, Zhangyang Wang

This paper seeks to answer the question: as the (near-) orthogonality of weights is found to be a favorable property for training deep convolutional neural networks, how can we enforce it in more effective and easy-to-use ways?

Can We Gain More from Orthogonality Regularizations in Training Deep Networks?

1 code implementation NeurIPS 2018 Nitin Bansal, Xiaohan Chen, Zhangyang Wang

This paper seeks to answer the question: as the (near-) orthogonality of weights is found to be a favorable property for training deep convolutional neural networks, how can we enforce it in more effective and easy-to-use ways?

Adaptive Activity Monitoring with Uncertainty Quantification in Switching Gaussian Process Models

no code implementations8 Jan 2019 Randy Ardywibowo, Guang Zhao, Zhangyang Wang, Bobak Mortazavi, Shuai Huang, Xiaoning Qian

This power-efficient sensing scheme can be achieved by deciding which group of sensors to use at a given time, requiring an accurate characterization of the trade-off between sensor energy usage and the uncertainty in ignoring certain sensor signals while monitoring.

Gaussian Processes Human Activity Recognition +1

Model Compression with Adversarial Robustness: A Unified Optimization Framework

2 code implementations NeurIPS 2019 Shupeng Gui, Haotao Wang, Chen Yu, Haichuan Yang, Zhangyang Wang, Ji Liu

Deep model compression has been extensively studied, and state-of-the-art methods can now achieve high compression ratios with minimal accuracy loss.

Adversarial Robustness Model Compression +1

Single Image Deraining: A Comprehensive Benchmark Analysis

1 code implementation CVPR 2019 Siyuan Li, Iago Breno Araujo, Wenqi Ren, Zhangyang Wang, Eric K. Tokuda, Roberto Hirata Junior, Roberto Cesar-Junior, Jiawan Zhang, Xiaojie Guo, Xiaochun Cao

We present a comprehensive study and evaluation of existing single image deraining algorithms, using a new large-scale benchmark consisting of both synthetic and real-world rainy images. This dataset highlights diverse data sources and image contents, and is divided into three subsets (rain streak, rain drop, rain and mist), each serving different training or evaluation purposes.

Single Image Deraining

UG$^{2+}$ Track 2: A Collective Benchmark Effort for Evaluating and Advancing Image Understanding in Poor Visibility Environments

no code implementations9 Apr 2019 Ye Yuan, Wenhan Yang, Wenqi Ren, Jiaying Liu, Walter J. Scheirer, Zhangyang Wang

The UG$^{2+}$ challenge in IEEE CVPR 2019 aims to evoke a comprehensive discussion and exploration about how low-level vision techniques can benefit the high-level automatic visual recognition in various scenarios.

Face Detection

ALISTA: Analytic Weights Are As Good As Learned Weights in LISTA

no code implementations ICLR 2019 Jialin Liu, Xiaohan Chen, Zhangyang Wang, Wotao Yin

In this work, we propose Analytic LISTA (ALISTA), where the weight matrix in LISTA is computed as the solution to a data-free optimization problem, leaving only the stepsize and threshold parameters to data-driven learning.

Controllable Artistic Text Style Transfer via Shape-Matching GAN

1 code implementation ICCV 2019 Shuai Yang, Zhangyang Wang, Zhaowen Wang, Ning Xu, Jiaying Liu, Zongming Guo

In this paper, we present the first text style transfer network that allows for real-time control of the crucial stylistic degree of the glyph through an adjustable parameter.

Style Transfer Text Style Transfer

Plug-and-Play Methods Provably Converge with Properly Trained Denoisers

1 code implementation14 May 2019 Ernest K. Ryu, Jialin Liu, Sicheng Wang, Xiaohan Chen, Zhangyang Wang, Wotao Yin

Plug-and-play (PnP) is a non-convex framework that integrates modern denoising priors, such as BM3D or deep learning-based denoisers, into ADMM or other proximal algorithms.

Denoising

Predicting Model Failure using Saliency Maps in Autonomous Driving Systems

1 code implementation19 May 2019 Sina Mohseni, Akshay Jagadeesh, Zhangyang Wang

While machine learning systems show high success rate in many complex tasks, research shows they can also fail in very unexpected situations.

Autonomous Driving BIG-bench Machine Learning +1

Segmentation-Aware Image Denoising without Knowing True Segmentation

2 code implementations22 May 2019 Sicheng Wang, Bihan Wen, Junru Wu, DaCheng Tao, Zhangyang Wang

Several recent works discussed application-driven image restoration neural networks, which are capable of not only removing noise in images but also preserving their semantic-aware details, making them suitable for various high-level computer vision tasks as the pre-processing step.

Image Denoising Image Restoration +2

Privacy-Preserving Deep Action Recognition: An Adversarial Learning Framework and A New Dataset

5 code implementations12 Jun 2019 Zhen-Yu Wu, Haotao Wang, Zhaowen Wang, Hailin Jin, Zhangyang Wang

We first discuss an innovative heuristic of cross-dataset training and evaluation, enabling the use of multiple single-task datasets (one with target task labels and the other with privacy labels) in our problem.

Action Recognition Privacy Preserving +1

EnlightenGAN: Deep Light Enhancement without Paired Supervision

8 code implementations17 Jun 2019 Yifan Jiang, Xinyu Gong, Ding Liu, Yu Cheng, Chen Fang, Xiaohui Shen, Jianchao Yang, Pan Zhou, Zhangyang Wang

Deep learning-based methods have achieved remarkable success in image restoration and enhancement, but are they still competitive when there is a lack of paired training data?

Generative Adversarial Network Image Restoration +1

Dual Dynamic Inference: Enabling More Efficient, Adaptive and Controllable Deep Inference

no code implementations10 Jul 2019 Yue Wang, Jianghao Shen, Ting-Kuei Hu, Pengfei Xu, Tan Nguyen, Richard Baraniuk, Zhangyang Wang, Yingyan Lin

State-of-the-art convolutional neural networks (CNNs) yield record-breaking predictive performance, yet at the cost of high-energy-consumption inference, that prohibits their widely deployments in resource-constrained Internet of Things (IoT) applications.

DeblurGAN-v2: Deblurring (Orders-of-Magnitude) Faster and Better

6 code implementations ICCV 2019 Orest Kupyn, Tetiana Martyniuk, Junru Wu, Zhangyang Wang

We present a new end-to-end generative adversarial network (GAN) for single image motion deblurring, named DeblurGAN-v2, which considerably boosts state-of-the-art deblurring efficiency, quality, and flexibility.

Blind Face Restoration Generative Adversarial Network +4

Universal Safeguarded Learned Convex Optimization with Guaranteed Convergence

no code implementations25 Sep 2019 Howard Heaton, Xiaohan Chen, Zhangyang Wang, Wotao Yin

Inferences by each network form solution estimates, and networks are trained to optimize these estimates for a particular distribution of data.

Neural Networks for Principal Component Analysis: A New Loss Function Provably Yields Ordered Exact Eigenvectors

no code implementations25 Sep 2019 Reza Oftadeh, Jiayi Shen, Zhangyang Wang, Dylan Shell

In this paper, we propose a new loss function for performing principal component analysis (PCA) using linear autoencoders (LAEs).

Decoder

Drawing Early-Bird Tickets: Towards More Efficient Training of Deep Networks

2 code implementations26 Sep 2019 Haoran You, Chaojian Li, Pengfei Xu, Yonggan Fu, Yue Wang, Xiaohan Chen, Richard G. Baraniuk, Zhangyang Wang, Yingyan Lin

In this paper, we discover for the first time that the winning tickets can be identified at the very early training stage, which we term as early-bird (EB) tickets, via low-cost training schemes (e. g., early stopping and low-precision training) at large learning rates.

E2-Train: Training State-of-the-art CNNs with Over 80% Energy Savings

no code implementations NeurIPS 2019 Yue Wang, Ziyu Jiang, Xiaohan Chen, Pengfei Xu, Yang Zhao, Yingyan Lin, Zhangyang Wang

Extensive simulations and ablation studies, with real energy measurements from an FPGA board, confirm the superiority of our proposed strategies and demonstrate remarkable energy savings for training.

Learning to Optimize in Swarms

1 code implementation NeurIPS 2019 Yue Cao, Tianlong Chen, Zhangyang Wang, Yang shen

Learning to optimize has emerged as a powerful framework for various optimization and machine learning tasks.

Calibrated Domain-Invariant Learning for Highly Generalizable Large Scale Re-Identification

1 code implementation26 Nov 2019 Ye Yuan, Wuyang Chen, Tianlong Chen, Yang Yang, Zhou Ren, Zhangyang Wang, Gang Hua

Many real-world applications, such as city-scale traffic monitoring and control, requires large-scale re-identification.

DAVID: Dual-Attentional Video Deblurring

no code implementations7 Dec 2019 Junru Wu, Xiang Yu, Ding Liu, Manmohan Chandraker, Zhangyang Wang

To train and evaluate on more diverse blur severity levels, we propose a Challenging DVD dataset generated from the raw DVD video set by pooling frames with different temporal windows.

Deblurring

In Defense of the Triplet Loss Again: Learning Robust Person Re-Identification with Fast Approximated Triplet Loss and Label Distillation

1 code implementation17 Dec 2019 Ye Yuan, Wuyang Chen, Yang Yang, Zhangyang Wang

This work addresses the above two shortcomings of triplet loss, extending its effectiveness to large-scale ReID datasets with potentially noisy labels.

Person Re-Identification

FasterSeg: Searching for Faster Real-time Semantic Segmentation

2 code implementations ICLR 2020 Wuyang Chen, Xinyu Gong, Xian-Ming Liu, Qian Zhang, Yuan Li, Zhangyang Wang

We present FasterSeg, an automatically designed semantic segmentation network with not only state-of-the-art performance but also faster speed than current methods.

Neural Architecture Search Real-Time Semantic Segmentation +1

Explainable Deep Relational Networks for Predicting Compound-Protein Affinities and Contacts

no code implementations29 Dec 2019 Mostafa Karimi, Di wu, Zhangyang Wang, Yang shen

DeepRelations shows superior interpretability to the state-of-the-art: without compromising affinity prediction, it boosts the AUPRC of contact prediction 9. 5, 16. 9, 19. 3 and 5. 7-fold for the test, compound-unique, protein-unique, and both-unique sets, respectively.

BIG-bench Machine Learning Drug Discovery +1

Fractional Skipping: Towards Finer-Grained Dynamic CNN Inference

1 code implementation3 Jan 2020 Jianghao Shen, Yonggan Fu, Yue Wang, Pengfei Xu, Zhangyang Wang, Yingyan Lin

The core idea of DFS is to hypothesize layer-wise quantization (to different bitwidths) as intermediate "soft" choices to be made between fully utilizing and skipping a layer.

Quantization

Deep Plastic Surgery: Robust and Controllable Image Editing with Human-Drawn Sketches

1 code implementation ECCV 2020 Shuai Yang, Zhangyang Wang, Jiaying Liu, Zongming Guo

We present a sketch refinement strategy, as inspired by the coarse-to-fine drawing process of the artists, which we show can help our model well adapt to casual and varied sketches without the need for real sketch training data.

VGAI: End-to-End Learning of Vision-Based Decentralized Controllers for Robot Swarms

no code implementations6 Feb 2020 Ting-Kuei Hu, Fernando Gama, Tianlong Chen, Zhangyang Wang, Alejandro Ribeiro, Brian M. Sadler

More specifically, we consider that each robot has access to a visual perception of the immediate surroundings, and communication capabilities to transmit and receive messages from other neighboring robots.

Triple Wins: Boosting Accuracy, Robustness and Efficiency Together by Enabling Input-Adaptive Inference

2 code implementations ICLR 2020 Ting-Kuei Hu, Tianlong Chen, Haotao Wang, Zhangyang Wang

Deep networks were recently suggested to face the odds between accuracy (on clean natural images) and robustness (on adversarially perturbed images) (Tsipras et al., 2019).

I Am Going MAD: Maximum Discrepancy Competition for Comparing Classifiers Adaptively

1 code implementation ICLR 2020 Haotao Wang, Tianlong Chen, Zhangyang Wang, Kede Ma

On the other hand, the trained classifiers have traditionally been evaluated on small and fixed sets of test images, which are deemed to be extremely sparsely distributed in the space of all natural images.

Image Classification

Uncertainty Quantification for Deep Context-Aware Mobile Activity Recognition and Unknown Context Discovery

no code implementations3 Mar 2020 Zepeng Huo, Arash Pakbin, Xiaohan Chen, Nathan Hurley, Ye Yuan, Xiaoning Qian, Zhangyang Wang, Shuai Huang, Bobak Mortazavi

Activity recognition in wearable computing faces two key challenges: i) activity characteristics may be context-dependent and change under different contexts or situations; ii) unknown contexts and activities may occur from time to time, requiring flexibility and adaptability of the algorithm.

Clustering Human Activity Recognition +1

Safeguarded Learned Convex Optimization

no code implementations4 Mar 2020 Howard Heaton, Xiaohan Chen, Zhangyang Wang, Wotao Yin

Our numerical examples show convergence of Safe-L2O algorithms, even when the provided data is not from the distribution of training data.

Adversarial Robustness: From Self-Supervised Pre-Training to Fine-Tuning

1 code implementation CVPR 2020 Tianlong Chen, Sijia Liu, Shiyu Chang, Yu Cheng, Lisa Amini, Zhangyang Wang

We conduct extensive experiments to demonstrate that the proposed framework achieves large performance margins (eg, 3. 83% on robust accuracy and 1. 3% on standard accuracy, on the CIFAR-10 dataset), compared with the conventional end-to-end adversarial training baseline.

Adversarial Robustness

L^2-GCN: Layer-Wise and Learned Efficient Training of Graph Convolutional Networks

2 code implementations CVPR 2020 Yuning You, Tianlong Chen, Zhangyang Wang, Yang shen

Graph convolution networks (GCN) are increasingly popular in many applications, yet remain notoriously hard to train over large graph datasets.

Drawing Early-Bird Tickets: Toward More Efficient Training of Deep Networks

1 code implementation ICLR 2020 Haoran You, Chaojian Li, Pengfei Xu, Yonggan Fu, Yue Wang, Xiaohan Chen, Richard G. Baraniuk, Zhangyang Wang, Yingyan Lin

Finally, we leverage the existence of EB tickets and the proposed mask distance to develop efficient training methods, which are achieved by first identifying EB tickets via low-cost schemes, and then continuing to train merely the EB tickets towards the target accuracy.

SmartExchange: Trading Higher-cost Memory Storage/Access for Lower-cost Computation

no code implementations7 May 2020 Yang Zhao, Xiaohan Chen, Yue Wang, Chaojian Li, Haoran You, Yonggan Fu, Yuan Xie, Zhangyang Wang, Yingyan Lin

We present SmartExchange, an algorithm-hardware co-design framework to trade higher-cost memory storage/access for lower-cost computation, for energy-efficient inference of deep neural networks (DNNs).

Model Compression Quantization

AutoSpeech: Neural Architecture Search for Speaker Recognition

3 code implementations7 May 2020 Shaojin Ding, Tianlong Chen, Xinyu Gong, Weiwei Zha, Zhangyang Wang

Speaker recognition systems based on Convolutional Neural Networks (CNNs) are often built with off-the-shelf backbones such as VGG-Net or ResNet.

Ranked #6 on Speaker Identification on VoxCeleb1 (using extra training data)

Image Classification Neural Architecture Search +3

NADS: Neural Architecture Distribution Search for Uncertainty Awareness

no code implementations ICML 2020 Randy Ardywibowo, Shahin Boluki, Xinyu Gong, Zhangyang Wang, Xiaoning Qian

Machine learning (ML) systems often encounter Out-of-Distribution (OoD) errors when dealing with testing data coming from a distribution different from training data.

Out of Distribution (OOD) Detection

AutoGAN-Distiller: Searching to Compress Generative Adversarial Networks

3 code implementations ICML 2020 Yonggan Fu, Wuyang Chen, Haotao Wang, Haoran Li, Yingyan Lin, Zhangyang Wang

Inspired by the recent success of AutoML in deep compression, we introduce AutoML to GAN compression and develop an AutoGAN-Distiller (AGD) framework.

AutoML Knowledge Distillation +2

When Does Self-Supervision Help Graph Convolutional Networks?

1 code implementation ICML 2020 Yuning You, Tianlong Chen, Zhangyang Wang, Yang shen

We first elaborate three mechanisms to incorporate self-supervision into GCNs, analyze the limitations of pretraining & finetuning and self-training, and proceed to focus on multi-task learning.

Multi-Task Learning Representation Learning +1

Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training

1 code implementation ICML 2020 Xuxi Chen, Wuyang Chen, Tianlong Chen, Ye Yuan, Chen Gong, Kewei Chen, Zhangyang Wang

Many real-world applications have to tackle the Positive-Unlabeled (PU) learning problem, i. e., learning binary classifiers from a large amount of unlabeled data and a few labeled positive examples.

Can 3D Adversarial Logos Cloak Humans?

1 code implementation25 Jun 2020 Yi Wang, Jingyang Zhou, Tianlong Chen, Sijia Liu, Shiyu Chang, Chandrajit Bajaj, Zhangyang Wang

Contrary to the traditional adversarial patch, this new form of attack is mapped into the 3D object world and back-propagates to the 2D image domain through differentiable rendering.

Object

GAN Slimming: All-in-One GAN Compression by A Unified Optimization Framework

2 code implementations ECCV 2020 Haotao Wang, Shupeng Gui, Haichuan Yang, Ji Liu, Zhangyang Wang

Generative adversarial networks (GANs) have gained increasing popularity in various computer vision applications, and recently start to be deployed to resource-constrained mobile devices.

Image-to-Image Translation Quantization +1

PCAL: A Privacy-preserving Intelligent Credit Risk Modeling Framework Based on Adversarial Learning

no code implementations6 Oct 2020 Yuli Zheng, Zhenyu Wu, Ye Yuan, Tianlong Chen, Zhangyang Wang

While machine learning is increasingly used in this field, the resulting large-scale collection of user private information has reinvigorated the privacy debate, considering dozens of data breach incidents every year caused by unauthorized hackers, and (potentially even more) information misuse/abuse by authorized parties.

BIG-bench Machine Learning Privacy Preserving

Training Stronger Baselines for Learning to Optimize

1 code implementation NeurIPS 2020 Tianlong Chen, Weiyi Zhang, Jingyang Zhou, Shiyu Chang, Sijia Liu, Lisa Amini, Zhangyang Wang

Learning to optimize (L2O) has gained increasing attention since classical optimizers require laborious problem-specific design and hyperparameter tuning.

Imitation Learning Rolling Shutter Correction

Graph Contrastive Learning with Augmentations

4 code implementations NeurIPS 2020 Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, Yang shen

In this paper, we propose a graph contrastive learning (GraphCL) framework for learning unsupervised representations of graph data.

Contrastive Learning Representation Learning +2

Robust Pre-Training by Adversarial Contrastive Learning

1 code implementation NeurIPS 2020 Ziyu Jiang, Tianlong Chen, Ting Chen, Zhangyang Wang

Recent work has shown that, when integrated with adversarial training, self-supervised pre-training can lead to state-of-the-art robustness In this work, we improve robustness-aware self-supervised pre-training by learning representations that are consistent under both data augmentations and adversarial perturbations.

Adversarial Robustness Contrastive Learning

What Does CNN Shift Invariance Look Like? A Visualization Study

1 code implementation9 Nov 2020 Jake Lee, Junfeng Yang, Zhangyang Wang

We present the results of three experiments comparing representations of millions of images with exhaustively shifted objects, examining both local invariance (within a few pixels) and global invariance (across the image frame).

Uncertainty-Aware Physically-Guided Proxy Tasks for Unseen Domain Face Anti-spoofing

no code implementations28 Nov 2020 Junru Wu, Xiang Yu, Buyu Liu, Zhangyang Wang, Manmohan Chandraker

Face anti-spoofing (FAS) seeks to discriminate genuine faces from fake ones arising from any type of spoofing attack.

Attribute Domain Generalization +1

MATE: Plugging in Model Awareness to Task Embedding for Meta Learning

1 code implementation NeurIPS 2020 Xiaohan Chen, Zhangyang Wang, Siyu Tang, Krikamol Muandet

Meta-learning improves generalization of machine learning models when faced with previously unseen tasks by leveraging experiences from different, yet related prior tasks.

feature selection Few-Shot Learning

The Lottery Tickets Hypothesis for Supervised and Self-supervised Pre-training in Computer Vision Models

1 code implementation CVPR 2021 Tianlong Chen, Jonathan Frankle, Shiyu Chang, Sijia Liu, Yang Zhang, Michael Carbin, Zhangyang Wang

We extend the scope of LTH and question whether matching subnetworks still exist in pre-trained computer vision models, that enjoy the same downstream transfer performance.

FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training

1 code implementation NeurIPS 2020 Yonggan Fu, Haoran You, Yang Zhao, Yue Wang, Chaojian Li, Kailash Gopalakrishnan, Zhangyang Wang, Yingyan Lin

Recent breakthroughs in deep neural networks (DNNs) have fueled a tremendous demand for intelligent edge devices featuring on-site learning, while the practical realization of such systems remains a challenge due to the limited resources available at the edge and the required massive training costs for state-of-the-art (SOTA) DNNs.

Quantization

Growing Deep Forests Efficiently with Soft Routing and Learned Connectivity

no code implementations29 Dec 2020 Jianghao Shen, Sicheng Wang, Zhangyang Wang

For example, our model with only 1 layer of 15 trees can perform comparably with the model in [3] with 2 layers of 2000 trees each.

EarlyBERT: Efficient BERT Training via Early-bird Lottery Tickets

1 code implementation ACL 2021 Xiaohan Chen, Yu Cheng, Shuohang Wang, Zhe Gan, Zhangyang Wang, Jingjing Liu

Heavily overparameterized language models such as BERT, XLNet and T5 have achieved impressive success in many NLP tasks.

Model Compression

Efficiently Troubleshooting Image Segmentation Models with Human-In-The-Loop

no code implementations1 Jan 2021 Haotao Wang, Tianlong Chen, Zhangyang Wang, Kede Ma

Image segmentation lays the foundation for many high-stakes vision applications such as autonomous driving and medical image analysis.

Autonomous Driving Image Segmentation +2

Weak NAS Predictor Is All You Need

no code implementations1 Jan 2021 Junru Wu, Xiyang Dai, Dongdong Chen, Yinpeng Chen, Mengchen Liu, Ye Yu, Zhangyang Wang, Zicheng Liu, Mei Chen, Lu Yuan

Rather than expecting a single strong predictor to model the whole space, we seek a progressive line of weak predictors that can connect a path to the best architecture, thus greatly simplifying the learning task of each predictor.

Neural Architecture Search

Robust Overfitting may be mitigated by properly learned smoothening

no code implementations ICLR 2021 Tianlong Chen, Zhenyu Zhang, Sijia Liu, Shiyu Chang, Zhangyang Wang

A recent study (Rice et al., 2020) revealed overfitting to be a dominant phenomenon in adversarially robust training of deep networks, and that appropriate early-stopping of adversarial training (AT) could match the performance gains of most recent algorithmic improvements.

Knowledge Distillation

Bayesian Learning to Optimize: Quantifying the Optimizer Uncertainty

no code implementations1 Jan 2021 Yue Cao, Tianlong Chen, Zhangyang Wang, Yang shen

Optimizing an objective function with uncertainty awareness is well-known to improve the accuracy and confidence of optimization solutions.

Image Classification Uncertainty Quantification +1

On Dynamic Noise Influence in Differential Private Learning

no code implementations1 Jan 2021 Junyuan Hong, Zhangyang Wang, Jiayu Zhou

In this paper, we provide comprehensive analysis of noise influence in dynamic privacy schedules to answer these critical questions.

ALFA: Adversarial Feature Augmentation for Enhanced Image Recognition

no code implementations1 Jan 2021 Tianlong Chen, Yu Cheng, Zhe Gan, Yu Hu, Zhangyang Wang, Jingjing Liu

Adversarial training is an effective method to combat adversarial attacks in order to create robust neural networks.

Learning A Minimax Optimizer: A Pilot Study

no code implementations ICLR 2021 Jiayi Shen, Xiaohan Chen, Howard Heaton, Tianlong Chen, Jialin Liu, Wotao Yin, Zhangyang Wang

We first present Twin L2O, the first dedicated minimax L2O framework consisting of two LSTMs for updating min and max variables, respectively.

SmartDeal: Re-Modeling Deep Network Weights for Efficient Inference and Training

1 code implementation4 Jan 2021 Xiaohan Chen, Yang Zhao, Yue Wang, Pengfei Xu, Haoran You, Chaojian Li, Yonggan Fu, Yingyan Lin, Zhangyang Wang

Results show that: 1) applied to inference, SD achieves up to 2. 44x energy efficiency as evaluated via real hardware implementations; 2) applied to training, SD leads to 10. 56x and 4. 48x reduction in the storage and training energy, with negligible accuracy loss compared to state-of-the-art training baselines.

Spending Your Winning Lottery Better After Drawing It

1 code implementation8 Jan 2021 Ajay Kumar Jaiswal, Haoyu Ma, Tianlong Chen, Ying Ding, Zhangyang Wang

In this paper, we demonstrate that it is unnecessary for spare retraining to strictly inherit those properties from the dense network.

Knowledge Distillation

Dynamic Privacy Budget Allocation Improves Data Efficiency of Differentially Private Gradient Descent

no code implementations19 Jan 2021 Junyuan Hong, Zhangyang Wang, Jiayu Zhou

In this paper, we provide comprehensive analysis of noise influence in dynamic privacy schedules to answer these critical questions.

A Unified Lottery Ticket Hypothesis for Graph Neural Networks

2 code implementations12 Feb 2021 Tianlong Chen, Yongduo Sui, Xuxi Chen, Aston Zhang, Zhangyang Wang

With graphs rapidly growing in size and deeper graph neural networks (GNNs) emerging, the training and inference of GNNs become increasingly expensive.

Link Prediction Node Classification

TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up

10 code implementations NeurIPS 2021 Yifan Jiang, Shiyu Chang, Zhangyang Wang

Our vanilla GAN architecture, dubbed TransGAN, consists of a memory-friendly transformer-based generator that progressively increases feature resolution, and correspondingly a multi-scale discriminator to capture simultaneously semantic contexts and low-level textures.

Data Augmentation Image Generation

Stronger NAS with Weaker Predictors

1 code implementation NeurIPS 2021 Junru Wu, Xiyang Dai, Dongdong Chen, Yinpeng Chen, Mengchen Liu, Ye Yu, Zhangyang Wang, Zicheng Liu, Mei Chen, Lu Yuan

We propose a paradigm shift from fitting the whole architecture space using one strong predictor, to progressively fitting a search path towards the high-performance sub-space through a set of weaker predictors.

Neural Architecture Search

Exposing Semantic Segmentation Failures via Maximum Discrepancy Competition

1 code implementation27 Feb 2021 Jiebin Yan, Yu Zhong, Yuming Fang, Zhangyang Wang, Kede Ma

A natural question then arises: Does the superior performance on the closed (and frequently re-used) test sets transfer to the open visual world with unconstrained variations?

Segmentation Semantic Segmentation

Data-Efficient GAN Training Beyond (Just) Augmentations: A Lottery Ticket Perspective

1 code implementation NeurIPS 2021 Tianlong Chen, Yu Cheng, Zhe Gan, Jingjing Liu, Zhangyang Wang

Training generative adversarial networks (GANs) with limited real image data generally results in deteriorated performance and collapsed models.

Data Augmentation

Adversarial Feature Augmentation and Normalization for Visual Recognition

1 code implementation22 Mar 2021 Tianlong Chen, Yu Cheng, Zhe Gan, JianFeng Wang, Lijuan Wang, Zhangyang Wang, Jingjing Liu

Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.

Classification Data Augmentation +2

UltraSR: Spatial Encoding is a Missing Key for Implicit Image Function-based Arbitrary-Scale Super-Resolution

1 code implementation23 Mar 2021 Xingqian Xu, Zhangyang Wang, Humphrey Shi

In this work, we propose UltraSR, a simple yet effective new network design based on implicit image functions in which we deeply integrated spatial coordinates and periodic encoding with the implicit neural representation.

Super-Resolution

Learning to Optimize: A Primer and A Benchmark

1 code implementation23 Mar 2021 Tianlong Chen, Xiaohan Chen, Wuyang Chen, Howard Heaton, Jialin Liu, Zhangyang Wang, Wotao Yin

It automates the design of an optimization method based on its performance on a set of training problems.

Benchmarking

The Elastic Lottery Ticket Hypothesis

1 code implementation NeurIPS 2021 Xiaohan Chen, Yu Cheng, Shuohang Wang, Zhe Gan, Jingjing Liu, Zhangyang Wang

Based on these results, we articulate the Elastic Lottery Ticket Hypothesis (E-LTH): by mindfully replicating (or dropping) and re-ordering layers for one network, its corresponding winning ticket could be stretched (or squeezed) into a subnetwork for another deeper (or shallower) network from the same family, whose performance is nearly the same competitive as the latter's winning ticket directly found by IMP.

Bootstrapping Your Own Positive Sample: Contrastive Learning With Electronic Health Record Data

no code implementations7 Apr 2021 Tingyi Wanyan, Jing Zhang, Ying Ding, Ariful Azad, Zhangyang Wang, Benjamin S Glicksberg

Electronic Health Record (EHR) data has been of tremendous utility in Artificial Intelligence (AI) for healthcare such as predicting future clinical events.

Attribute Contrastive Learning +1

Knowledge-Augmented Contrastive Learning for Abnormality Classification and Localization in Chest X-rays with Radiomics using a Feedback Loop

no code implementations11 Apr 2021 Yan Han, Chongyan Chen, Ahmed Tewfik, Benjamin Glicksberg, Ying Ding, Yifan Peng, Zhangyang Wang

The key knob of our framework is a unique positive sampling approach tailored for the medical images, by seamlessly integrating radiomic features as a knowledge augmentation.

Contrastive Learning

"BNN - BN = ?": Training Binary Neural Networks without Batch Normalization

1 code implementation16 Apr 2021 Tianlong Chen, Zhenyu Zhang, Xu Ouyang, Zechun Liu, Zhiqiang Shen, Zhangyang Wang

However, the BN layer is costly to calculate and is typically implemented with non-binary parameters, leaving a hurdle for the efficient implementation of BNN training.

Image Classification

InstantNet: Automated Generation and Deployment of Instantaneously Switchable-Precision Networks

1 code implementation22 Apr 2021 Yonggan Fu, Zhongzhi Yu, Yongan Zhang, Yifan Jiang, Chaojian Li, Yongyuan Liang, Mingchao Jiang, Zhangyang Wang, Yingyan Lin

The promise of Deep Neural Network (DNN) powered Internet of Thing (IoT) devices has motivated a tremendous demand for automated solutions to enable fast development and deployment of efficient (1) DNNs equipped with instantaneous accuracy-efficiency trade-off capability to accommodate the time-varying resources at IoT devices and (2) dataflows to optimize DNNs' execution efficiency on different devices.

Learning Transferable 3D Adversarial Cloaks for Deep Trained Detectors

1 code implementation22 Apr 2021 Arman Maesumi, Mingkang Zhu, Yi Wang, Tianlong Chen, Zhangyang Wang, Chandrajit Bajaj

This paper presents a novel patch-based adversarial attack pipeline that trains adversarial patches on 3D human meshes.

Adversarial Attack Object

SAFIN: Arbitrary Style Transfer With Self-Attentive Factorized Instance Normalization

1 code implementation13 May 2021 Aaditya Singh, Shreeshail Hingane, Xinyu Gong, Zhangyang Wang

We demonstrate that plugging SAFIN into the base network of another state-of-the-art method results in enhanced stylization.

Style Transfer

Troubleshooting Blind Image Quality Models in the Wild

no code implementations CVPR 2021 Zhihua Wang, Haotao Wang, Tianlong Chen, Zhangyang Wang, Kede Ma

Recently, the group maximum differentiation competition (gMAD) has been used to improve blind image quality assessment (BIQA) models, with the help of full-reference metrics.

Blind Image Quality Assessment Network Pruning

Undistillable: Making A Nasty Teacher That CANNOT teach students

1 code implementation ICLR 2021 Haoyu Ma, Tianlong Chen, Ting-Kuei Hu, Chenyu You, Xiaohui Xie, Zhangyang Wang

Knowledge Distillation (KD) is a widely used technique to transfer knowledge from pre-trained teacher models to (usually more lightweight) student models.

Knowledge Distillation

Improving Contrastive Learning on Imbalanced Data via Open-World Sampling

1 code implementation NeurIPS 2021 Ziyu Jiang, Tianlong Chen, Ting Chen, Zhangyang Wang

Contrastive learning approaches have achieved great success in learning visual representations with few labels of the target classes.

Attribute Contrastive Learning

Efficient Lottery Ticket Finding: Less Data is More

1 code implementation6 Jun 2021 Zhenyu Zhang, Xuxi Chen, Tianlong Chen, Zhangyang Wang

We observe that a high-quality winning ticket can be found with training and pruning the dense network on the very compact PrAC set, which can substantially save training iterations for the ticket finding process.

Self-Damaging Contrastive Learning

1 code implementation6 Jun 2021 Ziyu Jiang, Tianlong Chen, Bobak Mortazavi, Zhangyang Wang

Hence, the key innovation in SDCLR is to create a dynamic self-competitor model to contrast with the target model, which is a pruned version of the latter.

Contrastive Learning Linear evaluation +2

Chasing Sparsity in Vision Transformers: An End-to-End Exploration

1 code implementation NeurIPS 2021 Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang

For example, our sparsified DeiT-Small at (5%, 50%) sparsity for (data, architecture), improves 0. 28% top-1 accuracy, and meanwhile enjoys 49. 32% FLOPs and 4. 40% running time savings.

Efficient ViTs

Taxonomy of Machine Learning Safety: A Survey and Primer

no code implementations9 Jun 2021 Sina Mohseni, Haotao Wang, Zhiding Yu, Chaowei Xiao, Zhangyang Wang, Jay Yadawa

The open-world deployment of Machine Learning (ML) algorithms in safety-critical applications such as autonomous vehicles needs to address a variety of ML vulnerabilities such as interpretability, verifiability, and performance limitations.

Autonomous Vehicles BIG-bench Machine Learning +1

Sparse and Imperceptible Adversarial Attack via a Homotopy Algorithm

1 code implementation10 Jun 2021 Mingkang Zhu, Tianlong Chen, Zhangyang Wang

Compared to state-of-the-art methods, our homotopy attack leads to significantly fewer perturbations, e. g., reducing 42. 91% on CIFAR-10 and 75. 03% on ImageNet (average case, targeted attack), at similar maximal perturbation magnitudes, when still achieving 100% attack success rates.

Adversarial Attack

Graph Contrastive Learning Automated

2 code implementations10 Jun 2021 Yuning You, Tianlong Chen, Yang shen, Zhangyang Wang

Unfortunately, unlike its counterpart on image data, the effectiveness of GraphCL hinges on ad-hoc data augmentations, which have to be manually picked per dataset, by either rules of thumb or trial-and-errors, owing to the diverse nature of graph data.

Contrastive Learning Representation Learning +1

Federated Robustness Propagation: Sharing Robustness in Heterogeneous Federated Learning

1 code implementation18 Jun 2021 Junyuan Hong, Haotao Wang, Zhangyang Wang, Jiayu Zhou

In this paper, we study a novel FL strategy: propagating adversarial robustness from rich-resource users that can afford AT, to those with poor resources that cannot afford it, during federated learning.

Adversarial Robustness Federated Learning

Sparse Training via Boosting Pruning Plasticity with Neuroregeneration

2 code implementations NeurIPS 2021 Shiwei Liu, Tianlong Chen, Xiaohan Chen, Zahra Atashgahi, Lu Yin, Huanyu Kou, Li Shen, Mykola Pechenizkiy, Zhangyang Wang, Decebal Constantin Mocanu

Works on lottery ticket hypothesis (LTH) and single-shot network pruning (SNIP) have raised a lot of attention currently on post-training pruning (iterative magnitude pruning), and before-training pruning (pruning at initialization).

Network Pruning Sparse Learning

Scalable Perception-Action-Communication Loops with Convolutional and Graph Neural Networks

1 code implementation24 Jun 2021 Ting-Kuei Hu, Fernando Gama, Tianlong Chen, Wenqing Zheng, Zhangyang Wang, Alejandro Ribeiro, Brian M. Sadler

Our framework is implemented by a cascade of a convolutional and a graph neural network (CNN / GNN), addressing agent-level visual perception and feature learning, as well as swarm-level communication, local information aggregation and agent action inference, respectively.

Imitation Learning

Sanity Checks for Lottery Tickets: Does Your Winning Ticket Really Win the Jackpot?

2 code implementations NeurIPS 2021 Xiaolong Ma, Geng Yuan, Xuan Shen, Tianlong Chen, Xuxi Chen, Xiaohan Chen, Ning Liu, Minghai Qin, Sijia Liu, Zhangyang Wang, Yanzhi Wang

Based on our analysis, we summarize a guideline for parameter settings in regards of specific architecture characteristics, which we hope to catalyze the research progress on the topic of lottery ticket hypothesis.

DANCE: DAta-Network Co-optimization for Efficient Segmentation Model Training and Inference

no code implementations16 Jul 2021 Chaojian Li, Wuyang Chen, Yuchen Gu, Tianlong Chen, Yonggan Fu, Zhangyang Wang, Yingyan Lin

Semantic segmentation for scene understanding is nowadays widely demanded, raising significant challenges for the algorithm efficiency, especially its applications on resource-limited platforms.

Scene Understanding Segmentation +1

Black-Box Diagnosis and Calibration on GAN Intra-Mode Collapse: A Pilot Study

1 code implementation23 Jul 2021 Zhenyu Wu, Zhaowen Wang, Ye Yuan, Jianming Zhang, Zhangyang Wang, Hailin Jin

Existing diversity tests of samples from GANs are usually conducted qualitatively on a small scale, and/or depends on the access to original training data as well as the trained model parameters.

Image Generation

CERL: A Unified Optimization Framework for Light Enhancement with Realistic Noise

1 code implementation1 Aug 2021 Zeyuan Chen, Yifan Jiang, Dong Liu, Zhangyang Wang

We present \underline{C}oordinated \underline{E}nhancement for \underline{R}eal-world \underline{L}ow-light Noisy Images (CERL), that seamlessly integrates light enhancement and noise suppression parts into a unified and physics-grounded optimization framework.

Denoising

SSH: A Self-Supervised Framework for Image Harmonization

1 code implementation ICCV 2021 Yifan Jiang, He Zhang, Jianming Zhang, Yilin Wang, Zhe Lin, Kalyan Sunkavalli, Simon Chen, Sohrab Amirghodsi, Sarah Kong, Zhangyang Wang

Image harmonization aims to improve the quality of image compositing by matching the "appearance" (\eg, color tone, brightness and contrast) between foreground and background images.

Benchmarking Data Augmentation +1

Bag of Tricks for Training Deeper Graph Neural Networks: A Comprehensive Benchmark Study

1 code implementation24 Aug 2021 Tianlong Chen, Kaixiong Zhou, Keyu Duan, Wenqing Zheng, Peihao Wang, Xia Hu, Zhangyang Wang

In view of those, we present the first fair and reproducible benchmark dedicated to assessing the "tricks" of training deep GNNs.

Understanding and Accelerating Neural Architecture Search with Training-Free and Theory-Grounded Metrics

1 code implementation26 Aug 2021 Wuyang Chen, Xinyu Gong, Junru Wu, Yunchao Wei, Humphrey Shi, Zhicheng Yan, Yi Yang, Zhangyang Wang

This work targets designing a principled and unified training-free framework for Neural Architecture Search (NAS), with high performance, low cost, and in-depth interpretation.

Neural Architecture Search

Font Completion and Manipulation by Cycling Between Multi-Modality Representations

1 code implementation30 Aug 2021 Ye Yuan, Wuyang Chen, Zhaowen Wang, Matthew Fisher, Zhifei Zhang, Zhangyang Wang, Hailin Jin

The novel graph constructor maps a glyph's latent code to its graph representation that matches expert knowledge, which is trained to help the translation task.

Image-to-Image Translation Representation Learning +2

GDP: Stabilized Neural Network Pruning via Gates with Differentiable Polarization

no code implementations ICCV 2021 Yi Guo, Huan Yuan, Jianchao Tan, Zhangyang Wang, Sen yang, Ji Liu

During the training process, the polarization effect will drive a subset of gates to smoothly decrease to exact zero, while other gates gradually stay away from zero by a large margin.

Model Compression Network Pruning

Generalizable Learning to Optimize into Wide Valleys

no code implementations29 Sep 2021 Junjie Yang, Tianlong Chen, Mingkang Zhu, Fengxiang He, DaCheng Tao, Yingbin Liang, Zhangyang Wang

Learning to optimize (L2O) has gained increasing popularity in various optimization tasks, since classical optimizers usually require laborious, problem-specific design and hyperparameter tuning.

CheXT: Knowledge-Guided Cross-Attention Transformer for Abnormality Classification and Localization in Chest X-rays

no code implementations29 Sep 2021 Yan Han, Ying Ding, Ahmed Tewfik, Yifan Peng, Zhangyang Wang

During training, the image branch leverages its learned attention to estimate pathology localization, which is then utilized to extract radiomic features from images in the radiomics branch.

Peek-a-Boo: What (More) is Disguised in a Randomly Weighted Neural Network, and How to Find It Efficiently

no code implementations ICLR 2022 Xiaohan Chen, Jason Zhang, Zhangyang Wang

In this work, we define an extended class of subnetworks in randomly initialized NNs called disguised subnetworks, which are not only "hidden" in the random networks but also "disguised" -- hence can only be "unmasked" with certain transformations on weights.

Lottery Image Prior

no code implementations29 Sep 2021 Qiming Wu, Xiaohan Chen, Yifan Jiang, Pan Zhou, Zhangyang Wang

Drawing inspirations from the recently prosperous research on lottery ticket hypothesis (LTH), we conjecture and study a novel “lottery image prior” (LIP), stated as: given an (untrained or trained) DNN-based image prior, it will have a sparse subnetwork that can be training in isolation, to match the original DNN’s performance when being applied as a prior to various image inverse problems.

Compressive Sensing Image Reconstruction +1

Equalized Robustness: Towards Sustainable Fairness Under Distributional Shifts

no code implementations29 Sep 2021 Haotao Wang, Junyuan Hong, Jiayu Zhou, Zhangyang Wang

In this paper, we first propose a new fairness goal, termed Equalized Robustness (ER), to impose fair model robustness against unseen distribution shifts across majority and minority groups.

Fairness

Scaling the Depth of Vision Transformers via the Fourier Domain Analysis

no code implementations ICLR 2022 Peihao Wang, Wenqing Zheng, Tianlong Chen, Zhangyang Wang

The first technique, termed AttnScale, decomposes a self-attention block into low-pass and high-pass components, then rescales and combines these two filters to produce an all-pass self-attention matrix.

Audio Lottery: Speech Recognition Made Ultra-Lightweight, Noise-Robust, and Transferable

no code implementations ICLR 2022 Shaojin Ding, Tianlong Chen, Zhangyang Wang

In this paper, we investigate the tantalizing possibility of using lottery ticket hypothesis to discover lightweight speech recognition models, that are (1) robust to various noise existing in speech; (2) transferable to fit the open-world personalization; and 3) compatible with structured sparsity.

speech-recognition Speech Recognition

Universality of Deep Neural Network Lottery Tickets: A Renormalization Group Perspective

no code implementations29 Sep 2021 William T Redman, Tianlong Chen, Akshunna S. Dogra, Zhangyang Wang

Foundational work on the Lottery Ticket Hypothesis has suggested an exciting corollary: winning tickets found in the context of one task can be transferred to similar tasks, possibly even across different architectures.

AutoCoG: A Unified Data-Modal Co-Search Framework for Graph Neural Networks

no code implementations29 Sep 2021 Duc N.M Hoang, Kaixiong Zhou, Tianlong Chen, Xia Hu, Zhangyang Wang

Despite the preliminary success, we argue that for GNNs, NAS has to be customized further, due to the topological complicacy of GNN input data (graph) as well as the notorious training instability.

Data Augmentation Language Modelling +1

Stingy Teacher: Sparse Logits Suffice to Fail Knowledge Distillation

no code implementations29 Sep 2021 Haoyu Ma, Yifan Huang, Tianlong Chen, Hao Tang, Chenyu You, Zhangyang Wang, Xiaohui Xie

However, it is unclear why the distorted distribution of the logits is catastrophic to the student model.

Knowledge Distillation

Lottery Tickets can have Structural Sparsity

no code implementations29 Sep 2021 Tianlong Chen, Xuxi Chen, Xiaolong Ma, Yanzhi Wang, Zhangyang Wang

The lottery ticket hypothesis (LTH) has shown that dense models contain highly sparse subnetworks (i. e., $\textit{winning tickets}$) that can be trained in isolation to match full accuracy.

Learning Pruning-Friendly Networks via Frank-Wolfe: One-Shot, Any-Sparsity, And No Retraining

1 code implementation ICLR 2022 Lu Miao, Xiaolong Luo, Tianlong Chen, Wuyang Chen, Dong Liu, Zhangyang Wang

Conventional methods often require (iterative) pruning followed by re-training, which not only incurs large overhead beyond the original DNN training but also can be sensitive to retraining hyperparameters.

Universality of Winning Tickets: A Renormalization Group Perspective

no code implementations7 Oct 2021 William T. Redman, Tianlong Chen, Zhangyang Wang, Akshunna S. Dogra

Foundational work on the Lottery Ticket Hypothesis has suggested an exciting corollary: winning tickets found in the context of one task can be transferred to similar tasks, possibly even across different architectures.

Towards Lifelong Learning of Multilingual Text-To-Speech Synthesis

1 code implementation9 Oct 2021 Mu Yang, Shaojin Ding, Tianlong Chen, Tong Wang, Zhangyang Wang

This work presents a lifelong learning approach to train a multilingual Text-To-Speech (TTS) system, where each language was seen as an individual task and was learned sequentially and continually.

Speech Synthesis Text-To-Speech Synthesis

Hyperparameter Tuning is All You Need for LISTA

1 code implementation NeurIPS 2021 Xiaohan Chen, Jialin Liu, Zhangyang Wang, Wotao Yin

Learned Iterative Shrinkage-Thresholding Algorithm (LISTA) introduces the concept of unrolling an iterative algorithm and training it like a neural network.

Rolling Shutter Correction

Delayed Propagation Transformer: A Universal Computation Engine towards Practical Control in Cyber-Physical Systems

1 code implementation NeurIPS 2021 Wenqing Zheng, Qiangqiang Guo, Hao Yang, Peihao Wang, Zhangyang Wang

This paper presents the Delayed Propagation Transformer (DePT), a new transformer-based model that specializes in the global modeling of CPS while taking into account the immutable constraints from the physical world.

Inductive Bias

DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models

1 code implementation30 Oct 2021 Xuxi Chen, Tianlong Chen, Weizhu Chen, Ahmed Hassan Awadallah, Zhangyang Wang, Yu Cheng

To address these pain points, we propose a framework for resource- and parameter-efficient fine-tuning by leveraging the sparsity prior in both weight updates and the final model weights.

You are caught stealing my winning lottery ticket! Making a lottery ticket claim its ownership

1 code implementation NeurIPS 2021 Xuxi Chen, Tianlong Chen, Zhenyu Zhang, Zhangyang Wang

The lottery ticket hypothesis (LTH) emerges as a promising framework to leverage a special sparse subnetwork (i. e., winning ticket) instead of a full model for both training and inference, that can lower both costs without sacrificing the performance.

Improving Contrastive Learning on Imbalanced Seed Data via Open-World Sampling

1 code implementation NeurIPS 2021 Ziyu Jiang, Tianlong Chen, Ting Chen, Zhangyang Wang

Contrastive learning approaches have achieved great success in learning visual representations with few labels of the target classes.

Attribute Contrastive Learning

A Simple Single-Scale Vision Transformer for Object Localization and Instance Segmentation

3 code implementations17 Dec 2021 Wuyang Chen, Xianzhi Du, Fan Yang, Lucas Beyer, Xiaohua Zhai, Tsung-Yi Lin, Huizhong Chen, Jing Li, Xiaodan Song, Zhangyang Wang, Denny Zhou

In this paper, we comprehensively study three architecture design choices on ViT -- spatial reduction, doubled channels, and multiscale features -- and demonstrate that a vanilla ViT architecture can fulfill this goal without handcrafting multiscale features, maintaining the original ViT design philosophy.

Image Classification Instance Segmentation +6

Federated Dynamic Sparse Training: Computing Less, Communicating Less, Yet Learning Better

1 code implementation18 Dec 2021 Sameer Bibikar, Haris Vikalo, Zhangyang Wang, Xiaohan Chen

Federated learning (FL) enables distribution of machine learning workloads from the cloud to resource-limited edge devices.

Federated Learning

CADTransformer: Panoptic Symbol Spotting Transformer for CAD Drawings

1 code implementation CVPR 2022 Zhiwen Fan, Tianlong Chen, Peihao Wang, Zhangyang Wang

CADTransformer tokenizes directly from the set of graphical primitives in CAD drawings, and correspondingly optimizes line-grained semantic and instance symbol spotting altogether by a pair of prediction heads.

Data Augmentation

Fast and High-Quality Image Denoising via Malleable Convolutions

no code implementations2 Jan 2022 Yifan Jiang, Bartlomiej Wronski, Ben Mildenhall, Jonathan T. Barron, Zhangyang Wang, Tianfan Xue

These spatially-varying kernels are produced by an efficient predictor network running on a downsampled input, making them much more efficient to compute than per-pixel kernels produced by a full-resolution image, and also enlarging the network's receptive field compared with static kernels.

Image Denoising Image Restoration +1

Bringing Your Own View: Graph Contrastive Learning without Prefabricated Data Augmentations

1 code implementation4 Jan 2022 Yuning You, Tianlong Chen, Zhangyang Wang, Yang shen

Accordingly, we have extended the prefabricated discrete prior in the augmentation set, to a learnable continuous prior in the parameter space of graph generators, assuming that graph priors per se, similar to the concept of image manifolds, can be learned by data generation.

Contrastive Learning Graph Learning

VAQF: Fully Automatic Software-Hardware Co-Design Framework for Low-Bit Vision Transformer

no code implementations17 Jan 2022 Mengshu Sun, Haoyu Ma, Guoliang Kang, Yifan Jiang, Tianlong Chen, Xiaolong Ma, Zhangyang Wang, Yanzhi Wang

To the best of our knowledge, this is the first time quantization has been incorporated into ViT acceleration on FPGAs with the help of a fully automatic framework to guide the quantization strategy on the software side and the accelerator implementations on the hardware side given the target frame rate.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.