no code implementations • 1 Aug 2024 • Kuo Gai, Sicong Wang, Shihua Zhang
OTAD opens a novel avenue for developing reliable and secure deep learning systems through the regularity of optimal transport maps.
no code implementations • 2 Jul 2024 • Kuo Gai, Shihua Zhang
In practice, deeper networks tend to be more powerful than shallow ones, but this has not been understood theoretically.
no code implementations • 2 May 2024 • Sicong Wang, Kuo Gai, Shihua Zhang
Overall, this study extends NC to PFC to model the collapse phenomenon of intermediate layers and its dependence on the input data, shedding light on the theoretical understanding of ResNet in classification problems.
1 code implementation • CVPR 2024 • Shihua Zhang, Zizhuo Li, Yuan Gao, Jiayi Ma
Specifically we first decompose the rough motion field that is contaminated by false matches into several different sub-fields which are highly smooth and contain the main energy of the original field.
1 code implementation • 10 Feb 2023 • Rui Zhang, Qi Meng, Rongchan Zhu, Yue Wang, Wenlei Shi, Shihua Zhang, Zhi-Ming Ma, Tie-Yan Liu
To address these limitations, we propose the Monte Carlo Neural PDE Solver (MCNP Solver) for training unsupervised neural solvers via the PDEs' probabilistic representation, which regards macroscopic phenomena as ensembles of random particles.
no code implementations • 15 Dec 2021 • Rui Zhang, Shihua Zhang
However, the classic implicit Hessian-vector product (IHVP) method for calculating IF is fragile, and theoretical analysis of IF in the context of neural networks is still lacking.
1 code implementation • 1 Sep 2021 • Chihao Zhang, Yiling Elaine Chen, Shihua Zhang, Jingyi Jessica Li
While practitioners commonly combine ambiguous outcome labels for all data points (instances) in an ad hoc way to improve the accuracy of multi-class classification, there lacks a principled approach to guide the label combination for all data points by any optimality criterion.
no code implementations • 28 Feb 2021 • Penglong Zhai, Shihua Zhang
The information bottleneck (IB) principle has been adopted to explain deep learning in terms of information compression and prediction, which are balanced by a trade-off hyperparameter.
no code implementations • 18 Feb 2021 • Kuo Gai, Shihua Zhang
In a word, we conclude a mathematical principle of deep learning is to learn the geodesic curve in the Wasserstein space; and deep learning is a great engineering realization of continuous transformation in high-dimensional space.
no code implementations • 16 Oct 2020 • Penglong Zhai, Shihua Zhang
Low-rank approximation models of data matrices have become important machine learning and data mining tools in many fields including computer vision, text mining, bioinformatics and many others.
no code implementations • 20 May 2020 • Kuo Gai, Shihua Zhang
Non-adversarial generative models such as variational auto-encoder (VAE), Wasserstein auto-encoders with maximum mean discrepancy (WAE-MMD), sliced-Wasserstein auto-encoder (SWAE) are relatively easy to train and have less mode collapse compared to Wasserstein auto-encoder with generative adversarial network (WAE-GAN).
2 code implementations • 10 Feb 2020 • Chihao Zhang, Yang Yang, Wei zhang, Shihua Zhang
Such a method should scale up well, model the heterogeneous noise, and address the communication issue in a distributed system.
no code implementations • 5 Dec 2019 • Zhiyang Zhang, Shihua Zhang
Inspired by these considerations, we propose two novel multi-layer models--residual convolutional sparse coding model (Res-CSC) and mixed-scale dense convolutional sparse coding model (MSD-CSC), which have close relationship with the residual neural network (ResNet) and mixed-scale (dilated) dense neural network (MSDNet), respectively.
no code implementations • 25 Nov 2019 • Chihao Zhang, Kuo Gai, Shihua Zhang
However, most of the existing methods only assume that the noise is correlated in the feature space while there may exist two-way structured noise.
no code implementations • 28 Jul 2018 • Wenwen Min, Juan Liu, Shihua Zhang
We employ an alternating direction method of multipliers (ADMM) to solve the proximal operator.
no code implementations • 9 Dec 2017 • Chihao Zhang, Shihua Zhang
A few of matrix decomposition methods have been extended for such multi-view data integration and pattern discovery.
no code implementations • 13 Oct 2017 • Wenwen Min, Juan Liu, Shihua Zhang
Given two data matrices $X$ and $Y$, sparse canonical correlation analysis (SCCA) is to seek two sparse canonical vectors $u$ and $v$ to maximize the correlation between $Xu$ and $Yv$.
no code implementations • 28 Jul 2017 • Zhenxing Guo, Shihua Zhang
Deep learning, however, with its carefully designed hierarchical structure, is able to combine hidden features to form more representative features for pattern recognition.
1 code implementation • 25 Jul 2017 • Lihua Zhang, Shihua Zhang
In this paper, we introduce a sparse multiple relationship data regularized joint matrix factorization (JMF) framework and two adapted prediction models for pattern recognition and data integration.
no code implementations • 21 Sep 2016 • Wenwen Min, Juan Liu, Shihua Zhang
To address it, we introduce a novel network-regularized sparse LR model with a new penalty $\lambda \|\bm{w}\|_1 + \eta|\bm{w}|^T\bm{M}|\bm{w}|$ to consider the difference between the absolute values of the coefficients.
no code implementations • 19 Mar 2016 • Wenwen Min, Juan Liu, Shihua Zhang
Motivated by the development of sparse coding and graph-regularized norm, we propose a novel sparse graph-regularized SVD as a powerful biclustering tool for analyzing high-dimensional data.