Search Results for author: Yingzhen Yang

Found 38 papers, 5 papers with code

Gradient Descent Finds Over-Parameterized Neural Networks with Sharp Generalization for Nonparametric Regression: A Distribution-Free Analysis

no code implementations5 Nov 2024 Yingzhen Yang, Ping Li

We show that, if the neural network is trained by GD with early stopping, then the trained network renders a sharp rate of the nonparametric regression risk of $\cO(\eps_n^2)$, which is the same rate as that for the classical kernel regression trained by GD with early stopping, where $\eps_n$ is the critical population rate of the Neural Tangent Kernel (NTK) associated with the network and $n$ is the size of the training data.

regression

Locally Regularized Sparse Graph by Fast Proximal Gradient Descent

no code implementations25 Sep 2024 Dongfang Sun, Yingzhen Yang

In order to obtain a sparse graph aligned with the local geometric structure of data, we propose a novel Support Regularized Sparse Graph, abbreviated as SRSG, for data clustering.

Clustering

Deep Geometric Moments Promote Shape Consistency in Text-to-3D Generation

no code implementations12 Aug 2024 Utkarsh Nath, Rajeev Goel, Eun Som Jeon, Changhoon Kim, Kyle Min, Yezhou Yang, Yingzhen Yang, Pavan Turaga

To address the data scarcity associated with 3D assets, 2D-lifting techniques such as Score Distillation Sampling (SDS) have become a widely adopted practice in text-to-3D generation pipelines.

3D Generation Text to 3D

IDNet: A Novel Dataset for Identity Document Analysis and Fraud Detection

no code implementations3 Aug 2024 Hong Guan, Yancheng Wang, Lulu Xie, Soham Nag, Rajeev Goel, Niranjan Erappa Narayana Swamy, Yingzhen Yang, Chaowei Xiao, Jonathan Prisby, Ross Maciejewski, Jia Zou

Effective fraud detection and analysis of government-issued identity documents, such as passports, driver's licenses, and identity cards, are essential in thwarting identity theft and bolstering security on online platforms.

Fraud Detection Privacy Preserving

Efficient Visual Transformer by Learnable Token Merging

1 code implementation21 Jul 2024 Yancheng Wang, Yingzhen Yang

In this paper, we propose a novel and compact transformer block, Transformer with Learnable Token Merging (LTM), or LTM-Transformer.

Preconditioned Gradient Descent Finds Over-Parameterized Neural Networks with Sharp Generalization for Nonparametric Regression

no code implementations16 Jul 2024 Yingzhen Yang

We show that, if the neural network is trained with a novel Preconditioned Gradient Descent (PGD) with early stopping and the target function has spectral bias widely studied in the deep learning literature, the trained network renders a particularly sharp generalization bound with a minimax optimal rate of $\cO({1}/{n^{4\alpha/(4\alpha+1)}})$, which is sharper the current standard rate of $\cO({1}/{n^{2\alpha/(2\alpha+1)}})$ with $2\alpha = d/(d-1)$ when the data is distributed uniformly on the unit sphere in $\RR^d$ and $n$ is the size of the training data.

Learning Low-Rank Feature for Thorax Disease Classification

no code implementations14 Feb 2024 Rajeev Goel, Utkarsh Nath, Yancheng Wang, Alvin C. Silva, Teresa Wu, Yingzhen Yang

To address this challenge, we propose a novel Low-Rank Feature Learning (LRFL) method in this paper, which is universally applicable to the training of all neural networks.

Classification Self-Supervised Learning

Low-Rank Graph Contrastive Learning for Node Classification

no code implementations14 Feb 2024 Yancheng Wang, Yingzhen Yang

To the best of our knowledge, our theoretical result is among the first to theoretically demonstrate the advantage of low-rank learning in graph contrastive learning supported by strong empirical performance.

Classification Contrastive Learning +2

Sketching for Convex and Nonconvex Regularized Least Squares with Sharp Guarantees

no code implementations3 Nov 2023 Yingzhen Yang, Ping Li

We present general theoretical result for the approximation error between the optimization results of the original problem and the sketched problem for regularized least square problems which can be convex or nonconvex.

Sparse Learning

Sharp Generalization of Transductive Learning: A Transductive Local Rademacher Complexity Approach

no code implementations28 Sep 2023 Yingzhen Yang

Using a peeling strategy and a new surrogate variance operator, we derive the a novel excess risk bound in the transductive setting which is consistent with the classical LRC-based excess risk bound in the inductive setting.

Generalization Bounds Learning Theory +1

RecMind: Large Language Model Powered Agent For Recommendation

no code implementations28 Aug 2023 Yancheng Wang, Ziyan Jiang, Zheng Chen, Fan Yang, Yingxue Zhou, Eunah Cho, Xing Fan, Xiaojiang Huang, Yanbin Lu, Yingzhen Yang

While the recommendation system (RS) has advanced significantly through deep learning, current RS approaches usually train and fine-tune models on task-specific datasets, limiting their generalizability to new recommendation tasks and their ability to leverage external knowledge due to model scale and data size constraints.

Explanation Generation Language Modelling +2

Projective Proximal Gradient Descent for A Class of Nonconvex Nonsmooth Optimization Problems: Fast Convergence Without Kurdyka-Lojasiewicz (KL) Property

no code implementations20 Apr 2023 Yingzhen Yang, Ping Li

It is proved that PPGD achieves a fast convergence rate of $\cO(1/k^2)$ when the iteration number $k \ge k_0$ for a finite $k_0$ on a class of nonconvex and nonsmooth problems under mild assumptions, which is locally Nesterov's optimal convergence rate of first-order methods on smooth and convex objective function with Lipschitz continuous gradient.

RNAS-CL: Robust Neural Architecture Search by Cross-Layer Knowledge Distillation

no code implementations19 Jan 2023 Utkarsh Nath, Yancheng Wang, Yingzhen Yang

In this paper, we propose Robust Neural Architecture Search by Cross-Layer Knowledge Distillation (RNAS-CL), a novel NAS algorithm that improves the robustness of NAS by learning from a robust teacher through cross-layer knowledge distillation.

Knowledge Distillation Neural Architecture Search

Noisy $\ell^{0}$-Sparse Subspace Clustering on Dimensionality Reduced Data

no code implementations22 Jun 2022 Yingzhen Yang, Ping Li

Our results provide theoretical guarantee on the correctness of noisy $\ell^{0}$-SSC in terms of SDP on noisy data for the first time, which reveals the advantage of noisy $\ell^{0}$-SSC in terms of much less restrictive condition on subspace affinity.

Clustering

Bayesian Robust Graph Contrastive Learning

1 code implementation27 May 2022 Yancheng Wang, Yingzhen Yang

In this work, we propose a novel and robust method, Bayesian Robust Graph Contrastive Learning (BRGCL), which trains a GNN encoder to learn robust node representations.

Contrastive Learning Node Classification

Eliciting Structural and Semantic Global Knowledge in Unsupervised Graph Contrastive Learning

1 code implementation17 Feb 2022 Kaize Ding, Yancheng Wang, Yingzhen Yang, Huan Liu

In general, the contrastive learning process in GCL is performed on top of the representations learned by a graph neural network (GNN) backbone, which transforms and propagates the node contextual information based on its local neighborhoods.

Contrastive Learning Graph Neural Network +1

Discriminative Similarity for Data Clustering

no code implementations ICLR 2022 Yingzhen Yang, Ping Li

Similarity-based clustering methods separate data into clusters according to the pairwise similarity between the data, and the pairwise similarity is crucial for their performance.

Clustering

Adjoined Networks: A Training Paradigm with Applications to Network Compression

1 code implementation10 Jun 2020 Utkarsh Nath, Shrinu Kushagra, Yingzhen Yang

In this paper, we introduce Adjoined Networks, or AN, a learning paradigm that trains both the original base network and the smaller compressed network together.

Knowledge Distillation Neural Architecture Search +1

FSNet: Compression of Deep Convolutional Neural Networks by Filter Summary

no code implementations ICLR 2020 Yingzhen Yang, Jiahui Yu, Nebojsa Jojic, Jun Huan, Thomas S. Huang

FSNet has the same architecture as that of the baseline CNN to be compressed, and each convolution layer of FSNet has the same number of filters from FS as that of the basline CNN in the forward process.

General Classification Image Classification +4

An Empirical Study on Regularization of Deep Neural Networks by Local Rademacher Complexity

no code implementations3 Feb 2019 Yingzhen Yang, Jiahui Yu, Xingjian Li, Jun Huan, Thomas S. Huang

In this paper, we investigate the role of Rademacher complexity in improving generalization of DNNs and propose a novel regularizer rooted in Local Rademacher Complexity (LRC).

Neural Architecture Search

Learning $3$D-FilterMap for Deep Convolutional Neural Networks

no code implementations5 Jan 2018 Yingzhen Yang, Jianchao Yang, Ning Xu, Wei Han

Due to the weight sharing scheme, the parameter size of the $3$D-FilterMap is much smaller than that of the filters to be learned in the conventional convolution layer when $3$D-FilterMap generates the same number of filters.

On the Suboptimality of Proximal Gradient Descent for $\ell^{0}$ Sparse Approximation

no code implementations5 Sep 2017 Yingzhen Yang, Jiashi Feng, Nebojsa Jojic, Jianchao Yang, Thomas S. Huang

We study the proximal gradient descent (PGD) method for $\ell^{0}$ sparse approximation problem as well as its accelerated optimization with randomized algorithms in this paper.

Compressive Sensing Dimensionality Reduction

Discriminative Similarity for Clustering and Semi-Supervised Learning

no code implementations5 Sep 2017 Yingzhen Yang, Feng Liang, Nebojsa Jojic, Shuicheng Yan, Jiashi Feng, Thomas S. Huang

By generalization analysis via Rademacher complexity, the generalization error bound for the kernel classifier learned from hypothetical labeling is expressed as the sum of pairwise similarity between the data from different classes, parameterized by the weights of the kernel classifier.

Clustering

D3: Deep Dual-Domain Based Fast Restoration of JPEG-Compressed Images

no code implementations CVPR 2016 Zhangyang Wang, Ding Liu, Shiyu Chang, Qing Ling, Yingzhen Yang, Thomas S. Huang

In this paper, we design a Deep Dual-Domain (D3) based fast restoration model to remove artifacts of JPEG compressed images.

Learning A Deep $\ell_\infty$ Encoder for Hashing

no code implementations6 Apr 2016 Zhangyang Wang, Yingzhen Yang, Shiyu Chang, Qing Ling, Thomas S. Huang

We investigate the $\ell_\infty$-constrained representation which demonstrates robustness to quantization errors, utilizing the tool of deep learning.

Quantization

$\mathbf{D^3}$: Deep Dual-Domain Based Fast Restoration of JPEG-Compressed Images

no code implementations16 Jan 2016 Zhangyang Wang, Ding Liu, Shiyu Chang, Qing Ling, Yingzhen Yang, Thomas S. Huang

In this paper, we design a Deep Dual-Domain ($\mathbf{D^3}$) based fast restoration model to remove artifacts of JPEG compressed images.

Learning with $\ell^{0}$-Graph: $\ell^{0}$-Induced Sparse Subspace Clustering

no code implementations28 Oct 2015 Yingzhen Yang, Jiashi Feng, Jianchao Yang, Thomas S. Huang

Sparse subspace clustering methods, such as Sparse Subspace Clustering (SSC) \cite{ElhamifarV13} and $\ell^{1}$-graph \cite{YanW09, ChengYYFH10}, are effective in partitioning the data that lie in a union of subspaces.

Clustering

Designing A Composite Dictionary Adaptively From Joint Examples

no code implementations12 Mar 2015 Zhangyang Wang, Yingzhen Yang, Jianchao Yang, Thomas S. Huang

We study the complementary behaviors of external and internal examples in image restoration, and are motivated to formulate a composite dictionary design framework.

Image Denoising Image Restoration +1

Learning Super-Resolution Jointly from External and Internal Examples

no code implementations3 Mar 2015 Zhangyang Wang, Yingzhen Yang, Zhaowen Wang, Shiyu Chang, Jianchao Yang, Thomas S. Huang

Single image super-resolution (SR) aims to estimate a high-resolution (HR) image from a lowresolution (LR) input.

Image Super-Resolution

On a Theory of Nonparametric Pairwise Similarity for Clustering: Connecting Clustering to Classification

no code implementations NeurIPS 2014 Yingzhen Yang, Feng Liang, Shuicheng Yan, Zhangyang Wang, Thomas S. Huang

Modeling the underlying data distribution by nonparametric kernel density estimation, the generalization error bounds for both unsupervised nonparametric classifiers are the sum of nonparametric pairwise similarity terms between the data points for the purpose of clustering.

Clustering Density Estimation +2

Nonparametric Unsupervised Classification

no code implementations2 Oct 2012 Yingzhen Yang, Thomas S. Huang

Unsupervised classification methods learn a discriminative classifier from unlabeled data, which has been proven to be an effective way of simultaneously clustering the data and training a classifier from the data.

Classification Clustering +1

Cannot find the paper you are looking for? You can Submit a new open access paper.