Search Results for author: Yuege Xie

Found 5 papers, 1 papers with code

Adaptively Weighted Data Augmentation Consistency Regularization for Robust Optimization under Concept Shift

no code implementations4 Oct 2022 Yijun Dong, Yuege Xie, Rachel Ward

At the saddle point of the underlying objective, the weights assign label-dense samples to the supervised loss and label-sparse samples to the unsupervised consistency regularization.

Data Augmentation Image Segmentation +3

SHRIMP: Sparser Random Feature Models via Iterative Magnitude Pruning

1 code implementation7 Dec 2021 Yuege Xie, Bobby Shi, Hayden Schaeffer, Rachel Ward

Inspired by the success of the iterative magnitude pruning technique in finding lottery tickets of neural networks, we propose a new method -- Sparser Random Feature Models via IMP (ShRIMP) -- to efficiently fit high-dimensional data with inherent low-dimensional structure in the form of sparse variable dependencies.

Additive models Computational Efficiency +1

AdaLoss: A computationally-efficient and provably convergent adaptive gradient method

no code implementations17 Sep 2021 Xiaoxia Wu, Yuege Xie, Simon Du, Rachel Ward

We propose a computationally-friendly adaptive learning rate schedule, "AdaLoss", which directly uses the information of the loss function to adjust the stepsize in gradient descent methods.

regression

Overparameterization and generalization error: weighted trigonometric interpolation

no code implementations15 Jun 2020 Yuege Xie, Hung-Hsu Chou, Holger Rauhut, Rachel Ward

Motivated by surprisingly good generalization properties of learned deep neural networks in overparameterized scenarios and by the related double descent phenomenon, this paper analyzes the relation between smoothness and low generalization error in an overparameterized linear learning problem.

Linear Convergence of Adaptive Stochastic Gradient Descent

no code implementations28 Aug 2019 Yuege Xie, Xiaoxia Wu, Rachel Ward

We prove that the norm version of the adaptive stochastic gradient method (AdaGrad-Norm) achieves a linear convergence rate for a subset of either strongly convex functions or non-convex functions that satisfy the Polyak Lojasiewicz (PL) inequality.

Cannot find the paper you are looking for? You can Submit a new open access paper.