no code implementations • 16 Jul 2024 • Yucheng Lu, Dovile Juodelyte, Jonathan D. Victor, Veronika Cheplygina
In this paper, we use spectral analysis to investigate transfer learning and study model sensitivity to frequency shortcuts in medical imaging.
1 code implementation • 7 Mar 2024 • Dovile Juodelyte, Yucheng Lu, Amelia Jiménez-Sánchez, Sabrina Bottazzi, Enzo Ferrante, Veronika Cheplygina
Transfer learning has become an essential part of medical imaging classification algorithms, often leveraging ImageNet weights.
1 code implementation • 21 Apr 2023 • Yucheng Lu, Zhixin Xu, Moon Hyung Choi, Jimin Kim, Seung-Won Jung
Computed tomography (CT) has been used worldwide as a non-invasive test to assist in diagnosis.
1 code implementation • NeurIPS 2023 • A. Feder Cooper, Wentao Guo, Khiem Pham, Tiancheng Yuan, Charlie F. Ruan, Yucheng Lu, Christopher De Sa
Recent research on online Gradient Balancing (GraB) has revealed that there exist permutation-based example orderings for SGD that are guaranteed to outperform random reshuffling (RR).
no code implementations • 2 Feb 2023 • Yucheng Lu, Shivani Agrawal, Suvinay Subramanian, Oleg Rybakov, Christopher De Sa, Amir Yazdanbakhsh
Recent innovations on hardware (e. g. Nvidia A100) have motivated learning N:M structured sparsity masks from scratch for fast model inference.
1 code implementation • 24 Aug 2022 • Jun-Sang Yoo, Dong-Wook Kim, Yucheng Lu, Seung-Won Jung
To advance ZSSR, we obtain reference image patches with rich textures and high-frequency details which are also extracted only from the input image using cross-scale matching.
3 code implementations • 22 May 2022 • Yucheng Lu, Wentao Guo, Christopher De Sa
To reduce the memory overhead, we leverage discrepancy minimization theory to propose an online Gradient Balancing algorithm (GraB) that enjoys the same rate as herding, while reducing the memory usage from $O(nd)$ to just $O(d)$ and computation from $O(n^2)$ to $O(n)$, where $d$ denotes the model dimension.
1 code implementation • 12 Feb 2022 • Yucheng Lu, Conglong Li, Minjia Zhang, Christopher De Sa, Yuxiong He
1-bit gradient compression and local steps are two representative techniques that enable drastic communication reduction in distributed SGD.
no code implementations • ICLR 2022 • Yucheng Lu, Si Yi Meng, Christopher De Sa
In this paper, we develop a broad condition on the sequence of examples used by SGD that is sufficient to prove tight convergence rates in both strongly convex and non-convex settings.
1 code implementation • 28 Jun 2021 • Yucheng Lu, Seung-Won Jung
Low-light imaging on mobile devices is typically challenging due to insufficient incident light coming through the relatively small aperture, resulting in a low signal-to-noise ratio.
no code implementations • 2 Mar 2021 • Yucheng Lu, Youngsuk Park, Lifan Chen, Yuyang Wang, Christopher De Sa, Dean Foster
In large-scale time series forecasting, one often encounters the situation where the temporal patterns of time series, while drifting over time, differ from one another in the same dataset.
1 code implementation • NeurIPS 2021 • A. Feder Cooper, Yucheng Lu, Jessica Zosa Forde, Christopher De Sa
Recent empirical work shows that inconsistent results based on choice of hyperparameter optimization (HPO) configuration are a widespread problem in ML research.
no code implementations • 15 Jun 2020 • Yucheng Lu, Christopher De Sa
Decentralization is a promising method of scaling up parallel machine learning systems.
no code implementations • 14 May 2020 • Yucheng Lu, Jack Nash, Christopher De Sa
Parallelism is a ubiquitous method for accelerating machine learning algorithms.
no code implementations • ICML 2020 • Yucheng Lu, Christopher De Sa
Running Stochastic Gradient Descent (SGD) in a decentralized fashion has shown promising results.