no code implementations • 9 Dec 2024 • Xinyu Yang, Jixuan Leng, Geyang Guo, Jiawei Zhao, Ryumei Nakada, Linjun Zhang, Huaxiu Yao, Beidi Chen
Utilizing this key insight, we propose a family of Structured Sparse Fine-Tuning (S$^{2}$FT) methods for LLMs, which concurrently achieve state-of-the-art fine-tuning performance, training efficiency, and inference scalability.
no code implementations • 2 Oct 2024 • Yibo Zhong, Haoxiang Jiang, Lincan Li, Ryumei Nakada, Tianci Liu, Linjun Zhang, Huaxiu Yao, Haoyu Wang
The nonlinear approximation directly models the cumulative updates, effectively capturing complex and non-linear structures in the weight updates.
1 code implementation • 5 Jun 2024 • Ryumei Nakada, Yichen Xu, Lexin Li, Linjun Zhang
Imbalanced data and spurious correlations are common challenges in machine learning and data science.
no code implementations • 25 Apr 2024 • Zhe Zhang, Ryumei Nakada, Linjun Zhang
Differentially private federated learning is crucial for maintaining privacy in distributed environments.
no code implementations • 22 Mar 2024 • Tianxi Cai, Feiqing Huang, Ryumei Nakada, Linjun Zhang, Doudou Zhou
To accommodate the statistical analysis of multimodal EHR data, in this paper, we propose a novel multimodal feature embedding generative model and design a multimodal contrastive loss to obtain the multimodal EHR feature representation.
1 code implementation • 13 Jun 2023 • Alyssa Huang, Peihan Liu, Ryumei Nakada, Linjun Zhang, Wanrong Zhang
The surge in multimodal AI's success has sparked concerns over data privacy in vision-and-language tasks.
1 code implementation • 13 Feb 2023 • Ryumei Nakada, Halil Ibrahim Gulluk, Zhun Deng, Wenlong Ji, James Zou, Linjun Zhang
We show that the algorithm can detect the ground-truth pairs and improve performance by fully exploiting unpaired datasets.
no code implementations • 6 Oct 2021 • Wenlong Ji, Zhun Deng, Ryumei Nakada, James Zou, Linjun Zhang
Contrastive learning has achieved state-of-the-art performance in various self-supervised learning tasks and even outperforms its supervised counterpart.
no code implementations • 28 Feb 2021 • Ryumei Nakada, Masaaki Imaizumi
We investigate the asymptotic risk of a general class of overparameterized likelihood models, including deep models.
no code implementations • 4 Jul 2019 • Ryumei Nakada, Masaaki Imaizumi
In this study, we prove that an intrinsic low dimensionality of covariates is the main factor that determines the performance of deep neural networks (DNNs).