1 code implementation • 22 Feb 2024 • Xuxi Chen, Zhendong Wang, Daouda Sow, Junjie Yang, Tianlong Chen, Yingbin Liang, Mingyuan Zhou, Zhangyang Wang
Our study starts from an empirical strategy for the light continual training of LLMs using their original pre-training data sets, with a specific focus on selective retention of samples that incur moderately high losses.
no code implementations • 1 Aug 2023 • Daouda Sow, Sen Lin, Zhangyang Wang, Yingbin Liang
Experiments on standard classification datasets demonstrate that our proposed approach outperforms related state-of-the-art baseline methods in terms of average robust performance, and at the same time improves the robustness against attacks on the weakest data points.
no code implementations • 2 Feb 2023 • Daouda Sow, Sen Lin, Yingbin Liang, Junshan Zhang
More specifically, we first propose two simple but effective detection mechanisms of task switches and distribution shift based on empirical observations, which serve as a key building block for more elegant online model updates in our algorithm: the task switch detection mechanism allows reusing of the best model available for the current task at hand, and the distribution shift detection mechanism differentiates the meta model update in order to preserve the knowledge for in-distribution tasks and quickly learn the new knowledge for out-of-distribution tasks.
no code implementations • 1 Mar 2022 • Daouda Sow, Kaiyi Ji, Ziwei Guan, Yingbin Liang
Existing algorithms designed for such a problem were applicable to restricted situations and do not come with a full guarantee of convergence.
1 code implementation • 13 Oct 2021 • Daouda Sow, Kaiyi Ji, Yingbin Liang
Bilevel optimization has arisen as a powerful tool in modern machine learning.
no code implementations • 29 Sep 2021 • Daouda Sow, Kaiyi Ji, Yingbin Liang
Bilevel optimization (BO) has arisen as a powerful tool for solving many modern machine learning problems.
no code implementations • 1 Nov 2018 • Daouda Sow, Zengchang Qin, Mouhamed Niasse, Tao Wan
The recent advances of deep learning in both computer vision (CV) and natural language processing (NLP) provide us a new way of understanding semantics, by which we can deal with more challenging tasks such as automatic description generation from natural images.