no code implementations • 19 Jun 2024 • Xiangfeng Wang, Zaiyi Chen, Zheyong Xie, Tong Xu, Yongyi He, Enhong Chen
With the rising popularity of Transformer-based large language models (LLMs), reducing their high inference costs has become a significant research focus.
no code implementations • 16 Jan 2024 • Qixin Zhang, Zongqi Wan, Zengde Deng, Zaiyi Chen, Xiaoming Sun, Jialin Zhang, Yu Yang
The fundamental idea of our boosting technique is to exploit non-oblivious search to derive a novel auxiliary function $F$, whose stationary points are excellent approximations to the global maximum of the original DR-submodular objective $f$.
no code implementations • 6 Mar 2023 • Yuwei Chen, Zengde Deng, Yinzhi Zhou, Zaiyi Chen, Yujie Chen, Haoyuan Hu
This paper studies the online stochastic resource allocation problem (RAP) with chance constraints.
no code implementations • 18 Aug 2022 • Qixin Zhang, Zengde Deng, Xiangru Jian, Zaiyi Chen, Haoyuan Hu, Yu Yang
Maximizing a monotone submodular function is a fundamental task in machine learning, economics, and statistics.
no code implementations • 16 Aug 2022 • Qixin Zhang, Zengde Deng, Zaiyi Chen, Kuangqi Zhou, Haoyuan Hu, Yu Yang
In this paper, we revisit the online non-monotone continuous DR-submodular maximization problem over a down-closed convex set, which finds wide real-world applications in the domain of machine learning, economics, and operations research.
no code implementations • 3 Jan 2022 • Qixin Zhang, Zengde Deng, Zaiyi Chen, Haoyuan Hu, Yu Yang
In the online setting, for the first time we consider the adversarial delays for stochastic gradient feedback, under which we propose a boosting online gradient algorithm with the same non-oblivious function $F$.
no code implementations • 28 Dec 2021 • Qixin Zhang, Wenbing Ye, Zaiyi Chen, Haoyuan Hu, Enhong Chen, Yang Yu
As a result, only limited violations of constraints or pessimistic competitive bounds could be guaranteed.
no code implementations • 1 Jan 2021 • Hui Zhong, Zaiyi Chen, Chuan Qin, Zai Huang, Vincent W. Zheng, Tong Xu, Enhong Chen
Though many algorithms, such as AMSGRAD and ADAMNC, have been proposed to fix the non-convergence issues, achieving a data-dependent regret bound similar to or better than ADAGRAD is still a challenge to these methods.
4 code implementations • ICCV 2019 • Yisen Wang, Xingjun Ma, Zaiyi Chen, Yuan Luo, Jin-Feng Yi, James Bailey
In this paper, we show that DNN learning with Cross Entropy (CE) exhibits overfitting to noisy labels on some classes ("easy" classes), but more surprisingly, it also suffers from significant under learning on some other classes ("hard" classes).
Ranked #43 on
Image Classification
on Clothing1M
no code implementations • 10 Jun 2019 • Dong-Dong Chen, Yisen Wang, Jin-Feng Yi, Zaiyi Chen, Zhi-Hua Zhou
Unsupervised domain adaptation aims to transfer the classifier learned from the source domain to the target domain in an unsupervised manner.
no code implementations • 14 Sep 2018 • Zaiyi Chen
Rank minimization (RM) is a wildly investigated task of finding solutions by exploiting low-rank structure of parameter matrices.
no code implementations • ICLR 2019 • Zaiyi Chen, Zhuoning Yuan, Jin-Feng Yi, Bo-Wen Zhou, Enhong Chen, Tianbao Yang
For example, there is still a lack of theories of convergence for SGD and its variants that use stagewise step size and return an averaged solution in practice.
no code implementations • ICML 2018 • Mingrui Liu, Xiaoxuan Zhang, Zaiyi Chen, Xiaoyu Wang, Tianbao Yang
In this paper, we consider statistical learning with AUC (area under ROC curve) maximization in the classical stochastic setting where one random data drawn from an unknown distribution is revealed at each iteration for updating the model.
no code implementations • ICML 2018 • Zaiyi Chen, Yi Xu, Enhong Chen, Tianbao Yang
Although the convergence rates of existing variants of ADAGRAD have a better dependence on the number of iterations under the strong convexity condition, their iteration complexities have a explicitly linear dependence on the dimensionality of the problem.