1 code implementation • 23 Apr 2024 • Yun Yue, Fangzhou Lin, Guanyi Mou, Ziming Zhang
In recent years, there has been a growing trend of incorporating hyperbolic geometry methods into computer vision.
1 code implementation • NeurIPS 2023 • Yun Yue, Zhiling Ye, Jiadi Jiang, Yongchao Liu, Ke Zhang
Additionally, we introduce an auto-switching function that enables the preconditioning matrix to switch dynamically between Stochastic Gradient Descent (SGD) and the adaptive optimizer.
1 code implementation • 25 May 2023 • Yun Yue, Jiadi Jiang, Zhiling Ye, Ning Gao, Yongchao Liu, Ke Zhang
Deep Neural Networks (DNNs) generalization is known to be closely related to the flatness of minima, leading to the development of Sharpness-Aware Minimization (SAM) for seeking flatter minima and better generalization.
no code implementations • 2 Feb 2023 • Yun Yue, Fangzhou Lin, Kazunori D Yamada, Ziming Zhang
Learning good image representations that are beneficial to downstream tasks is a challenging task in computer vision.
1 code implementation • ICCV 2023 • Fangzhou Lin, Yun Yue, Songlin Hou, Xuechu Yu, Yajun Xu, Kazunori D Yamada, Ziming Zhang
Chamfer distance (CD) is a standard metric to measure the shape dissimilarity between point clouds in point cloud completion, as well as a loss function for (deep) learning.
no code implementations • NeurIPS 2021 • Ziming Zhang, Yun Yue, Guojun Wu, Yanhua Li, Haichong Zhang
In this paper we consider the training stability of recurrent neural networks (RNNs) and propose a family of RNNs, namely SBO-RNN, that can be formulated using stochastic bilevel optimization (SBO).
no code implementations • 29 Sep 2021 • Guojun Wu, Yun Yue, Yanhua Li, Ziming Zhang
Lightweight neural networks refer to deep networks with small numbers of parameters, which are allowed to be implemented in resource-limited hardware such as embedded systems.
1 code implementation • 30 Jul 2021 • Yun Yue, Yongchao Liu, Suo Tong, Minghao Li, Zhen Zhang, Chunyang Wen, Huanjun Bao, Lihong Gu, Jinjie Gu, Yixiang Mu
We develop a novel framework that adds the regularizers of the sparse group lasso to a family of adaptive optimizers in deep learning, such as Momentum, Adagrad, Adam, AMSGrad, AdaHessian, and create a new class of optimizers, which are named Group Momentum, Group Adagrad, Group Adam, Group AMSGrad and Group AdaHessian, etc., accordingly.
no code implementations • 1 Jan 2021 • Yun Yue, Suo Tong, Zhen Zhang, Yongchao Liu, Chunyang Wen, Huanjun Bao, Jinjie Gu, Yixiang Mu
We develop a novel framework that adds the regularizers to a family of adaptive optimizers in deep learning, such as MOMENTUM, ADAGRAD, ADAM, AMSGRAD, ADAHESSIAN, and create a new class of optimizers, which are named GROUP MOMENTUM, GROUP ADAGRAD, GROUP ADAM, GROUP AMSGRAD and GROUP ADAHESSIAN, etc., accordingly.
1 code implementation • 12 Oct 2020 • Yun Yue, Ming Li, Venkatesh Saligrama, Ziming Zhang
We propose to utilize the Frank-Wolfe (FW) algorithm in this context.