no code implementations • 1 May 2023 • Ziyang Zhang, Huan Li, Yang Zhao, Changyao Lin, Jie Liu
As deep neural networks (DNNs) are being applied to a wide range of edge intelligent applications, it is critical for edge inference platforms to have both high-throughput and low-latency at the same time.
1 code implementation • 23 Feb 2023 • Zhichen Lai, Dalin Zhang, Huan Li, Christian S. Jensen, Hua Lu, Yan Zhao
Many deep learning models have been proposed to improve the accuracy of CTS forecasting.
Ranked #1 on
Traffic Prediction
on PeMS08
(FLOPs(M) metric)
Correlated Time Series Forecasting
Multivariate Time Series Forecasting
+2
3 code implementations • 13 Aug 2022 • Xingyu Xie, Pan Zhou, Huan Li, Zhouchen Lin, Shuicheng Yan
Adan first reformulates the vanilla Nesterov acceleration to develop a new Nesterov momentum estimation (NME) method, which avoids the extra overhead of computing gradient at the extrapolation point.
no code implementations • 15 Jun 2022 • Arpit Agarwal, Sanjeev Khanna, Huan Li, Prathamesh Patil
At the heart of our algorithmic results is a view of the objective in terms of cuts in the graph, which allows us to use a relaxed notion of cut sparsifiers to do hierarchical clustering while introducing only a small distortion in the objective function.
1 code implementation • 27 Jan 2022 • Huan Li, Zhouchen Lin
They do not invoke negative curvature exploitation or minimization of regularized surrogate functions as the subroutines.
no code implementations • NeurIPS 2021 • Anindya De, Sanjeev Khanna, Huan Li, MohammadHesam NikpeySalekde
We study the problem of minimizing a convex function given by a zeroth order oracle that is possibly corrupted by {\em outlier noise}.
no code implementations • 6 Apr 2021 • Huan Li, Zhouchen Lin
We prove the $O((\frac{\gamma}{1-\sigma_{\gamma}})^2\sqrt{\frac{L}{\epsilon}})$ and $O((\frac{\gamma}{1-\sigma_{\gamma}})^{1. 5}\sqrt{\frac{L}{\mu}}\log\frac{1}{\epsilon})$ complexities for the practical single loop accelerated gradient tracking over time-varying graphs when the problems are nonstrongly convex and strongly convex, respectively, where $\gamma$ and $\sigma_{\gamma}$ are two common constants charactering the network connectivity, $\epsilon$ is the desired precision, and $L$ and $\mu$ are the smoothness and strong convexity constants, respectively.
1 code implementation • 8 Oct 2020 • Tiantian Liu, Huan Li, Hua Lu, Muhammad Aamir Cheema, Lidan Shou
Indoor location-based services (LBS), such as POI search and routing, are often built on top of typical indoor spatial queries.
Databases Data Structures and Algorithms
no code implementations • 9 Sep 2020 • Huan Li, Zhouchen Lin, Yongchun Fang
Our stochastic gradient computation complexities are the same as the ones of single-machine VR methods, such as SAG, SAGA, and SVRG, and our communication complexities keep the same as those of EXTRA and DIGing, respectively.
no code implementations • 30 Apr 2020 • Fei Tang, Wanling Gao, Jianfeng Zhan, Chuanxin Lan, Xu Wen, Lei Wang, Chunjie Luo, Jiahui Dai, Zheng Cao, Xingwang Xiong, Zihan Jiang, Tianshu Hao, Fanda Fan, Fan Zhang, Yunyou Huang, Jianan Chen, Mengjia Du, Rui Ren, Chen Zheng, Daoyi Zheng, Haoning Tang, Kunlin Zhan, Biao Wang, Defei Kong, Minghe Yu, Chongkang Tan, Huan Li, Xinhui Tian, Yatao Li, Junchao Shao, Zhenyu Wang, Xiaoyu Wang, Hainan Ye
We use real-world benchmarks to cover the factors space that impacts the learning dynamics to the most considerable extent.
no code implementations • 24 Feb 2020 • Huan Li, Zhouchen Lin
EXTRA is a popular method for dencentralized distributed optimization and has broad applications.
no code implementations • 17 Feb 2020 • Wanling Gao, Fei Tang, Jianfeng Zhan, Chuanxin Lan, Chunjie Luo, Lei Wang, Jiahui Dai, Zheng Cao, Xiongwang Xiong, Zihan Jiang, Tianshu Hao, Fanda Fan, Xu Wen, Fan Zhang, Yunyou Huang, Jianan Chen, Mengjia Du, Rui Ren, Chen Zheng, Daoyi Zheng, Haoning Tang, Kunlin Zhan, Biao Wang, Defei Kong, Minghe Yu, Chongkang Tan, Huan Li, Xinhui Tian, Yatao Li, Gang Lu, Junchao Shao, Zhenyu Wang, Xiaoyu Wang, Hainan Ye
An end-to-end benchmark is a distillation of the essential attributes of an industry-scale application.
no code implementations • 13 Aug 2019 • Wanling Gao, Fei Tang, Lei Wang, Jianfeng Zhan, Chunxin Lan, Chunjie Luo, Yunyou Huang, Chen Zheng, Jiahui Dai, Zheng Cao, Daoyi Zheng, Haoning Tang, Kunlin Zhan, Biao Wang, Defei Kong, Tong Wu, Minghe Yu, Chongkang Tan, Huan Li, Xinhui Tian, Yatao Li, Junchao Shao, Zhenyu Wang, Xiaoyu Wang, Hainan Ye
On the basis of the AIBench framework, abstracting the real-world data sets and workloads from one of the top e-commerce providers, we design and implement the first end-to-end Internet service AI benchmark, which contains the primary modules in the critical paths of an industry scale application and is scalable to deploy on different cluster scales.
no code implementations • 6 Aug 2019 • Mihai Cucuringu, Huan Li, He Sun, Luca Zanetti
Graph clustering is a basic technique in machine learning, and has widespread applications in different domains.
no code implementations • 3 Oct 2018 • Huan Li, Yibo Yang, Dongmin Chen, Zhouchen Lin
In this paper, we propose the hypothesis that the neural network structure design can be inspired by optimization algorithms and a faster optimization algorithm may lead to a better neural network structure.
no code implementations • 17 Jun 2016 • Shoou-I Yu, Yi Yang, Zhongwen Xu, Shicheng Xu, Deyu Meng, Zexi Mao, Zhigang Ma, Ming Lin, Xuanchong Li, Huan Li, Zhenzhong Lan, Lu Jiang, Alexander G. Hauptmann, Chuang Gan, Xingzhong Du, Xiaojun Chang
The large number of user-generated videos uploaded on to the Internet everyday has led to many commercial video search engines, which mainly rely on text metadata for search.
no code implementations • NeurIPS 2015 • Huan Li, Zhouchen Lin
However, it is still unknown whether the usual APG can ensure the convergence to a critical point in nonconvex programming.
no code implementations • 14 Nov 2015 • Canyi Lu, Huan Li, Zhouchen Lin, Shuicheng Yan
The Augmented Lagragian Method (ALM) and Alternating Direction Method of Multiplier (ADMM) have been powerful optimization methods for general convex programming subject to linear constraint.
no code implementations • 13 Aug 2015 • Canyi Lu, Huan Li, Zhouchen Lin
To the best of our knowledge, this is the first work which directly minimizes the mutual coherence of the projected dictionary with a convergence guarantee.
no code implementations • 18 Oct 2013 • Zhouchen Lin, Risheng Liu, Huan Li
However, the traditional alternating direction method (ADM) and its linearized version (LADM, obtained by linearizing the quadratic penalty term) are for the two-block case and cannot be naively generalized to solve the multi-block case.