no code implementations • 10 Apr 2024 • Qi Deng, Zheng Fan, Zhi Li, Xinna Pan, Qi Kang, Mengchu Zhou
The application of evolutionary algorithms (EAs) to multi-objective optimization problems has been widespread.
no code implementations • 24 Mar 2024 • Qi Deng, Zhong-guo Zhou
We propose that the liquidity of an asset includes two components: liquidity jump and liquidity diffusion.
no code implementations • 25 Jan 2024 • Wenzhi Gao, Qi Deng
This paper considers stochastic weakly convex optimization without the standard Lipschitz continuity assumption.
1 code implementation • 21 Aug 2023 • Zhenwei Lin, Qi Deng
In this paper, we adapt the GBM for constrained learning tasks within the framework of Bregman proximal algorithms.
no code implementations • 21 Aug 2023 • Jiyuan Tan, Chenyu Xue, Chuwen Zhang, Qi Deng, Dongdong Ge, Yinyu Ye
In this paper, we propose the stochastic homogeneous second-order descent method (SHSODM) for stochastic functions enjoying gradient dominance property based on a recently proposed homogenization approach.
no code implementations • 18 Jul 2023 • Qi Deng, Linhong Zheng, Jiaqi Peng, Xu Li, Zhong-guo Zhou, Monica Hussein, Dingyi Chen, Mick Swartz
We find that the most efficient regulation regime in Chinese IPO pricing has four characteristics: 1) registration system, 2) no hard return caps nor trading curbs that restrict the initial return; 3) more specific listing rules for issuers, and 4) more stringent participation requirements for investors.
no code implementations • 27 Jun 2023 • Qi Deng, Zhong-guo Zhou
We establish innovative liquidity premium measures, and construct liquidity-adjusted return and volatility to model assets with extreme liquidity, represented by a portfolio of selected crypto assets, and upon which we develop a set of liquidity-adjusted ARMA-GARCH/EGARCH models.
no code implementations • 10 Apr 2023 • Digvijay Boob, Qi Deng
Second, to obtain the optimal operator complexity for smooth deterministic problems, we present a novel single-loop Adaptive Lagrangian Extrapolation~(\texttt{AdLagEx}) method that can adaptively search for and explicitly bound the Lagrange multipliers.
no code implementations • 28 Jan 2023 • Jinsong Liu, Chenghan Xie, Qi Deng, Dongdong Ge, Yinyu Ye
In this paper, we propose several new stochastic second-order algorithms for policy optimization that only require gradient and Hessian-vector product in each iteration, making them computationally efficient and comparable to policy gradient methods.
no code implementations • 21 Dec 2022 • Zhenwei Lin, Qi Deng
Before our work, the best complexity bound was $\mathcal{O}(1/{\varepsilon})$, and it remains unclear how to improve this result by leveraging the strong convexity assumption.
no code implementations • 19 Nov 2022 • Qi Deng, Lunge Dai, Zixin Yang, Zhong-guo Zhou, Monica Hussein, Dingyi Chen, Mick Swartz
Based on our findings, we argue that the differences among the levels and determinants of initial return, monthly return (intrinsic value) and intramonth return (overreaction) in different time periods can be largely explained by regulation regime changes along two dimensions: 1) approval vs. registration and 2) listing day trading curbs and return limits.
no code implementations • NeurIPS 2021 • Qi Deng, Wenzhi Gao
Second, motivated by the success of momentum stochastic gradient descent, we propose a new stochastic extrapolated model-based method, greatly extending the classic Polyak momentum technique to a wider class of stochastic algorithms for weakly convex optimization.
no code implementations • NeurIPS 2020 • Digvijay Boob, Qi Deng, Guanghui Lan, Yilin Wang
We also establish new convergence complexities to achieve an approximate KKT solution when the objective can be smooth/nonsmooth, deterministic/stochastic and convex/nonconvex with complexity that is on a par with gradient descent for unconstrained optimization problems in respective cases.
no code implementations • 26 Sep 2019 • Qi Deng
The AIBC implements a two-consensus scheme to enforce upper-layer economic policies and achieve fundamental layer performance and robustness: the DPoEV incentive consensus on the application and resource layers, and the DABFT distributed consensus on the fundamental layer.
no code implementations • 3 Sep 2019 • Qi Deng, Chenghao Lan
Third, we extend accelerated coordinate descent (ACD) to nonsmooth and nonconvex optimization to develop a novel randomized proximal DC algorithm whereby we solve the subproblem inexactly by ACD.
no code implementations • 7 Aug 2019 • Digvijay Boob, Qi Deng, Guanghui Lan
For large-scale and stochastic problems, we present a more practical proximal point method in which the approximate solutions of the subproblems are computed by the aforementioned ConEx method.
1 code implementation • 1 Oct 2018 • Qi Deng, Yi Cheng, Guanghui Lan
More specifically, we show that diagonal scaling, initially designed to improve vanilla stochastic gradient, can be incorporated into accelerated stochastic gradient descent to achieve the optimal rate of convergence for smooth stochastic optimization.