no code implementations • 18 Apr 2023 • Jing An, Jianfeng Lu
We study the convergence of stochastic gradient descent (SGD) for non-convex objective functions.
1 code implementation • 6 Mar 2023 • Pierre Bréchet, Katerina Papagiannouli, Jing An, Guido Montúfar
We consider a deep matrix factorization model of covariance matrices trained with the Bures-Wasserstein distance.
no code implementations • 31 May 2021 • Jing An, Lexing Ying
When the loss function is a sum of multiple terms, a popular method is the stochastic gradient descent.
no code implementations • 11 May 2021 • Xiaolong Wei, Lifang Yang, Xianglin Huang, Gang Cao, Tao Zhulin, Zhengyang Du, Jing An
This paper proposed a hierarchical transformers MADDPG based on RNN which we call it Hierarchical RNNs-Based Transformers MADDPG(HRTMADDPG).
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • ICLR 2021 • Jing An, Lexing Ying, Yuhua Zhu
We consider two commonly-used techniques, resampling and reweighting, that rebalance the proportions of the subgroups to maintain the desired objective function.
no code implementations • 21 May 2018 • Jing An, Jianfeng Lu, Lexing Ying
The resulting SME of Langevin type extracts more information about the ASGD dynamics and elucidates the relationship between different types of stochastic gradient algorithms.