no code implementations • CVPR 2022 • Jialian Li, Jingyi Zhang, Zhiyong Wang, Siqi Shen, Chenglu Wen, Yuexin Ma, Lan Xu, Jingyi Yu, Cheng Wang
Quantitative and qualitative experiments show that our method outperforms the techniques based only on RGB images.
Ranked #3 on 3D Human Pose Estimation on SLOPER4D (using extra training data)
no code implementations • 13 Mar 2022 • Jialian Li, Tongzheng Ren, Dong Yan, Hang Su, Jun Zhu
Our goal is to identify a near-optimal robust policy for the perturbed testing environment, which introduces additional technical difficulties as we need to simultaneously estimate the training environment uncertainty from samples and find the worst-case perturbation for testing.
no code implementations • NeurIPS 2021 • Tongzheng Ren, Jialian Li, Bo Dai, Simon S. Du, Sujay Sanghavi
To the best of our knowledge, these are the \emph{first} set of nearly horizon-free bounds for episodic time-homogeneous offline tabular MDP and linear MDP with anchor points.
no code implementations • ICLR 2020 • Yichi Zhou, Jialian Li, Jun Zhu
Posterior sampling for reinforcement learning (PSRL) is a useful framework for making decisions in an unknown environment.
no code implementations • ICLR 2020 • Yichi Zhou, Tongzheng Ren, Jialian Li, Dong Yan, Jun Zhu
In this paper, we present Lazy-CFR, a CFR algorithm that adopts a lazy update strategy to avoid traversing the whole game tree in each round.
no code implementations • 20 May 2019 • Yangbin Lin, Jialian Li, Cheng Wang, Zhonggui Chen, Zongyue Wang, Jonathan Li
Man-made environments typically comprise planar structures that exhibit numerous geometric relationships, such as parallelism, coplanarity, and orthogonality.
no code implementations • ICLR 2019 • Jialian Li, Hang Su, Jun Zhu
We can solve these tasks by first building models for other agents and then finding the optimal policy with these models.
no code implementations • 10 Oct 2018 • Yichi Zhou, Tongzheng Ren, Jialian Li, Dong Yan, Jun Zhu
In this paper, we present a novel technique, lazy update, which can avoid traversing the whole game tree in CFR, as well as a novel analysis on the regret of CFR with lazy update.
no code implementations • ICML 2017 • Yichi Zhou, Jialian Li, Jun Zhu
We study the problem on how to learn the pure Nash Equilibrium of a two-player zero-sum static game with random payoffs under unknown distributions via efficient payoff queries.
1 code implementation • 28 Jun 2017 • Haosheng Zou, Kun Xu, Jialian Li, Jun Zhu
We took part in the YouTube-8M Video Understanding Challenge hosted on Kaggle, and achieved the 10th place within less than one month's time.
no code implementations • NeurIPS 2016 • Yong Ren, Jialian Li, Yucen Luo, Jun Zhu
Maximum mean discrepancy (MMD) has been successfully applied to learn deep generative models for characterizing a joint distribution of variables via kernel mean embedding.