no code implementations • 6 Jun 2023 • Minting Pan, Yitao Zheng, Wendong Zhang, Yunbo Wang, Xiaokang Yang
Pretraining RL models on offline video datasets is a promising way to improve their training efficiency in online tasks, but challenging due to the inherent mismatch in tasks, dynamics, and behaviors across domains.
1 code implementation • 27 Mar 2023 • Minting Pan, Xiangming Zhu, Yitao Zheng, Yunbo Wang, Xiaokang Yang
On top of our previous work, we further consider the sparse dependencies between controllable and noncontrollable states, address the training collapse problem of state decoupling, and validate our approach in transfer learning setups.
1 code implementation • 9 Dec 2020 • Qi Zhou, Haipeng Chen, Yitao Zheng, Zhen Wang
As one of the most powerful topic models, Latent Dirichlet Allocation (LDA) has been used in a vast range of tasks, including document understanding, information retrieval and peer-reviewer assignment.