no code implementations • 27 Oct 2024 • Jing Zhang, Linjiajie Fang, Kexin Shi, Wenjia Wang, Bing-Yi Jing
A learning policy may take actions beyond the behavior policy's knowledge, referred to as Out-of-Distribution (OOD) actions.
1 code implementation • 31 May 2024 • Linjiajie Fang, Ruoxue Liu, Jing Zhang, Wenjia Wang, Bing-Yi Jing
In this paper, we propose Diffusion Actor-Critic (DAC) that formulates the Kullback-Leibler (KL) constraint policy iteration as a diffusion noise regression problem, enabling direct representation of target policies as diffusion models.
1 code implementation • NeurIPS 2023 • Jing Zhang, Chi Zhang, Wenjia Wang, Bing-Yi Jing
Due to the inability to interact with the environment, offline reinforcement learning (RL) methods face the challenge of estimating the Out-of-Distribution (OOD) points.
no code implementations • 10 Feb 2020 • Bing-Yi Jing, Ting Li, Zhongyuan Lyu, Dong Xia
We show that the TWIST procedure can accurately detect the communities with small misclassification error as the number of nodes and/or the number of layers increases.
no code implementations • 2 Sep 2017 • Ting Li, Bing-Yi Jing, Ningchen Ying, Xianshi Yu
Simulations are conducted to illustrate the advantages of our new scaling method.
no code implementations • 18 Jan 2015 • Jim Jing-Yan Wang, Yunji Wang, Bing-Yi Jing, Xin Gao
To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework.