no code implementations • 15 Nov 2024 • Hanzhong Guo, Jianfeng Zhang, Cheng Zou, Jun Li, Meng Wang, Ruxue Wen, Pingzhong Tang, Jingdong Chen, Ming Yang
A key challenge of try-on is to generate realistic images of the model wearing the garments while preserving the details of the garments.
no code implementations • 17 Nov 2021 • Hangyu Mao, Chao Wang, Xiaotian Hao, Yihuan Mao, Yiming Lu, Chengjie WU, Jianye Hao, Dong Li, Pingzhong Tang
The MineRL competition is designed for the development of reinforcement learning and imitation learning algorithms that can efficiently leverage human demonstrations to drastically reduce the number of environment interactions needed to solve the complex \emph{ObtainDiamond} task with sparse rewards.
no code implementations • 29 Sep 2021 • Mingyang Liu, Chengjie WU, Qihan Liu, Yansen Jing, Jun Yang, Pingzhong Tang, Chongjie Zhang
Search algorithms have been playing a vital role in the success of superhuman AI in both perfect information and imperfect information games.
1 code implementation • 3 Oct 2020 • Chuheng Zhang, Yuanqi Li, Xi Chen, Yifei Jin, Pingzhong Tang, Jian Li
Modern machine learning models (such as deep neural networks and boosting decision tree models) have become increasingly popular in financial market prediction, due to their superior capacity to extract complex non-linear patterns.
no code implementations • 9 Sep 2019 • Qingpeng Cai, Ling Pan, Pingzhong Tang
Based on this theoretical guarantee, we propose a class of the deterministic value gradient algorithm (DVG) with infinite horizon, and different rollout steps of the analytical gradients by the learned model trade off between the variance of the value gradients and the model bias.
no code implementations • 26 May 2019 • Feiyang Pan, Xiang Ao, Pingzhong Tang, Min Lu, Dapeng Liu, Lei Xiao, Qing He
It is often observed that the probabilistic predictions given by a machine learning model can disagree with averaged actual outcomes on specific subsets of data, which is also known as the issue of miscalibration.
no code implementations • ICLR 2019 • Qi Tan, Pingzhong Tang, Ke Xu, Weiran Shen, Song Zuo
Generative neural networks map a standard, possibly distribution to a complex high-dimensional distribution, which represents the real world data set.
1 code implementation • 25 Apr 2019 • Feiyang Pan, Shuokai Li, Xiang Ao, Pingzhong Tang, Qing He
We propose Meta-Embedding, a meta-learning-based approach that learns to generate desirable initial embeddings for new ad IDs.
no code implementations • 18 Nov 2018 • Feiyang Pan, Qingpeng Cai, An-Xiang Zeng, Chun-Xiang Pan, Qing Da, Hua-Lin He, Qing He, Pingzhong Tang
Model-free reinforcement learning methods such as the Proximal Policy Optimization algorithm (PPO) have successfully applied in complex decision-making problems such as Atari games.
no code implementations • 10 Jul 2018 • Qingpeng Cai, Ling Pan, Pingzhong Tang
Such a setting generalizes the widely-studied stochastic state transition setting, namely the setting of deterministic policy gradient (DPG).
no code implementations • 9 May 2018 • Weiran Shen, Pingzhong Tang, Song Zuo
We then apply our framework to a number of multi-item revenue optimal design settings, for a few of which the theoretically optimal mechanisms are unknown.
no code implementations • 13 Feb 2018 • Ling Pan, Qingpeng Cai, Zhixuan Fang, Pingzhong Tang, Longbo Huang
Different from existing methods that often ignore spatial information and rely heavily on accurate prediction, HRP captures both spatial and temporal dependencies using a divide-and-conquer structure with an embedded localized module.
no code implementations • 12 Feb 2018 • Feiyang Pan, Qingpeng Cai, Pingzhong Tang, Fuzhen Zhuang, Qing He
We evaluate PGCR on toy datasets as well as a real-world dataset of personalized music recommendations.
no code implementations • 25 Aug 2017 • Qingpeng Cai, Aris Filos-Ratsikas, Pingzhong Tang, Yiwei Zhang
We study the problem of allocating impressions to sellers in e-commerce websites, such as Amazon, eBay or Taobao, aiming to maximize the total revenue generated by the platform.
no code implementations • 6 Jul 2017 • Mengjing Chen, Weiran Shen, Pingzhong Tang, Song Zuo
To this end, we use a so-called "ironing" technique to convert the problem into an equivalent convex optimization one via a clean Markov decision process (MDP) formulation, where the states are the driver distributions and the decision variables are the prices for each pair of locations.