no code implementations • 28 Aug 2023 • Falcon Z. Dai
In our inquiry of various kinds of "costs" associated with reinforcement learning inspired by the demands in robotic applications, rewards are central to understanding the structure of a Markov decision process and reward-centric notions can elucidate important concepts in reinforcement learning.
no code implementations • 24 Oct 2020 • Falcon Z. Dai
Being inspired by the success of \texttt{word2vec} \citep{mikolov2013distributed} in capturing analogies, we study the conjecture that analogical relations can be represented by vector spaces.
1 code implementation • 15 Feb 2020 • Falcon Z. Dai, Matthew R. Walter
At the working heart of policy iteration algorithms commonly used and studied in the discounted setting of reinforcement learning, the policy evaluation step estimates the value of states with samples from a Markov reward process induced by following a Markov policy in a Markov decision process.
2 code implementations • 1 Aug 2019 • Igor Vasiljevic, Nick Kolkin, Shanyi Zhang, Ruotian Luo, Haochen Wang, Falcon Z. Dai, Andrea F. Daniele, Mohammadreza Mostajabi, Steven Basart, Matthew R. Walter, Gregory Shakhnarovich
We introduce DIODE, a dataset that contains thousands of diverse high resolution color images with accurate, dense, long-range depth measurements.
1 code implementation • ACL 2019 • Falcon Z. Dai, Zheng Cai
We show that the imperceptibility of several existing linguistic steganographic systems (Fang et al., 2017; Yang et al., 2018) relies on implicit assumptions on statistical behaviors of fluent text.
no code implementations • 3 Jul 2019 • Falcon Z. Dai, Matthew R. Walter
We propose a new complexity measure for Markov decision processes (MDPs), the maximum expected hitting cost (MEHC).
1 code implementation • WS 2018 • Sebastian Gehrmann, Falcon Z. Dai, Henry Elder, Alexander M. Rush
Learning to generate fluent natural language from structured data with neural networks has become an common approach for NLG.
1 code implementation • WS 2017 • Falcon Z. Dai, Zheng Cai
Given the advantage and recent success of English character-level and subword-unit models in several NLP tasks, we consider the equivalent modeling problem for Chinese.