no code implementations • 23 Feb 2023 • Vincent Liu, Yash Chandak, Philip Thomas, Martha White
In this work, we consider the off-policy policy evaluation problem for contextual bandits and finite horizon reinforcement learning in the nonstationary setting.
1 code implementation • 22 Feb 2023 • Zhuohan Li, Lianmin Zheng, Yinmin Zhong, Vincent Liu, Ying Sheng, Xin Jin, Yanping Huang, Zhifeng Chen, Hao Zhang, Joseph E. Gonzalez, Ion Stoica
Model parallelism is conventionally viewed as a method to scale a single large deep learning model beyond the memory limits of a single device.
no code implementations • 18 May 2022 • Han Wang, Archit Sakhadeo, Adam White, James Bell, Vincent Liu, Xutong Zhao, Puer Liu, Tadashi Kozuno, Alona Fyshe, Martha White
The performance of reinforcement learning (RL) agents is sensitive to the choice of hyperparameters.
no code implementations • 30 Mar 2022 • Han Wang, Erfan Miahi, Martha White, Marlos C. Machado, Zaheer Abbas, Raksha Kumaraswamy, Vincent Liu, Adam White
In this paper we investigate the properties of representations learned by deep reinforcement learning systems.
1 code implementation • 23 Nov 2021 • Alex Tamkin, Vincent Liu, Rongfei Lu, Daniel Fein, Colin Schultz, Noah Goodman
Self-supervised learning algorithms, including BERT and SimCLR, have enabled significant strides in fields like natural language processing, computer vision, and speech processing.
Ranked #1 on
Self-Supervised Learning
on DABS
no code implementations • 15 Nov 2021 • Vincent Liu, James R. Wright, Martha White
Offline reinforcement learning -- learning a policy from a batch of data -- is known to be hard for general MDPs.
no code implementations • 1 Jan 2021 • Vincent Liu, Adam M White, Hengshuai Yao, Martha White
Catastrophic interference is common in many network-based learning systems, and many proposals exist for mitigating it.
no code implementations • 7 Jul 2020 • Vincent Liu, Adam White, Hengshuai Yao, Martha White
In this work, we provide a definition of interference for control in reinforcement learning.
no code implementations • 2 Jun 2020 • Alejandro Schuler, Aashish Bhardwaj, Vincent Liu
Clinical researchers often select among and evaluate risk prediction models using standard machine learning metrics based on confusion matrices.
no code implementations • ICLR 2020 • Somjit Nath, Vincent Liu, Alan Chan, Xin Li, Adam White, Martha White
Recurrent neural networks (RNNs) allow an agent to construct a state-representation from a stream of experience, which is essential in partially observable problems.
no code implementations • 5 Jul 2019 • Brendan Bennett, Wesley Chung, Muhammad Zaheer, Vincent Liu
Temporal difference methods enable efficient estimation of value functions in reinforcement learning in an incremental fashion, and are of broader interest because they correspond learning as observed in biological systems.
no code implementations • 6 Jan 2019 • Vincent Liu, Ademi Adeniji, Nathaniel Lee, Jason Zhao, Mario Srouji
Central Pattern Generators (CPGs) are biological neural circuits capable of producing coordinated rhythmic outputs in the absence of rhythmic input.
no code implementations • 15 Nov 2018 • Vincent Liu, Raksha Kumaraswamy, Lei Le, Martha White
We investigate sparse representations for control in reinforcement learning.
no code implementations • 20 Oct 2018 • Wen-Hao Chen, Chin-Chi Hsu, Yi-An Lai, Vincent Liu, Mi-Yen Yeh, Shou-De Lin
Attribute-aware CF models aims at rating prediction given not only the historical rating from users to items, but also the information associated with users (e. g. age), items (e. g. price), or even ratings (e. g. rating time).