no code implementations • 13 Mar 2023 • Mrigank Raman, Pratyush Maini, J. Zico Kolter, Zachary C. Lipton, Danish Pruthi
In recent years, NLP practitioners have converged on the following practice: (i) import an off-the-shelf pretrained (masked) language model; (ii) append a multilayer perceptron atop the CLS token's hidden representation (with randomly initialized weights); and (iii) fine-tune the entire model on a downstream task (MLP).
no code implementations • 9 Mar 2023 • Alex Mei, Michael Saxon, Shiyu Chang, Zachary C. Lipton, William Yang Wang
We conduct a broad literature survey, identifying many clusters of similar conceptions of transparency, tying each back to our north star with analysis of how it furthers or hinders our ideal AI transparency goals.
1 code implementation • 16 Feb 2023 • Shantanu Gupta, David Childers, Zachary C. Lipton
Even when the causal graph underlying our data is unknown, we can use observational data to narrow down the possible values that an average treatment effect (ATE) can take by (1) identifying the graph up to a Markov equivalence class; and (2) estimating that ATE for each graph in the class.
1 code implementation • 6 Feb 2023 • Zachary Novack, Saurabh Garg, Julian McAuley, Zachary C. Lipton
Open vocabulary models (e. g.
1 code implementation • 6 Feb 2023 • Saurabh Garg, Nick Erickson, James Sharpnack, Alex Smola, Sivaraman Balakrishnan, Zachary C. Lipton
Meanwhile, popular deep domain adaptation heuristics tend to falter when faced with shifts in label proportions.
no code implementations • 27 Jan 2023 • Jacob Tyo, Zachary C. Lipton
Supervised learning typically optimizes the expected value risk functional of the loss, but in many cases, we want to optimize for other risk functionals.
1 code implementation • 29 Nov 2022 • Zachary Novack, Simran Kaur, Tanya Marwah, Saurabh Garg, Zachary C. Lipton
A number of competing hypotheses have been proposed to explain why small-batch Stochastic Gradient Descent (SGD)leads to improved generalization over the full-batch regime, with recent work crediting the implicit regularization of various quantities throughout training.
1 code implementation • 14 Nov 2022 • Helen Zhou, YuWen Chen, Zachary C. Lipton
Machine learning models deployed in healthcare systems face data drawn from continually evolving environments.
no code implementations • 3 Nov 2022 • Helen Zhou, Sivaraman Balakrishnan, Zachary C. Lipton
Rates of missing data often depend on record-keeping policies and thus may change across times and locations, even when the underlying features are comparatively stable.
1 code implementation • 26 Oct 2022 • Pratyush Maini, Saurabh Garg, Zachary C. Lipton, J. Zico Kolter
Popular metrics derived from these dynamics include (i) the epoch at which examples are first correctly classified; (ii) the number of times their predictions flip during training; and (iii) whether their prediction flips if they are held out.
no code implementations • 21 Oct 2022 • Tanya Marwah, Zachary C. Lipton, Jianfeng Lu, Andrej Risteski
We show that if composing a function with Barron norm $b$ with $L$ produces a function of Barron norm at most $B_L b^p$, the solution to the PDE can be $\epsilon$-approximated in the $L^2$ sense by a function with Barron norm $O\left(\left(dB_L\right)^{p^{\log(1/\epsilon)}}\right)$.
no code implementations • 4 Oct 2022 • Rasool Fakoor, Jonas Mueller, Zachary C. Lipton, Pratik Chaudhari, Alexander J. Smola
We introduce a novel time-varying importance weight estimator that can detect gradual shifts in the distribution of data.
no code implementations • 28 Sep 2022 • Kundan Krishna, Saurabh Garg, Jeffrey P. Bigham, Zachary C. Lipton
For most natural language processing tasks, the dominant practice is to finetune large pretrained transformer models (e. g., BERT) using smaller downstream datasets.
1 code implementation • 14 Sep 2022 • Jacob Tyo, Bhuwan Dhingra, Zachary C. Lipton
Despite decades of research on authorship attribution (AA) and authorship verification (AV), inconsistent dataset splits/filtering and mismatched evaluation methods make it difficult to assess the state of the art.
no code implementations • 28 Aug 2022 • Helen Zhou, Cheng Cheng, Kelly J. Shields, Gursimran Kochhar, Tariq Cheema, Zachary C. Lipton, Jeremy C. Weiss
With COVID-19 now pervasive, identification of high-risk individuals is crucial.
2 code implementations • 26 Jul 2022 • Manley Roberts, Pranav Mani, Saurabh Garg, Zachary C. Lipton
Thus motivated, we introduce a practical algorithm that leverages domain-discriminative models as follows: (i) push examples through domain discriminator $p(d|\mathbf{x})$; (ii) discretize the data by clustering examples in $p(d|\mathbf{x})$ space; (iii) perform non-negative matrix factorization on the discrete data; (iv) combine the recovered $p(y|d)$ with the discriminator outputs $p(d|\mathbf{x})$ to compute $p_d(y|x) \; \forall d$.
1 code implementation • 26 Jul 2022 • Saurabh Garg, Sivaraman Balakrishnan, Zachary C. Lipton
We introduce the problem of domain adaptation under Open Set Label Shift (OSLS) where the label distribution can change arbitrarily and a new class may arrive during deployment, but the class-conditional distributions p(x|y) are domain-invariant.
no code implementations • 27 Jun 2022 • Liu Leqi, Audrey Huang, Zachary C. Lipton, Kamyar Azizzadenesheli
Standard uniform convergence results bound the generalization gap of the expected loss over a hypothesis class.
no code implementations • 21 Jun 2022 • Simran Kaur, Jeremy Cohen, Zachary C. Lipton
The mechanisms by which certain training interventions, such as increasing learning rates and applying batch normalization, improve the generalization of deep networks remains a mystery.
no code implementations • 8 Jun 2022 • Divyansh Kaushik, Zachary C. Lipton, Alex John London
In recent years, machine learning (ML) has come to rely more heavily on crowdworkers, both for building bigger datasets and for addressing research questions requiring human interaction or judgment.
no code implementations • 25 Mar 2022 • Omer Ben-Porat, Lee Cohen, Liu Leqi, Zachary C. Lipton, Yishay Mansour
We first address the case where all users share the same type, demonstrating that a recent UCB-based algorithm is optimal.
1 code implementation • 2 Feb 2022 • Yi-Fan Zhang, HANLIN ZHANG, Zachary C. Lipton, Li Erran Li, Eric P. Xing
Previous works on Treatment Effect Estimation (TEE) are not in widespread use because they are predominantly theoretical, where strong parametric assumptions are made but untractable for practical application.
1 code implementation • ICLR 2022 • Saurabh Garg, Sivaraman Balakrishnan, Zachary C. Lipton, Behnam Neyshabur, Hanie Sedghi
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions that may cause performance drops.
1 code implementation • 17 Dec 2021 • Siddhant Arora, Danish Pruthi, Norman Sadeh, William W. Cohen, Zachary C. Lipton, Graham Neubig
Through our evaluation, we observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
1 code implementation • NeurIPS 2021 • Saurabh Garg, Yifan Wu, Alex Smola, Sivaraman Balakrishnan, Zachary C. Lipton
Formally, this task is broken down into two subtasks: (i) Mixture Proportion Estimation (MPE) -- determining the fraction of positive examples in the unlabeled data; and (ii) PU-learning -- given such an estimate, learning the desired positive-versus-negative classifier.
no code implementations • 14 Oct 2021 • Anurag Katakkar, Clay H. Yoo, Weiqin Wang, Zachary C. Lipton, Divyansh Kaushik
In attempts to develop sample-efficient and interpretable algorithms, researcher have explored myriad mechanisms for collecting and exploiting feature feedback (or rationales) auxiliary annotations provided for training (but not test) instances that highlight salient evidence.
1 code implementation • Findings (EMNLP) 2021 • Kundan Krishna, Jeffrey Bigham, Zachary C. Lipton
Pretraining techniques leveraging enormous datasets have driven recent advances in text summarization.
1 code implementation • NeurIPS 2021 • Shantanu Gupta, Zachary C. Lipton, David Childers
Researchers often face data fusion problems, where multiple data sources are available, each capturing a distinct subset of variables.
1 code implementation • 21 Jun 2021 • Aston Zhang, Zachary C. Lipton, Mu Li, Alexander J. Smola
This open-source book represents our attempt to make deep learning approachable, teaching readers the concepts, the context, and the code.
1 code implementation • 13 Jun 2021 • Shantanu Gupta, Hao Wang, Zachary C. Lipton, Yuyang Wang
Link prediction methods are frequently applied in recommender systems, e. g., to suggest citations for academic papers or friends in social networks.
1 code implementation • ACL 2021 • Divyansh Kaushik, Douwe Kiela, Zachary C. Lipton, Wen-tau Yih
In adversarial data collection (ADC), a human workforce interacts with a model in real time, attempting to produce examples that elicit incorrect predictions.
1 code implementation • 1 May 2021 • Saurabh Garg, Sivaraman Balakrishnan, J. Zico Kolter, Zachary C. Lipton
To assess generalization, machine learning scientists typically either (i) bound the generalization gap and then (after training) plug in the empirical risk to obtain a bound on the true risk; or (ii) validate empirically on holdout data.
no code implementations • NeurIPS 2021 • Audrey Huang, Liu Leqi, Zachary C. Lipton, Kamyar Azizzadenesheli
Even when unable to run experiments, practitioners can evaluate prospective policies, using previously logged data.
no code implementations • 4 Mar 2021 • Audrey Huang, Liu Leqi, Zachary C. Lipton, Kamyar Azizzadenesheli
Because optimizing the coherent risk is difficult in Markov decision processes, recent work tends to focus on the Markov coherent risk (MCR), a time-consistent surrogate.
no code implementations • NeurIPS 2021 • Tanya Marwah, Zachary C. Lipton, Andrej Risteski
Recent experiments have shown that deep networks can approximate solutions to high-dimensional PDEs, seemingly escaping the curse of dimensionality.
no code implementations • 20 Feb 2021 • Saurabh Garg, Joshua Zhanson, Emilio Parisotto, Adarsh Prasad, J. Zico Kolter, Zachary C. Lipton, Sivaraman Balakrishnan, Ruslan Salakhutdinov, Pradeep Ravikumar
In this paper, we present a detailed empirical study to characterize the heavy-tailed nature of the gradients of the PPO surrogate reward function.
1 code implementation • 1 Dec 2020 • Danish Pruthi, Rachit Bansal, Bhuwan Dhingra, Livio Baldini Soares, Michael Collins, Zachary C. Lipton, Graham Neubig, William W. Cohen
While many methods purport to explain predictions by highlighting salient features, what aims these explanations serve and how they ought to be evaluated often go unstated.
no code implementations • 26 Nov 2020 • Nicholas Roberts, Davis Liang, Graham Neubig, Zachary C. Lipton
This makes human-level BLEU a misleading benchmark in that modern MT systems cannot approach human-level BLEU while simultaneously maintaining human-level translation diversity.
no code implementations • NeurIPS 2021 • Liu Leqi, Fatma Kilinc-Karzan, Zachary C. Lipton, Alan L. Montgomery
Psychological research shows that enjoyment of many goods is subject to satiation, with short-term satisfaction declining after repeated exposures to the same item.
no code implementations • 7 Nov 2020 • Jessica Dai, Sina Fazelpour, Zachary C. Lipton
If k% of employers were to voluntarily adopt a fairness-promoting intervention, should we expect k% progress (in aggregate) towards the benefits of universal adoption, or will the dynamics of partial compliance wash out the hoped-for benefits?
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Danish Pruthi, Bhuwan Dhingra, Graham Neubig, Zachary C. Lipton
For many prediction tasks, stakeholders desire not only predictions but also supporting evidence that a human can use to verify its correctness.
no code implementations • EMNLP 2021 • David Lowell, Brian E. Howard, Zachary C. Lipton, Byron C. Wallace
Unsupervised Data Augmentation (UDA) is a semi-supervised technique that applies a consistency loss to penalize differences between a model's predictions on (a) observed (unlabeled) examples; and (b) corresponding 'noised' examples produced via data augmentation.
1 code implementation • EMNLP 2020 • ZiRui Wang, Zachary C. Lipton, Yulia Tsvetkov
Modern multilingual models are trained on concatenated text from multiple languages in hopes of conferring benefits to each (positive transfer), with the most pronounced benefits accruing to low-resource languages.
no code implementations • ICLR 2021 • Divyansh Kaushik, Amrith Setlur, Eduard Hovy, Zachary C. Lipton
In attempts to produce ML models less reliant on spurious patterns in NLP datasets, researchers have recently proposed curating counterfactually augmented data (CAD) via a human-in-the-loop process in which given some documents and their (initial) labels, humans must revise the text to make a counterfactual label applicable.
no code implementations • 14 Jul 2020 • Kundan Krishna, Amy Pavel, Benjamin Schloss, Jeffrey P. Bigham, Zachary C. Lipton
In this exploratory study, we describe a new dataset consisting of conversation transcripts, post-visit summaries, corresponding supporting evidence (in the transcript), and structured labels.
no code implementations • 7 Jul 2020 • Lakshay Chauhan, John Alberg, Zachary C. Lipton
On a periodic basis, publicly traded companies report fundamentals, financial data including revenue, earnings, debt, among others.
1 code implementation • 2 Jun 2020 • Helen Zhou, Cheng Cheng, Zachary C. Lipton, George H. Chen, Jeremy C. Weiss
Finally, the PEER score is provided in the form of a nomogram for direct calculation of patient risk, and can be used to highlight at-risk patients among critical care patients eligible for ECMO.
no code implementations • ACL 2021 • Kundan Krishna, Sopan Khosla, Jeffrey P. Bigham, Zachary C. Lipton
Following each patient visit, physicians draft long semi-structured clinical summaries called SOAP notes.
no code implementations • 26 Mar 2020 • Shantanu Gupta, Zachary C. Lipton, David Childers
We show that it strictly outperforms the backdoor and frontdoor estimators and that this improvement can be unbounded.
no code implementations • NeurIPS 2020 • Saurabh Garg, Yifan Wu, Sivaraman Balakrishnan, Zachary C. Lipton
Our contributions include (i) consistency conditions for MLLS, which include calibration of the classifier and a confusion matrix invertibility condition that BBSE also requires; (ii) a unified framework, casting BBSE as roughly equivalent to MLLS for a particular choice of calibration method; and (iii) a decomposition of MLLS's finite-sample error into terms reflecting miscalibration and estimation error.
no code implementations • 25 Feb 2020 • Kyra Gan, Andrew A. Li, Zachary C. Lipton, Sridhar Tayur
In this paper, we consider the benefit of incorporating a large confounded observational dataset (confounder unobserved) alongside a small deconfounded observational dataset (confounder revealed) when estimating the ATE.
no code implementations • 8 Jan 2020 • Sina Fazelpour, Zachary C. Lipton
Inspired by recent breakthroughs in predictive modeling, practitioners in both industry and government have turned to machine learning with hopes of operationalizing predictions to drive automated decisions.
no code implementations • ICLR 2020 • Haohan Wang, Xindi Wu, Songwei Ge, Zachary C. Lipton, Eric P. Xing
Recent research has shown that CNNs are often overly sensitive to high-frequency textural patterns.
no code implementations • NeurIPS 2019 • Fan Yang, Liu Leqi, Yifan Wu, Zachary C. Lipton, Pradeep Ravikumar, William W. Cohen, Tom Mitchell
The ability to inferring latent psychological traits from human behavior is key to developing personalized human-interacting machine learning systems.
no code implementations • 18 Oct 2019 • Simran Kaur, Jeremy Cohen, Zachary C. Lipton
For a standard convolutional neural network, optimizing over the input pixels to maximize the score of some target class will generally produce a grainy-looking version of the original image.
1 code implementation • 2 Oct 2019 • Angela H. Jiang, Daniel L. -K. Wong, Giulio Zhou, David G. Andersen, Jeffrey Dean, Gregory R. Ganger, Gauri Joshi, Michael Kaminksy, Michael Kozuch, Zachary C. Lipton, Padmanabhan Pillai
This paper introduces Selective-Backprop, a technique that accelerates the training of deep neural networks (DNNs) by prioritizing examples with high loss at each iteration.
2 code implementations • ICLR 2020 • Divyansh Kaushik, Eduard Hovy, Zachary C. Lipton
While classifiers trained on either original or manipulated data alone are sensitive to spurious features (e. g., mentions of genre), models trained on the combined data are less sensitive to this signal.
3 code implementations • ACL 2020 • Danish Pruthi, Mansi Gupta, Bhuwan Dhingra, Graham Neubig, Zachary C. Lipton
Attention mechanisms are ubiquitous components in neural architectures applied to natural language processing.
1 code implementation • IJCNLP 2019 • Alankar Jain, Bhargavi Paranjape, Zachary C. Lipton
Although over 100 languages are supported by strong off-the-shelf machine translation systems, only a subset of them possess large annotated corpora for named entity recognition.
1 code implementation • 12 Aug 2019 • Mansi Gupta, Nitish Kulkarni, Raghuveer Chanda, Anirudha Rayasam, Zachary C. Lipton
Observing that many questions can be answered based upon the available product reviews, we propose the task of review-based QA.
no code implementations • 1 Jul 2019 • Xinyang Feng, Zachary C. Lipton, Jie Yang, Scott A. Small, Frank A. Provenzano
Numerous studies have established that estimated brain age, as derived from statistical models trained on healthy populations, constitutes a valuable biomarker that is predictive of cognitive decline and various neurological diseases.
no code implementations • 25 Jun 2019 • Amy Zhang, Zachary C. Lipton, Luis Pineda, Kamyar Azizzadenesheli, Anima Anandkumar, Laurent Itti, Joelle Pineau, Tommaso Furlanello
In this paper, we propose an algorithm to approximate causal states, which are the coarsest partition of the joint history of actions and observations in partially-observable Markov decision processes (POMDP).
4 code implementations • NeurIPS 2019 • Haohan Wang, Songwei Ge, Eric P. Xing, Zachary C. Lipton
Despite their renowned predictive power on i. i. d.
Ranked #90 on
Domain Generalization
on PACS
no code implementations • 27 May 2019 • Lee Cohen, Zachary C. Lipton, Yishay Mansour
We analyze the optimal employer policy both when the employer sets a fixed number of tests per candidate and when the employer can set a dynamic policy, assigning further tests adaptively based on results from the previous tests.
3 code implementations • ACL 2019 • Danish Pruthi, Bhuwan Dhingra, Zachary C. Lipton
To combat adversarial spelling mistakes, we propose placing a word recognition model in front of the downstream classifier.
no code implementations • 9 Apr 2019 • Tingfung Lau, Nathan Ng, Julian Gingold, Nina Desai, Julian McAuley, Zachary C. Lipton
First, noting that in each image the embryo occupies a small subregion, we jointly train a region proposal network with the downstream classifier to isolate the embryo.
no code implementations • ICLR 2019 • Haohan Wang, Zexue He, Zachary C. Lipton, Eric P. Xing
We test our method on the battery of standard domain generalization data sets and, interestingly, achieve comparable or better performance as compared to other domain generalization methods that explicitly require samples from the target distribution for training.
Ranked #99 on
Domain Generalization
on PACS
1 code implementation • 8 Dec 2018 • Jonathon Byrd, Zachary C. Lipton
Importance-weighted risk minimization is a key ingredient in many machine learning algorithms for causal inference, domain adaptation, class imbalance, and off-policy reinforcement learning.
1 code implementation • NeurIPS 2019 • Stephan Rabanser, Stephan Günnemann, Zachary C. Lipton
We might hope that when faced with unexpected inputs, well-designed software systems would fire off warnings.
no code implementations • EMNLP 2018 • Aditya Siddhant, Zachary C. Lipton
This paper provides a large scale empirical study of deep active learning, addressing multiple tasks and, for each, multiple datasets, multiple models, and a full suite of acquisition functions.
no code implementations • EMNLP 2018 • Divyansh Kaushik, Zachary C. Lipton
Many recent papers address reading comprehension, where examples consist of (question, passage, answer) tuples.
no code implementations • 17 Jul 2018 • Davis Liang, Zhiheng Huang, Zachary C. Lipton
Despite rapid advances in speech recognition, current models remain brittle to superficial perturbations to their inputs.
no code implementations • IJCNLP 2019 • David Lowell, Zachary C. Lipton, Byron C. Wallace
Active learning (AL) is a widely-used training strategy for maximizing predictive performance subject to a fixed annotation budget.
no code implementations • 9 Jul 2018 • Zachary C. Lipton, Jacob Steinhardt
Collectively, machine learning (ML) researchers are engaged in the creation and dissemination of knowledge about data-driven algorithms.
3 code implementations • ICLR 2019 • Kamyar Azizzadenesheli, Brandon Yang, Weitang Liu, Zachary C. Lipton, Animashree Anandkumar
We deploy this model and propose generative adversarial tree search (GATS) a deep RL algorithm that learns the environment model and implements Monte Carlo tree search (MCTS) on the learned model for planning.
1 code implementation • ICML 2018 • Tommaso Furlanello, Zachary C. Lipton, Michael Tschannen, Laurent Itti, Anima Anandkumar
Knowledge distillation (KD) consists of transferring knowledge from one machine learning model (the teacher}) to another (the student).
no code implementations • 12 Mar 2018 • Subarna Tripathi, Zachary C. Lipton, Truong Q. Nguyen
In this paper, we propose to denoise corrupted images by finding the nearest point on the GAN manifold, recovering latent vectors by minimizing distances in image space.
no code implementations • ICLR 2018 • Guneet S. Dhillon, Kamyar Azizzadenesheli, Zachary C. Lipton, Jeremy Bernstein, Jean Kossaifi, Aran Khanna, Anima Anandkumar
Neural networks are known to be vulnerable to adversarial examples.
1 code implementation • ICLR 2019 • Peiyun Hu, Zachary C. Lipton, Anima Anandkumar, Deva Ramanan
While many active learning papers assume that the learner can simply ask for a label and receive it, real annotation often presents a mismatch between the form of a label (say, one among many classes), and the form of an annotation (typically yes/no binary feedback).
1 code implementation • ICML 2018 • Zachary C. Lipton, Yu-Xiang Wang, Alex Smola
Faced with distribution shift between training and test set, we wish to detect and quantify the shift, and to correct our classifiers without test set labels.
no code implementations • ICLR 2018 • Zachary C. Lipton, Kamyar Azizzadenesheli, Abhishek Kumar, Lihong Li, Jianfeng Gao, Li Deng
Many practical reinforcement learning problems contain catastrophic states that the optimal policy visits infrequently or never.
1 code implementation • ICLR 2018 • Ashish Khetan, Zachary C. Lipton, Anima Anandkumar
We propose a new algorithm for jointly modeling labels and worker quality from noisy crowd-sourced data.
no code implementations • 20 Nov 2017 • Zachary C. Lipton
For the field of interpretable machine learning to advance, we must ask the following questions: What precisely won't various stakeholders accept?
1 code implementation • NeurIPS 2018 • Zachary C. Lipton, Alexandra Chouldechova, Julian McAuley
Following related work in law and policy, two notions of disparity have come to shape the study of fairness in algorithmic decision-making.
no code implementations • 13 Nov 2017 • John Alberg, Zachary C. Lipton
Academic research has identified some factors, i. e. computed features of the reported data, that are known through retrospective analysis to outperform the market average.
no code implementations • IJCNLP 2017 • Jianmo Ni, Zachary C. Lipton, Sharad Vikram, Julian McAuley
Natural language approaches that model information like product reviews have proved to be incredibly useful in improving the performance of such methods, as reviews provide valuable auxiliary information that can be used to better estimate latent user preferences and item properties.
no code implementations • 26 Jul 2017 • Jean Kossaifi, Zachary C. Lipton, Arinbjorn Kolbeinsson, Aran Khanna, Tommaso Furlanello, Anima Anandkumar
First, we introduce Tensor Contraction Layers (TCLs) that reduce the dimensionality of their input while preserving their multilinear structure using tensor contraction.
2 code implementations • WS 2017 • Yanyao Shen, Hyokun Yun, Zachary C. Lipton, Yakov Kronrod, Animashree Anandkumar
In this work, we demonstrate that the amount of labeled training data can be drastically reduced when deep learning is combined with active learning.
no code implementations • 1 Jun 2017 • Jean Kossaifi, Aran Khanna, Zachary C. Lipton, Tommaso Furlanello, Anima Anandkumar
Specifically, we propose the Tensor Contraction Layer (TCL), the first attempt to incorporate tensor contractions as end-to-end trainable neural network layers.
1 code implementation • ICLR 2018 • Chris Donahue, Zachary C. Lipton, Akshay Balsubramani, Julian McAuley
Corresponding samples from the real dataset consist of two distinct photographs of the same subject.
1 code implementation • ICML 2017 • Chris Donahue, Zachary C. Lipton, Julian McAuley
For the step placement task, we combine recurrent and convolutional neural networks to ingest spectrograms of low-level audio features to predict steps, conditioned on chart difficulty.
no code implementations • 17 Feb 2017 • Nathan Ng, Rodney A Gabriel, Julian McAuley, Charles Elkan, Zachary C. Lipton
Scheduling surgeries is a challenging task due to the fundamental uncertainty of the clinical environment, as well as the risks and costs associated with under- and over-booking.
1 code implementation • 15 Feb 2017 • Zachary C. Lipton, Subarna Tripathi
Generative adversarial networks (GANs) transform latent vectors into visually plausible images.
10 code implementations • 17 Dec 2016 • Xiujun Li, Zachary C. Lipton, Bhuwan Dhingra, Lihong Li, Jianfeng Gao, Yun-Nung Chen
Then, one can train reinforcement learning agents in an online fashion as they interact with the simulator.
no code implementations • 3 Nov 2016 • Zachary C. Lipton, Kamyar Azizzadenesheli, Abhishek Kumar, Lihong Li, Jianfeng Gao, Li Deng
We introduce intrinsic fear (IF), a learned reward shaping that guards DRL agents against periodic catastrophes.
no code implementations • 17 Aug 2016 • Zachary C. Lipton, Xiujun Li, Jianfeng Gao, Lihong Li, Faisal Ahmed, Li Deng
We present a new algorithm that significantly improves the efficiency of exploration for deep Q-learning agents in dialogue systems.
no code implementations • 15 Jul 2016 • Subarna Tripathi, Zachary C. Lipton, Serge Belongie, Truong Nguyen
Then we train a recurrent neural network that takes as input sequences of pseudo-labeled frames and optimizes an objective that encourages both accuracy on the target frame and consistency across consecutive frames.
no code implementations • 13 Jun 2016 • Zachary C. Lipton, David C. Kale, Randall Wetzel
For linear models, we show an alternative strategy to capture this signal.
2 code implementations • 10 Jun 2016 • Zachary C. Lipton
First, we examine the motivations underlying interest in interpretability, finding them to be diverse and occasionally discordant.
no code implementations • 23 Feb 2016 • Zachary C. Lipton
As neural networks are typically over-complete, it's easy to show the existence of vast continuous regions through weight space with equal loss.
no code implementations • 11 Nov 2015 • Zachary C. Lipton, David C. Kale, Charles Elkan, Randall Wetzel
We present the first study to empirically evaluate the ability of LSTMs to recognize patterns in multivariate time series of clinical measurements.
1 code implementation • 11 Nov 2015 • Zachary C. Lipton, Sharad Vikram, Julian McAuley
A recommender system's basic task is to estimate how users will respond to unseen items.
no code implementations • 26 Oct 2015 • Zachary C. Lipton, David C. Kale, Randall C. Wetzel
We present a novel application of LSTM recurrent neural networks to multilabel classification of diagnoses given variable-length time series of clinical measurements.
2 code implementations • 29 May 2015 • Zachary C. Lipton, John Berkowitz, Charles Elkan
Recurrent neural networks (RNNs) are connectionist models that capture the dynamics of sequences via cycles in the network of nodes.
no code implementations • 24 May 2015 • Zachary C. Lipton, Charles Elkan
This paper provides closed-form updates for the popular squared norm $\ell^2_2$ and elastic net regularizers.
no code implementations • 24 Dec 2014 • Zhanglong Ji, Zachary C. Lipton, Charles Elkan
The objective of machine learning is to extract useful information from data, while privacy is preserved by concealing information.