Search Results for author: Michael Oberst

Found 15 papers, 11 papers with code

The Limited Impact of Medical Adaptation of Large Language and Vision-Language Models

1 code implementation13 Nov 2024 Daniel P. Jeong, Pranav Mani, Saurabh Garg, Zachary C. Lipton, Michael Oberst

Several recent works seek to develop foundation models specifically for medical applications, adapting general-purpose large language models (LLMs) and vision-language models (VLMs) via continued pretraining on publicly available biomedical corpora.

Question Answering

Medical Adaptation of Large Language and Vision-Language Models: Are We Making Progress?

1 code implementation6 Nov 2024 Daniel P. Jeong, Saurabh Garg, Zachary C. Lipton, Michael Oberst

For instance, across the tasks and model pairs we consider in the 3-shot setting, medical LLMs only outperform their base models in 12. 1% of cases, reach a (statistical) tie in 49. 8% of cases, and are significantly worse than their base models in the remaining 38. 2% of cases.

Question Answering

Auditing Fairness under Unobserved Confounding

1 code implementation18 Mar 2024 Yewon Byun, Dylan Sam, Michael Oberst, Zachary C. Lipton, Bryan Wilder

The presence of inequity is a fundamental problem in the outcomes of decision-making systems, especially when human lives are at stake.

Decision Making Fairness

Benchmarking Observational Studies with Experimental Data under Right-Censoring

no code implementations23 Feb 2024 Ilker Demirel, Edward De Brouwer, Zeshan Hussain, Michael Oberst, Anthony Philippakis, David Sontag

Drawing causal inferences from observational studies (OS) requires unverifiable validity assumptions; however, one can falsify those assumptions by benchmarking the OS with experimental data from a randomized controlled trial (RCT).

Benchmarking

Falsification of Internal and External Validity in Observational Studies via Conditional Moment Restrictions

no code implementations30 Jan 2023 Zeshan Hussain, Ming-Chieh Shih, Michael Oberst, Ilker Demirel, David Sontag

Our approach is interpretable, allowing a practitioner to visualize which subgroups in the population lead to falsification of an observational study.

counterfactual

Falsification before Extrapolation in Causal Effect Estimation

1 code implementation27 Sep 2022 Zeshan Hussain, Michael Oberst, Ming-Chieh Shih, David Sontag

Under the assumption that at least one observational estimator is asymptotically normal and consistent for both the validation and extrapolated effects, we provide guarantees on the coverage probability of the intervals output by our algorithm.

Selection bias

Evaluating Robustness to Dataset Shift via Parametric Robustness Sets

1 code implementation31 May 2022 Nikolaj Thams, Michael Oberst, David Sontag

We give a method for proactively identifying small, plausible shifts in distribution which lead to large differences in model performance.

Regularizing towards Causal Invariance: Linear Models with Proxies

1 code implementation3 Mar 2021 Michael Oberst, Nikolaj Thams, Jonas Peters, David Sontag

In the case of two proxy variables, we propose a modified estimator that is prediction optimal under interventions up to a known strength.

Treatment Policy Learning in Multiobjective Settings with Fully Observed Outcomes

1 code implementation1 Jun 2020 Soorajnath Boominathan, Michael Oberst, Helen Zhou, Sanjat Kanjilal, David Sontag

In several medical decision-making problems, such as antibiotic prescription, laboratory testing can provide precise indications for how a patient will respond to different treatment options.

Decision Making

Counterfactual Off-Policy Evaluation with Gumbel-Max Structural Causal Models

1 code implementation14 May 2019 Michael Oberst, David Sontag

We introduce an off-policy evaluation procedure for highlighting episodes where applying a reinforcement learned (RL) policy is likely to have produced a substantially different outcome than the observed policy.

counterfactual Management +1

Cannot find the paper you are looking for? You can Submit a new open access paper.