2 code implementations • 25 Jan 2021 • Satya Narayan Shukla, Benjamin M. Marlin
Irregular sampling occurs in many time series modeling applications where it presents a significant challenge to standard deep learning models.
1 code implementation • 31 Aug 2023 • Lucas Bandarkar, Davis Liang, Benjamin Muller, Mikel Artetxe, Satya Narayan Shukla, Donald Husa, Naman Goyal, Abhinandan Krishnan, Luke Zettlemoyer, Madian Khabsa
We use this dataset to evaluate the capabilities of multilingual masked language models (MLMs) and large language models (LLMs).
1 code implementation • 3 Dec 2018 • Satya Narayan Shukla, Benjamin M. Marlin
In this paper, we present a new deep learning architecture for addressing this problem based on the use of a semi-parametric interpolation network followed by the application of a prediction network.
1 code implementation • ICLR 2019 • Satya Narayan Shukla, Benjamin M. Marlin
The interpolation network allows for information to be shared across multiple dimensions of a multivariate time series during the interpolation stage, while any standard deep learning model can be used for the prediction network.
1 code implementation • 30 Sep 2019 • Satya Narayan Shukla, Anit Kumar Sahu, Devin Willmott, J. Zico Kolter
We focus on the problem of black-box adversarial attacks, where the aim is to generate adversarial examples using information limited to loss function evaluations of input-output pairs.
1 code implementation • ICLR 2023 • Satya Narayan Shukla, Benjamin M. Marlin
Irregularly sampled time series commonly occur in several domains where they present a significant challenge to standard deep learning models.
1 code implementation • 13 Jul 2020 • Satya Narayan Shukla, Anit Kumar Sahu, Devin Willmott, J. Zico Kolter
We focus on the problem of black-box adversarial attacks, where the aim is to generate adversarial examples for deep learning models solely based on information limited to output label~(hard label) to a queried data input.
1 code implementation • 8 Oct 2020 • Anit Kumar Sahu, Satya Narayan Shukla, J. Zico Kolter
We study the problem of generating adversarial examples in a black-box setting, where we only have access to a zeroth order oracle, providing us with loss function evaluations.
no code implementations • 7 Feb 2020 • Meet P. Vadera, Satya Narayan Shukla, Brian Jalaian, Benjamin M. Marlin
In this paper, we consider the problem of assessing the adversarial robustness of deep neural network models under both Markov chain Monte Carlo (MCMC) and Bayesian Dark Knowledge (BDK) inference approximations.
no code implementations • 24 Mar 2020 • Satya Narayan Shukla, Benjamin M. Marlin
Intensive Care Unit Electronic Health Records (ICU EHRs) store multimodal data about patients including clinical notes, sparse and irregularly sampled physiological time series, lab results, and more.
no code implementations • 30 Nov 2020 • Satya Narayan Shukla, Benjamin M. Marlin
Such data represent fundamental challenges to many classical models from machine learning and statistics due to the presence of non-uniform intervals between observations.
no code implementations • 20 Sep 2023 • Mohamed Afham, Satya Narayan Shukla, Omid Poursaeed, Pengchuan Zhang, Ashish Shah, SerNam Lim
While most modern video understanding models operate on short-range clips, real-world videos are often several minutes long with semantically consistent segments of variable length.
no code implementations • 26 Dec 2023 • Ping-Yeh Chiang, Yipin Zhou, Omid Poursaeed, Satya Narayan Shukla, Ashish Shah, Tom Goldstein, Ser-Nam Lim
Recently, Pyramid Adversarial training (Herrmann et al., 2022) has been shown to be very effective for improving clean accuracy and distribution-shift robustness of vision transformers.
no code implementations • 11 Apr 2024 • Kanchana Ranasinghe, Satya Narayan Shukla, Omid Poursaeed, Michael S. Ryoo, Tsung-Yu Lin
Integration of Large Language Models (LLMs) into visual domain tasks, resulting in visual-LLMs (V-LLMs), has enabled exceptional performance in vision-language tasks, particularly for visual question answering (VQA).