no code implementations • 23 Apr 2024 • Derek Powell, Walter Gerych, Thomas Hartvigsen
We then use TAXI to evaluate popular editors' consistency, measuring how often editing a subject's category appropriately edits its properties.
1 code implementation • 29 Feb 2024 • ShangHua Gao, Teddy Koker, Owen Queen, Thomas Hartvigsen, Theodoros Tsiligkaridis, Marinka Zitnik
However, current foundation models apply to sequence data but not to time series, which present unique challenges due to the inherent diverse and multidomain time series datasets, diverging task specifications across forecasting, classification and other types of tasks, and the apparent need for task-specialized models.
1 code implementation • 24 Feb 2024 • Bryan R Christ, Jonathan Kropko, Thomas Hartvigsen
MATHWELL's performance despite being trained by finetuning only highlights the quality of our synthetic data for training age-appropriate word problem generators.
1 code implementation • 13 Feb 2024 • Kyle O'Brien, Nathan Ng, Isha Puri, Jorge Mendez, Hamid Palangi, Yoon Kim, Marzyeh Ghassemi, Thomas Hartvigsen
Most techniques for improving OOD robustness are not applicable to settings where the model is effectively a black box, such as when the weights are frozen, retraining is costly, or the model is leveraged via an API.
no code implementations • 6 Feb 2024 • Sujay Nagaraj, Walter Gerych, Sana Tonekaboni, Anna Goldenberg, Berk Ustun, Thomas Hartvigsen
We first demonstrate the importance of modelling the temporal nature of the label noise function and how existing methods will consistently underperform.
no code implementations • 1 Dec 2023 • Stefan Hegselmann, Antonio Parziale, Divya Shanmugam, Shengpu Tang, Mercy Nyamewaa Asiedu, Serina Chang, Thomas Hartvigsen, Harvineet Singh
A collection of the accepted Findings papers that were presented at the 3rd Machine Learning for Health symposium (ML4H 2023), which was held on December 10, 2023, in New Orleans, Louisiana, USA.
no code implementations • 4 Nov 2023 • Hang Yin, Yao Su, Xinyue Liu, Thomas Hartvigsen, Yanhua Li, Xiangnan Kong
We refer to such brain networks as multi-state, and this mixture can help us understand human behavior.
1 code implementation • 25 Jul 2023 • Taylor W. Killian, Haoran Zhang, Thomas Hartvigsen, Ava P. Amini
Prevalent in many real-world settings such as healthcare, irregular time series are challenging to formulate predictions from.
1 code implementation • 7 Apr 2023 • Tianhua Zhang, Hongyin Luo, Yung-Sung Chuang, Wei Fang, Luc Gaitskell, Thomas Hartvigsen, Xixin Wu, Danny Fox, Helen Meng, James Glass
Despite recent concerns about undesirable behaviors generated by large language models (LLMs), including non-factual, biased, and hateful language, we find LLMs are inherent multi-task language checkers based on their latent representations of natural and social knowledge.
no code implementations • 8 Feb 2023 • Thomas Hartvigsen, Jidapa Thadajarassiri, Xiangnan Kong, Elke Rundensteiner
Using this insight, we then propose CAT, a model that classifies multivariate ITS by explicitly seeking highly-relevant portions of an input series' timeline.
1 code implementation • NeurIPS 2023 • Thomas Hartvigsen, Swami Sankaranarayanan, Hamid Palangi, Yoon Kim, Marzyeh Ghassemi
We propose GRACE, a lifelong model editing method, which implements spot-fixes on streaming errors of a deployed model, ensuring minimal impact on unrelated inputs.
1 code implementation • 11 Oct 2022 • Ramesh Doddaiah, Prathyush Parvatharaju, Elke Rundensteiner, Thomas Hartvigsen
Instead, when a classifier is choosing between many classes, an effective explanation must show what sets the chosen class apart from the rest.
1 code implementation • 21 Aug 2022 • Thomas Hartvigsen, Walter Gerych, Jidapa Thadajarassiri, Xiangnan Kong, Elke Rundensteiner
We bridge this gap and study early classification of irregular time series, a new setting for early classifiers that opens doors to more real-world problems.
no code implementations • LREC 2022 • Ruofan Hu, Dongyu Zhang, Dandan Tao, Thomas Hartvigsen, Hao Feng, Elke Rundensteiner
To accelerate the development of machine learning-based models for foodborne outbreak detection, we thus present TWEET-FID (TWEET-Foodborne Illness Detection), the first publicly available annotated dataset for multiple foodborne illness incident detection tasks.
no code implementations • 6 May 2022 • Aparna Balagopalan, Haoran Zhang, Kimia Hamidieh, Thomas Hartvigsen, Frank Rudzicz, Marzyeh Ghassemi
Across two different blackbox model architectures and four popular explainability methods, we find that the approximation quality of explanation models, also known as the fidelity, differs significantly between subgroups.
1 code implementation • ACL 2022 • Thomas Hartvigsen, Saadia Gabriel, Hamid Palangi, Maarten Sap, Dipankar Ray, Ece Kamar
To help mitigate these issues, we create ToxiGen, a new large-scale and machine-generated dataset of 274k toxic and benign statements about 13 minority groups.
no code implementations • 1 Jan 2021 • Walter Gerych, Thomas Hartvigsen, Luke Buquicchio, Kavin Chandrasekaran, Hamid Mansoor, Abdulaziz alajaji
In this work, we propose DeepSPU, the first method to address this sequential bias problem.
no code implementations • ACL 2020 • Cansu Sen, Thomas Hartvigsen, Biao Yin, Xiangnan Kong, Elke Rundensteiner
Motivated by human attention, computational attention mechanisms have been designed to help neural networks adjust their focus on specific parts of the input data.
no code implementations • 25 Sep 2019 • Thomas Hartvigsen, Cansu Sen, Xiangnan Kong, Elke Rundensteiner
As a result, even for high-dimensional hidden states, all dimensions are updated at each timestep regardless of the recurrent memory cell.
1 code implementation • KDD 2019 • Thomas Hartvigsen, Cansu Sen, Xiangnan Kong, Elke Rundensteiner
Early classification of time series is the prediction of the class label of a time series before it is observed in its entirety.