no code implementations • 9 Dec 2023 • Vivek Miglani, Aobo Yang, Aram H. Markosyan, Diego Garcia-Olano, Narine Kokhlikyan
Captum is a comprehensive library for model explainability in PyTorch, offering a range of methods from the interpretability literature to enhance users' understanding of PyTorch models.
no code implementations • NeurIPS 2023 • Fulton Wang, Julius Adebayo, Sarah Tan, Diego Garcia-Olano, Narine Kokhlikyan
We present a method for identifying groups of test examples -- slices -- on which a model under-performs, a task now known as slice discovery.
1 code implementation • 3 Dec 2022 • Diego Garcia-Olano, Yasumasa Onoe, Joydeep Ghosh, Byron C. Wallace
However, while fine-tuning sparse, interpretable representations improves accuracy on downstream tasks, it destroys the semantics of the dimensions which were enforced in pre-training.
no code implementations • 13 Dec 2021 • Diego Garcia-Olano, Yasumasa Onoe, Joydeep Ghosh
In this work, we empirically study how and whether such methods, applied in a bi-modal setting, can improve an existing VQA system's performance on the KBVQA task.
2 code implementations • Findings (ACL) 2021 • Diego Garcia-Olano, Yasumasa Onoe, Ioana Baldini, Joydeep Ghosh, Byron C. Wallace, Kush R. Varshney
Pre-trained language models induce dense entity representations that offer strong performance on entity-centric NLP tasks, but such representations are not immediately interpretable.
no code implementations • CONLL 2019 • Daniel Gillick, Sayali Kulkarni, Larry Lansing, Alessandro Presta, Jason Baldridge, Eugene Ie, Diego Garcia-Olano
We show that it is feasible to perform entity linking by training a dual encoder (two-tower) model that encodes mentions and entities in the same dense vector space, where candidate entities are retrieved by approximate nearest neighbor search.
1 code implementation • 18 Apr 2019 • Alan H. Gee, Diego Garcia-Olano, Joydeep Ghosh, David Paydarfar
We improve upon existing models by optimizing for increased prototype diversity and robustness, visualize how these prototypes in the latent space are used by the model to distinguish classes, and show that prototypes are capable of learning features on two dimensional time-series data to produce explainable insights during classification tasks.