no code implementations • 25 Oct 2023 • Yoshinari Fujinuma, Siddharth Varia, Nishant Sankaran, Srikar Appalaraju, Bonan Min, Yogarshi Vyas
Document image classification is different from plain-text document classification and consists of classifying a document by understanding the content and structure of documents such as forms, emails, and other such documents.
1 code implementation • 26 May 2023 • Tyler A. Chang, Kishaloy Halder, Neha Anna John, Yogarshi Vyas, Yassine Benajiba, Miguel Ballesteros, Dan Roth
In this paper, we propose three dimensions of linguistic dataset drift: vocabulary, structural, and semantic drift.
no code implementations • 22 May 2023 • Karthikeyan K, Yogarshi Vyas, Jie Ma, Giovanni Paolini, Neha Anna John, Shuai Wang, Yassine Benajiba, Vittorio Castelli, Dan Roth, Miguel Ballesteros
We experiment with 6 diverse datasets and show that PLM consistently performs better than most other approaches (0. 5 - 2. 5 F1), including in novel settings for taxonomy expansion not considered in prior work.
no code implementations • 18 May 2023 • Sharon Levy, Neha Anna John, Ling Liu, Yogarshi Vyas, Jie Ma, Yoshinari Fujinuma, Miguel Ballesteros, Vittorio Castelli, Dan Roth
As a result, it is critical to examine biases within each language and attribute.
no code implementations • 21 Mar 2023 • Ming Shen, Jie Ma, Shuai Wang, Yogarshi Vyas, Kalpit Dixit, Miguel Ballesteros, Yassine Benajiba
Opinion summarization provides an important solution for summarizing opinions expressed among a large number of reviews.
no code implementations • 23 Feb 2023 • Katerina Margatina, Shuai Wang, Yogarshi Vyas, Neha Anna John, Yassine Benajiba, Miguel Ballesteros
Temporal concept drift refers to the problem of data changing over time.
no code implementations • 11 Oct 2022 • Muhammad Khalifa, Yogarshi Vyas, Shuai Wang, Graham Horwood, Sunil Mallya, Miguel Ballesteros
The standard classification setting where categories are fixed during both training and testing falls short in dynamic environments where new document categories could potentially emerge.
no code implementations • ACL 2022 • Hyunji Hayley Park, Yogarshi Vyas, Kashif Shah
Several methods have been proposed for classifying long textual documents using Transformers.
1 code implementation • 28 Jun 2021 • Paula Czarnowska, Yogarshi Vyas, Kashif Shah
Measuring bias is key for better understanding and addressing unfairness in NLP/ML models.
no code implementations • NAACL 2021 • Yogarshi Vyas, Miguel Ballesteros
In entity linking, mentions of named entities in raw text are disambiguated against a knowledge base (KB).
no code implementations • EMNLP 2020 • Miguel Ballesteros, Rishita Anubhai, Shuai Wang, Nima Pourdamghani, Yogarshi Vyas, Jie Ma, Parminder Bhatia, Kathleen McKeown, Yaser Al-Onaizan
In this paper, we propose a neural architecture and a set of training methods for ordering events by predicting temporal relations.
no code implementations • IJCNLP 2019 • Yogarshi Vyas, Marine Carpuat
Our classifier relies on a novel attention-based distillation approach to account for translation ambiguity when transferring knowledge from English to cross-lingual settings.
1 code implementation • NAACL 2018 • Shyam Upadhyay, Yogarshi Vyas, Marine Carpuat, Dan Roth
We propose BISPARSE-DEP, a family of unsupervised approaches for cross-lingual hypernymy detection, which learns sparse, bilingual word embeddings based on dependency contexts.
1 code implementation • NAACL 2018 • Yogarshi Vyas, Xing Niu, Marine Carpuat
Recognizing that even correct translations are not always semantically equivalent, we automatically detect meaning divergences in parallel sentence pairs with a deep neural model of bilingual semantic similarity which can be trained for any parallel corpus without any manual annotation.
no code implementations • SEMEVAL 2017 • Yogarshi Vyas, Marine Carpuat
We introduce WHiC, a challenging testbed for detecting hypernymy, an asymmetric relation between words.
no code implementations • WS 2017 • Marine Carpuat, Yogarshi Vyas, Xing Niu
Parallel corpora are often not as parallel as one might assume: non-literal translations and noisy translations abound, even in curated corpora routinely used for training and evaluation.
2 code implementations • CVPR 2017 • Mohit Iyyer, Varun Manjunatha, Anupam Guha, Yogarshi Vyas, Jordan Boyd-Graber, Hal Daumé III, Larry Davis
While computers can now describe what is explicitly depicted in natural images, in this paper we examine whether they can understand the closure-driven narratives conveyed by stylized artwork and dialogue in comic book panels.
no code implementations • 26 Oct 2015 • Sudha Rao, Yogarshi Vyas, Hal Daume III, Philip Resnik
We develop a novel technique to parse English sentences into Abstract Meaning Representation (AMR) using SEARN, a Learning to Search approach, by modeling the concept and the relation learning in a unified framework.