1 code implementation • EMNLP 2021 • Prafulla Kumar Choubey, Anna Currey, Prashant Mathur, Georgiana Dinu
Targeted evaluations have found that machine translation systems often output incorrect gender in translations, even when the gender is clear from context.
no code implementations • ACL 2022 • Bohan Zhang, Prafulla Kumar Choubey, Ruihong Huang
Document-level text simplification often deletes some sentences besides performing lexical, grammatical or structural simplification to reduce text complexity.
no code implementations • Findings (EMNLP) 2021 • Prafulla Kumar Choubey, Ruihong Huang
We present an actor-critic framework to induce subtopical structures in a news article for news discourse profiling.
Ranked #4 on
Text Classification
on NewsDiscourse
no code implementations • 24 Feb 2025 • Prafulla Kumar Choubey, Xiangyu Peng, Shilpa Bhagavath, Caiming Xiong, Shiva Kumar Pentyala, Chien-Sheng Wu
Automated service agents require well-structured workflows to provide consistent and accurate responses to customer queries.
no code implementations • 16 Dec 2024 • Xiangyu Peng, Prafulla Kumar Choubey, Caiming Xiong, Chien-Sheng Wu
Existing evaluation frameworks for retrieval-augmented generation (RAG) systems focus on answerable queries, but they overlook the importance of appropriately rejecting unanswerable requests.
no code implementations • 9 Dec 2024 • Nan Zhang, Prafulla Kumar Choubey, Alexander Fabbri, Gabriel Bernadett-Shapiro, Rui Zhang, Prasenjit Mitra, Caiming Xiong, Chien-Sheng Wu
On the similarity side, we follow existing work and explore some variances to construct a similarity tree based on recursive summarization.
no code implementations • 22 Oct 2024 • Prafulla Kumar Choubey, Xin Su, Man Luo, Xiangyu Peng, Caiming Xiong, Tiep Le, Shachar Rosenman, Vasudev Lal, Phil Mui, Ricky Ho, Phillip Howard, Chien-Sheng Wu
Additionally, there is a gap in evaluation datasets and methodologies for ontology-free KG construction.
1 code implementation • 20 Oct 2024 • Kaige Xie, Philippe Laban, Prafulla Kumar Choubey, Caiming Xiong, Chien-Sheng Wu
Using this categorization, we introduce a fine-grained evaluation protocol that provides insights into the retrieval and generation characteristics of RAG systems, including three commercial generative answer engines: You. com, Perplexity AI, and Bing Chat.
no code implementations • 15 Nov 2023 • Prafulla Kumar Choubey, Alexander R. Fabbri, Caiming Xiong, Chien-Sheng Wu
Ideal summarization models should generalize to novel summary-worthy content without remembering reference training summaries by rote.
1 code implementation • 17 Sep 2023 • Kung-Hsiang Huang, Philippe Laban, Alexander R. Fabbri, Prafulla Kumar Choubey, Shafiq Joty, Caiming Xiong, Chien-Sheng Wu
In this paper, we propose a new task of summarizing diverse information encountered in multiple news articles encompassing the same event.
1 code implementation • 7 Sep 2023 • Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause, Senthil Purushwalkam, Tong Niu, Wojciech Kryściński, Lidiya Murakhovs'ka, Prafulla Kumar Choubey, Alex Fabbri, Ye Liu, Rui Meng, Lifu Tu, Meghana Bhat, Chien-Sheng Wu, Silvio Savarese, Yingbo Zhou, Shafiq Joty, Caiming Xiong
Most open-source LLMs, on the other hand, are limited in their ability to support longer sequence lengths, which is a key requirement for many tasks that require inference over an input context.
1 code implementation • 11 Nov 2022 • Alexander R. Fabbri, Prafulla Kumar Choubey, Jesse Vig, Chien-Sheng Wu, Caiming Xiong
We propose to use sentence-compression data to train the post-editing model to take a summary with extrinsic entity errors marked with special tokens and output a compressed, well-formed summary with those errors removed.
no code implementations • 23 Oct 2022 • Prafulla Kumar Choubey, Yu Bai, Chien-Sheng Wu, Wenhao Liu, Nazneen Rajani
Pre-trained language models (PLMs) have been shown effective for zero-shot (0shot) text classification.
no code implementations • 23 Oct 2022 • Xiangyu Peng, Chen Xing, Prafulla Kumar Choubey, Chien-Sheng Wu, Caiming Xiong
Through this way, SESoM inherits the superior generalization of model ensemble approaches and simultaneously captures the sample-specific competence of each source prompt.
1 code implementation • 21 Oct 2022 • Prafulla Kumar Choubey, Ruihong Huang
We propose to leverage news discourse profiling to model document-level temporal structures for building temporal dependency graphs.
no code implementations • 14 Oct 2021 • Prafulla Kumar Choubey, Alexander R. Fabbri, Jesse Vig, Chien-Sheng Wu, Wenhao Liu, Nazneen Fatema Rajani
Then, we fine-tune a base summarization model, which is trained on all training samples, on the clean (noisy) subset to obtain an \textit{expert} (\textit{anti-expert}) model.
1 code implementation • ICLR 2022 • Benjamin Newman, Prafulla Kumar Choubey, Nazneen Rajani
They take LLM embeddings as input and output continuous prompts that are used to query the LLM.
no code implementations • 15 Apr 2021 • Prafulla Kumar Choubey, Anna Currey, Prashant Mathur, Georgiana Dinu
Targeted evaluations have found that machine translation systems often output incorrect gender, even when the gender is clear from context.
1 code implementation • EACL 2021 • Prafulla Kumar Choubey, Ruihong Huang
We propose to leverage lexical paraphrases and high precision rules informed by news discourse structure to automatically collect coreferential and non-coreferential event pairs from unlabeled English news articles.
no code implementations • ACL 2020 • Prafulla Kumar Choubey, Aaron Lee, Ruihong Huang, Lu Wang
Understanding discourse structures of news articles is vital to effectively contextualize the occurrence of a news event.
Ranked #5 on
Text Classification
on NewsDiscourse
no code implementations • LREC 2020 • Prafulla Kumar Choubey, Ruihong Huang
Most supervised word sense disambiguation (WSD) systems build word-specific classifiers by leveraging labeled data.
1 code implementation • IJCNLP 2019 • Lisa Fan, Marshall White, Eva Sharma, Ruisi Su, Prafulla Kumar Choubey, Ruihong Huang, Lu Wang
The increasing prevalence of political bias in news media calls for greater public awareness of it, as well as robust methods for its detection.
no code implementations • NAACL 2019 • Lei Gao, Prafulla Kumar Choubey, Ruihong Huang
We aim to comprehensively identify all the event causal relations in a document, both within a sentence and across sentences, which is important for reconstructing pivotal event structures.
no code implementations • NAACL 2019 • Sanuj Sharma, Prafulla Kumar Choubey, Ruihong Huang
Specifically, we use the current user utterance and the most recent system utterance to determine the relevance of a system utterance.
no code implementations • ACL 2018 • Prafulla Kumar Choubey, Ruihong Huang
This paper proposes a novel approach for event coreference resolution that models correlations between event coreference chains and document topical structures through an Integer Linear Programming formulation.
no code implementations • NAACL 2018 • Prafulla Kumar Choubey, Kaushik Raju, Ruihong Huang
Identifying the most dominant and central event of a document, which governs and connects other foreground and background events in the document, is useful for many applications, such as text summarization, storyline generation and text segmentation.
1 code implementation • 6 Nov 2017 • Prafulla Kumar Choubey, Ruihong Huang
Our simple system designed using minimal features achieved the micro-average F1 scores of 57. 72, 44. 27 and 42. 47 for event span detection, type identification and realis status classification tasks respectively.
no code implementations • EMNLP 2017 • Prafulla Kumar Choubey, Ruihong Huang
We present a sequential model for temporal relation classification between intra-sentence events.
no code implementations • EMNLP 2017 • Prafulla Kumar Choubey, Ruihong Huang
We introduce a novel iterative approach for event coreference resolution that gradually builds event clusters by exploiting inter-dependencies among event mentions within the same chain as well as across event chains.