Search Results for author: Marzena Karpinska

Found 11 papers, 6 papers with code

Revisiting Statistical Laws of Semantic Shift in Romance Cognates

no code implementations COLING 2022 Yoshifumi Kawasaki, Maëlys Salingre, Marzena Karpinska, Hiroya Takamura, Ryo Nagata

This article revisits statistical relationships across Romance cognates between lexical semantic shift and six intra-linguistic variables, such as frequency and polysemy.

Word Embeddings

FABLES: Evaluating faithfulness and content selection in book-length summarization

3 code implementations1 Apr 2024 Yekyung Kim, Yapei Chang, Marzena Karpinska, Aparna Garimella, Varun Manjunatha, Kyle Lo, Tanya Goyal, Mohit Iyyer

While LLM-based auto-raters have proven reliable for factuality and coherence in other settings, we implement several LLM raters of faithfulness and find that none correlates strongly with human annotations, especially with regard to detecting unfaithful claims.

Long-Context Understanding

Large language models effectively leverage document-level context for literary translation, but critical errors persist

1 code implementation6 Apr 2023 Marzena Karpinska, Mohit Iyyer

Large language models (LLMs) are competitive with the state of the art on a wide range of sentence-level translation datasets.

Sentence Translation

Paraphrasing evades detectors of AI-generated text, but retrieval is an effective defense

1 code implementation NeurIPS 2023 Kalpesh Krishna, Yixiao Song, Marzena Karpinska, John Wieting, Mohit Iyyer

To increase the robustness of AI-generated text detection to paraphrase attacks, we introduce a simple defense that relies on retrieving semantically-similar generations and must be maintained by a language model API provider.

Language Modelling Outlier Detection +3

Exploring Document-Level Literary Machine Translation with Parallel Paragraphs from World Literature

1 code implementation25 Oct 2022 Katherine Thai, Marzena Karpinska, Kalpesh Krishna, Bill Ray, Moira Inghilleri, John Wieting, Mohit Iyyer

Using Par3, we discover that expert literary translators prefer reference human translations over machine-translated paragraphs at a rate of 84%, while state-of-the-art automatic MT metrics do not correlate with those preferences.

Machine Translation Translation

DEMETR: Diagnosing Evaluation Metrics for Translation

1 code implementation25 Oct 2022 Marzena Karpinska, Nishant Raj, Katherine Thai, Yixiao Song, Ankita Gupta, Mohit Iyyer

While machine translation evaluation metrics based on string overlap (e. g., BLEU) have their limitations, their computations are transparent: the BLEU score assigned to a particular candidate translation can be traced back to the presence or absence of certain words.

Machine Translation Translation

The Perils of Using Mechanical Turk to Evaluate Open-Ended Text Generation

no code implementations EMNLP 2021 Marzena Karpinska, Nader Akoury, Mohit Iyyer

Recent text generation research has increasingly focused on open-ended domains such as story and poetry generation.

Text Generation

NarrativeTime: Dense Temporal Annotation on a Timeline

no code implementations29 Aug 2019 Anna Rogers, Marzena Karpinska, Ankita Gupta, Vladislav Lialin, Gregory Smelkov, Anna Rumshisky

For the past decade, temporal annotation has been sparse: only a small portion of event pairs in a text was annotated.

Chunking

Cannot find the paper you are looking for? You can Submit a new open access paper.