Search Results for author: Micha Elsner

Found 28 papers, 6 papers with code

A Word-and-Paradigm Workflow for Fieldwork Annotation

no code implementations ComputEL (ACL) 2022 Maria Copot, Sara Court, Noah Diewald, Stephanie Antetomaso, Micha Elsner

There are many challenges in morphological fieldwork annotation, it heavily relies on segmentation and feature labeling (which have both practical and theoretical drawbacks), it’s time-intensive, and the annotator needs to be linguistically trained and may still annotate things inconsistently.

Active Learning Segmentation

OSU at SigMorphon 2022: Analogical Inflection With Rule Features

1 code implementation NAACL (SIGMORPHON) 2022 Micha Elsner, Sara Court

OSU’s inflection system is a transformer whose input is augmented with an analogical exemplar showing how to inflect a different word into the target cell.

Exploring How Generative Adversarial Networks Learn Phonological Representations

no code implementations21 May 2023 Jingyi Chen, Micha Elsner

This paper explores how Generative Adversarial Networks (GANs) learn representations of phonological phenomena.

Analogy in Contact: Modeling Maltese Plural Inflection

no code implementations20 May 2023 Sara Court, Andrea D. Sims, Micha Elsner

Maltese is often described as having a hybrid morphological system resulting from extensive contact between Semitic and Romance language varieties.

Privacy Policy Question Answering Assistant: A Query-Guided Extractive Summarization Approach

no code implementations29 Sep 2021 Moniba Keymanesh, Micha Elsner, Srinivasan Parthasarathy

We address these problems by paraphrasing to bring the style and language of the user's question closer to the language of privacy policies.

Extractive Summarization Question Answering

Fairness-aware Summarization for Justified Decision-Making

no code implementations13 Jul 2021 Moniba Keymanesh, Tanya Berger-Wolf, Micha Elsner, Srinivasan Parthasarathy

In other words, decision-relevant features should provide sufficient information for the predicted outcome and should be independent of the membership of individuals in protected groups such as race and gender.

Data Poisoning Decision Making +1

Acquiring language from speech by learning to remember and predict

1 code implementation CONLL 2020 Cory Shain, Micha Elsner

Classical accounts of child language learning invoke memory limits as a pressure to discover sparse, language-like representations of speech, while more recent proposals stress the importance of prediction for language learning.

The Paradigm Discovery Problem

1 code implementation ACL 2020 Alexander Erdmann, Micha Elsner, Shijie Wu, Ryan Cotterell, Nizar Habash

Our benchmark system first makes use of word embeddings and string similarity to cluster forms by cell and by paradigm.

Clustering Word Embeddings

Measuring the perceptual availability of phonological features during language acquisition using unsupervised binary stochastic autoencoders

no code implementations NAACL 2019 Cory Shain, Micha Elsner

In this paper, we deploy binary stochastic neural autoencoder networks as models of infant language learning in two typologically unrelated languages (Xitsonga and English).

Language Acquisition

Lexical Networks in !Xung

no code implementations WS 2018 Syed-Amad Hussain, Micha Elsner, Am Miller, a

We investigate the lexical network properties of the large phoneme inventory Southern African language Mangetti Dune ! Xung as it compares to English and other commonly-studied languages.

Retrieval

Speech segmentation with a neural encoder model of working memory

no code implementations EMNLP 2017 Micha Elsner, Cory Shain

We present the first unsupervised LSTM speech segmenter as a cognitive model of the acquisition of words from unsegmented input.

Challenges and Solutions for Latin Named Entity Recognition

no code implementations WS 2016 Alex Erdmann, er, Christopher Brown, Brian Joseph, Mark Janse, Petra Ajaka, Micha Elsner, Marie-Catherine de Marneffe

Although spanning thousands of years and genres as diverse as liturgy, historiography, lyric and other forms of prose and poetry, the body of Latin texts is still relatively sparse compared to English.

Active Learning Domain Adaptation +5

You Talking to Me? A Corpus and Algorithm for Conversation Disentanglement

no code implementations1 Jun 2008 Micha Elsner, Eugene Charniak

We present a corpus of Internet Relay Chat (IRC) dialogue in which the various conversations have been manually disentangled, and evaluate annotator reliability.

Conversation Disentanglement Disentanglement

Cannot find the paper you are looking for? You can Submit a new open access paper.