Search Results for author: David Mortensen

Found 18 papers, 7 papers with code

Leveraging Allophony in Self-Supervised Speech Models for Atypical Pronunciation Assessment

1 code implementation10 Feb 2025 Kwanghee Choi, Eunjung Yeo, Kalvin Chang, Shinji Watanabe, David Mortensen

Allophony refers to the variation in the phonetic realization of a phoneme based on its phonetic environment.

Potential Applications of Artificial Intelligence for Cross-language Intelligibility Assessment of Dysarthric Speech

no code implementations27 Jan 2025 Eunjung Yeo, Julie Liss, Visar Berisha, David Mortensen

Purpose: This commentary introduces how artificial intelligence (AI) can be leveraged to advance cross-language intelligibility assessment of dysarthric speech.

Derivational Morphology Reveals Analogical Generalization in Large Language Models

no code implementations12 Nov 2024 Valentin Hofmann, Leonie Weissweiler, David Mortensen, Hinrich Schütze, Janet Pierrehumbert

As expected, rule-based and analogical models explain the predictions of GPT-J equally well for adjectives with regular nominalization patterns.

ELCC: the Emergent Language Corpus Collection

no code implementations4 Jul 2024 Brendon Boldt, David Mortensen

The availability of a substantial collection of well-documented emergent language corpora, then, will enable research which can analyze a wider variety of emergent languages, which more effectively uncovers general principles in emergent communication rather than artifacts of particular environments.

Transfer Learning

XferBench: a Data-Driven Benchmark for Emergent Language

no code implementations3 Jul 2024 Brendon Boldt, David Mortensen

In this paper, we introduce a benchmark for evaluating the overall quality of emergent languages using data-driven methods.

A Review of the Applications of Deep Learning-Based Emergent Communication

no code implementations3 Jul 2024 Brendon Boldt, David Mortensen

Emergent communication, or emergent language, is the field of research which studies how human language-like communication systems emerge de novo in deep multi-agent reinforcement learning environments.

Multi-agent Reinforcement Learning

Can Large Language Models Code Like a Linguist?: A Case Study in Low Resource Sound Law Induction

no code implementations18 Jun 2024 Atharva Naik, Kexun Zhang, Nathaniel Robinson, Aravind Mysore, Clayton Marr, Hong Sng Rebecca Byrnes, Anna Cai, Kalvin Chang, David Mortensen

Historical linguists have long written a kind of incompletely formalized ''program'' that converts reconstructed words in an ancestor language into words in one of its attested descendants that consist of a series of ordered string rewrite functions (called sound laws).

Transformed Protoform Reconstruction

1 code implementation4 Jul 2023 Young Min Kim, Kalvin Chang, Chenxuan Cui, David Mortensen

We update their model with the state-of-the-art seq2seq model: the Transformer.

Decoder

Mathematically Modeling the Lexicon Entropy of Emergent Language

1 code implementation28 Nov 2022 Brendon Boldt, David Mortensen

We formulate a stochastic process, FiLex, as a mathematical model of lexicon entropy in deep learning-based emergent language systems.

Recommendations for Systematic Research on Emergent Language

no code implementations22 Jun 2022 Brendon Boldt, David Mortensen

Emergent language is unique among fields within the discipline of machine learning for its open-endedness, not obviously presenting well-defined problems to be solved.

Modeling Emergent Lexicon Formation with a Self-Reinforcing Stochastic Process

1 code implementation22 Jun 2022 Brendon Boldt, David Mortensen

We introduce FiLex, a self-reinforcing stochastic process which models finite lexicons in emergent language experiments.

Quantifying Cognitive Factors in Lexical Decline

1 code implementation12 Oct 2021 David Francis, Ella Rabinovich, Farhan Samir, David Mortensen, Suzanne Stevenson

Specifically, we propose a variety of psycholinguistic factors -- semantic, distributional, and phonological -- that we hypothesize are predictive of lexical decline, in which words greatly decrease in frequency over time.

Diversity

Polyglot Neural Language Models: A Case Study in Cross-Lingual Phonetic Representation Learning

no code implementations NAACL 2016 Yulia Tsvetkov, Sunayana Sitaram, Manaal Faruqui, Guillaume Lample, Patrick Littell, David Mortensen, Alan W. black, Lori Levin, Chris Dyer

We introduce polyglot language models, recurrent neural network models trained to predict symbol sequences in many different languages using shared representations of symbols and conditioning on typological information about the language to be predicted.

Representation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.