no code implementations • ICML 2020 • Justin DeBenedetto, David Chiang
Unordered, variable-sized inputs arise in many settings across multiple fields.
1 code implementation • 17 Aug 2024 • Kenneth J. Sible, David Chiang
In machine translation, rare words continue to be a problem for the dominant encoder-decoder architecture, especially in low-resource and out-of-domain translation settings.
1 code implementation • 13 Jun 2024 • Chihiro Taguchi, David Chiang
We investigate what linguistic factors affect the performance of Automatic Speech Recognition (ASR) models.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 25 Apr 2024 • Stephen Bothwell, Brian DuSell, David Chiang, Brian Krostenko
To assist historical linguists in the study of Italic sound change, we introduce the Proto-Italic to Latin (PILA) dataset, which consists of roughly 3, 000 pairs of forms from Proto-Italic and Latin.
1 code implementation • 23 Apr 2024 • Chihiro Taguchi, Jefferson Saransig, Dayana Velásquez, David Chiang
This dataset, the ASR model, and the code used to develop them will be publicly available.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 11 Apr 2024 • Stephen Bothwell, Abigail Swenor, David Chiang
This paper describes submissions from the team Nostra Domina to the EvaLatin 2024 shared task of emotion polarity detection.
1 code implementation • 10 Apr 2024 • Aarohi Srivastava, David Chiang
We present a suite of experiments that allow us to understand the underlying challenges of language model adaptation to nonstandard text.
no code implementations • 5 Apr 2024 • Andy Yang, David Chiang
Deriving formal bounds on the expressivity of transformers, as well as studying transformers that are constructed to implement known algorithms, are both effective methods for better understanding the computational power of transformers.
no code implementations • 2 Apr 2024 • Lena Strobl, Dana Angluin, David Chiang, Jonathan Rawski, Ashish Sabharwal
We study the sequence-to-sequence mapping capacity of transformers by relating them to finite transducers, and find that they can express surprisingly large classes of transductions.
1 code implementation • 16 Mar 2024 • Fahim Faisal, Orevaoghene Ahia, Aarohi Srivastava, Kabir Ahuja, David Chiang, Yulia Tsvetkov, Antonios Anastasopoulos
This allows for a comprehensive evaluation of NLP system performance on different language varieties.
4 code implementations • 30 Nov 2023 • Stephen Bothwell, Justin DeBenedetto, Theresa Crnkovich, Hildegund Müller, David Chiang
Rhetoric, both spoken and written, involves not only content but also style.
no code implementations • 1 Nov 2023 • Lena Strobl, William Merrill, Gail Weiss, David Chiang, Dana Angluin
As transformers have gained prominence in natural language processing, some researchers have investigated theoretically what problems they can and cannot solve, by treating problems as formal languages.
no code implementations • 31 Oct 2023 • Aarohi Srivastava, David Chiang
Real-world NLP applications often deal with nonstandard text (e. g., dialectal, informal, or misspelled text).
no code implementations • 23 Oct 2023 • Alexandra Butoi, Tim Vieira, Ryan Cotterell, David Chiang
From these, we also immediately obtain stringsum and allsum algorithms for TAG, LIG, PAA, and EPDA.
no code implementations • 21 Oct 2023 • Andy Yang, David Chiang, Dana Angluin
The expressive power of transformers over inputs of unbounded size can be studied through their ability to recognize classes of formal languages.
1 code implementation • 3 Oct 2023 • Brian DuSell, David Chiang
Attention, specifically scaled dot-product attention, has proven effective for natural language, but it does not have a mechanism for handling hierarchical patterns of arbitrary nesting depth, which limits its ability to recognize certain syntactic structures.
1 code implementation • 7 Aug 2023 • Chihiro Taguchi, Yusuke Sakai, Parisa Haghani, David Chiang
This paper presents a state-of-the-art model for transcribing speech in any language into the International Phonetic Alphabet (IPA).
no code implementations • 6 Jun 2023 • Alexandra Butoi, Ryan Cotterell, David Chiang
Furthermore, using an even stricter notion of equivalence called d-strong equivalence, we make precise the intuition that a CFG controlling a CFG is a TAG, a PDA controlling a PDA is an embedded PDA, and a PDA controlling a CFG is a LIG.
no code implementations • 30 Mar 2023 • Aarohi Srivastava, David Chiang
In this work, we induce character-level noise in various forms when fine-tuning BERT to enable zero-shot cross-lingual transfer to unseen dialects and languages.
no code implementations • 25 Jan 2023 • David Chiang, Peter Cholak, Anand Pillay
Characterizing neural networks in terms of better-understood formal systems has the potential to yield new insights into the power and limitations of these networks.
no code implementations • 13 Dec 2022 • Patrick Soga, David Chiang
A current goal in the graph neural network literature is to enable transformers to operate on graph-structured data, given their success on language and vision tasks.
no code implementations • 19 Oct 2022 • Darcey Riley, David Chiang
Language models suffer from various degenerate behaviors.
1 code implementation • 13 Oct 2022 • Alexandra Butoi, Brian DuSell, Tim Vieira, Ryan Cotterell, David Chiang
Weighted pushdown automata (WPDAs) are at the core of many natural language processing tasks, like syntax-based statistical machine translation and transition-based dependency parsing.
2 code implementations • 4 Oct 2022 • Brian DuSell, David Chiang
Second, it can recognize languages with much larger alphabet sizes than one might expect given the size of its stack alphabet.
1 code implementation • ACL 2022 • David Chiang, Peter Cholak
We examine this limitation using two languages: PARITY, the language of bit strings with an odd number of 1s, and FIRST, the language of bit strings starting with a 1.
1 code implementation • ICLR 2022 • Brian DuSell, David Chiang
Learning hierarchical structures in sequential data -- from simple algorithmic patterns to natural language -- in a reliable, generalizable way remains a challenging problem for neural language models.
no code implementations • NAACL 2021 • Colin McDonald, David Chiang
We present a simple method for extending transformers to source-side trees.
no code implementations • ACL (IWSLT) 2021 • Toan Q. Nguyen, Kenton Murray, David Chiang
In this paper, we investigate the driving factors behind concatenation, a simple but effective data augmentation method for low-resource neural machine translation.
1 code implementation • 25 Feb 2021 • David Chiang, Alexander M. Rush, Boaz Barak
We propose a notation for tensors with named axes, which relieves the author, reader, and future implementers of machine learning models from the burden of keeping track of the order of axes and the purpose of each.
no code implementations • 22 Oct 2020 • David Chiang, Chung-chieh Shan
It is natural for probabilistic programs to use conditionals to express alternative substructures in models, and loops (recursion) to express repeated substructures in models.
1 code implementation • NeurIPS 2020 • David Chiang, Darcey Riley
We propose the use of hyperedge replacement graph grammars for factor graphs, or factor graph grammars (FGGs) for short.
no code implementations • WMT (EMNLP) 2020 • Xing Jie Zhong, David Chiang
Despite advances in neural machine translation (NMT) quality, rare words continue to be problematic.
1 code implementation • CONLL 2020 • Brian DuSell, David Chiang
We present a differentiable stack data structure that simultaneously and tractably encodes an exponential number of stack configurations, based on Lang's algorithm for simulating nondeterministic pushdown automata.
no code implementations • 2 Jan 2020 • Justin DeBenedetto, David Chiang
Unordered, variable-sized inputs arise in many settings across multiple fields.
no code implementations • WS 2019 • Kenton Murray, Brian DuSell, David Chiang
We investigated the impact of auto-sizing (Murray and Chiang, 2015; Murray et al., 2019) to the Transformer network (Vaswani et al., 2017) with the goal of substantially reducing the number of parameters in the model.
1 code implementation • WS 2019 • Kenton Murray, Jeffery Kinnison, Toan Q. Nguyen, Walter Scheirer, David Chiang
Neural sequence-to-sequence models, particularly the Transformer, are the state of the art in machine translation.
no code implementations • ACL 2019 • Arturo Argueta, David Chiang
Operations using sparse structures are common in natural language models at the input and output layers, because these models operate on sequences over discrete alphabets.
no code implementations • 7 Apr 2019 • Samuel Grieggs, Bingyu Shen, Greta Rauch, Pei Li, Jiaqi Ma, David Chiang, Brian Price, Walter J. Scheirer
The subtleties of human perception, as measured by vision scientists through the use of psychophysics, are important clues to the internal workings of visual recognition.
no code implementations • WS 2018 • Kenton Murray, David Chiang
We study two problems in neural machine translation (NMT).
2 code implementations • NAACL 2019 • Antonios Anastasopoulos, Alison Lui, Toan Nguyen, David Chiang
Neural Machine Translation (NMT) systems are known to degrade when confronted with noisy data, especially when the system is trained only on clean data.
1 code implementation • COLING 2018 • Antonis Anastasopoulos, Marika Lekakou, Josep Quer, Eleni Zimianiti, Justin DeBenedetto, David Chiang
Most work on part-of-speech (POS) tagging is focused on high resource languages, or examines low-resource and active learning settings through simulated studies.
no code implementations • NAACL 2018 • Huadong Chen, Shu-Jian Huang, David Chiang, Xin-yu Dai, Jia-Jun Chen
Natural language sentences, being hierarchical, can be represented at different levels of granularity, like words, subwords, or characters.
no code implementations • ACL 2018 • Arturo Argueta, David Chiang
Weighted finite-state transducers (FSTs) are frequently used in language processing to handle tasks such as part-of-speech tagging and speech recognition.
1 code implementation • 23 Mar 2018 • Antonis Anastasopoulos, David Chiang
Recently proposed data collection frameworks for endangered language documentation aim not only to collect speech in the language of interest, but also to collect translations into a high-resource language that will render the collected resource interpretable.
no code implementations • CL 2018 • David Chiang, Frank Drewes, Daniel Gildea, Adam Lopez, Giorgio Satta
Graphs have a variety of uses in natural language processing, particularly as representations of linguistic meaning.
no code implementations • NAACL 2018 • Antonios Anastasopoulos, David Chiang
We explore multitask models for neural translation of speech, augmenting them in order to reflect two intuitive notions.
4 code implementations • NAACL 2018 • Toan Q. Nguyen, David Chiang
We explore two solutions to the problem of mistranslating rare words in neural machine translation.
no code implementations • WS 2017 • Antonios Anastasopoulos, Sameer Bansal, David Chiang, Sharon Goldwater, Adam Lopez
Vast amounts of speech data collected for language documentation and research remain untranscribed and unsearchable, but often a small amount of speech may have text translations available.
no code implementations • IJCNLP 2017 • Toan Q. Nguyen, David Chiang
We present a simple method to improve neural translation of a low-resource language pair using parallel data from a related, also low-resource, language pair.
no code implementations • CONLL 2017 • Huadong Chen, Shu-Jian Huang, David Chiang, Xin-yu Dai, Jia-Jun Chen
We propose a listwise learning framework for structure prediction problems such as machine translation.
1 code implementation • ACL 2017 • Huadong Chen, Shu-Jian Huang, David Chiang, Jia-Jun Chen
Most neural machine translation (NMT) models are based on the sequential encoder-decoder framework, which makes no use of syntactic information.
no code implementations • WS 2017 • Antonios Anastasopoulos, David Chiang
For many low-resource or endangered languages, spoken language resources are more likely to be annotated with translations than with transcriptions.
4 code implementations • 15 Jan 2017 • Graham Neubig, Chris Dyer, Yoav Goldberg, Austin Matthews, Waleed Ammar, Antonios Anastasopoulos, Miguel Ballesteros, David Chiang, Daniel Clothiaux, Trevor Cohn, Kevin Duh, Manaal Faruqui, Cynthia Gan, Dan Garrette, Yangfeng Ji, Lingpeng Kong, Adhiguna Kuncoro, Gaurav Kumar, Chaitanya Malaviya, Paul Michel, Yusuke Oda, Matthew Richardson, Naomi Saphra, Swabha Swayamdipta, Pengcheng Yin
In the static declaration strategy that is used in toolkits like Theano, CNTK, and TensorFlow, the user first defines a computation graph (a symbolic representation of the computation), and then examples are fed into an engine that executes this computation and computes its derivatives.
no code implementations • EACL 2017 • Arturo Argueta, David Chiang
Weighted finite automata and transducers (including hidden Markov models and conditional random fields) are widely used in natural language processing (NLP) to perform tasks such as morphological analysis, part-of-speech tagging, chunking, named entity recognition, speech recognition, and others.
1 code implementation • EMNLP 2016 • Antonios Anastasopoulos, David Chiang, Long Duong
For many low-resource languages, spoken language resources are more likely to be annotated with translations than with transcriptions.
1 code implementation • 10 Aug 2016 • Salvador Aguiñaga, Rodrigo Palacios, David Chiang, Tim Weninger
In experiments on large real world networks, we show that random graphs, generated from extracted graph grammars, exhibit a wide range of properties that are very similar to the original graphs.
no code implementations • EMNLP 2015 • Kenton Murray, David Chiang
Neural networks have been shown to improve performance across a range of natural-language tasks.