no code implementations • 30 Jan 2020 • Ashiqur R. KhudaBukhsh, Shriphani Palakodety, Jaime G. Carbonell
Code mixing (or code switching) is a common phenomenon observed in social-media content generated by a linguistically diverse user-base.
no code implementations • 8 Oct 2019 • Shriphani Palakodety, Ashiqur R. KhudaBukhsh, Jaime G. Carbonell
The Rohingya refugee crisis is one of the biggest humanitarian crises of modern times with more than 600, 000 Rohingyas rendered homeless according to the United Nations High Commissioner for Refugees.
no code implementations • 11 Sep 2019 • Shriphani Palakodety, Ashiqur R. KhudaBukhsh, Jaime G. Carbonell
The recent Pulwama terror attack (February 14, 2019, Pulwama, Kashmir) triggered a chain of escalating events between India and Pakistan adding another episode to their 70-year-old dispute over Kashmir.
1 code implementation • IJCNLP 2019 • Aditi Chaudhary, Jiateng Xie, Zaid Sheikh, Graham Neubig, Jaime G. Carbonell
Most state-of-the-art models for named entity recognition (NER) rely on the availability of large amounts of labeled data, making them challenging to extend to new, lower-resourced languages.
1 code implementation • WS 2019 • Aditi Chaudhary, Elizabeth Salesky, Gayatri Bhat, David R. Mortensen, Jaime G. Carbonell, Yulia Tsvetkov
This paper presents the submission by the CMU-01 team to the SIGMORPHON 2019 task 2 of Morphological Analysis and Lemmatization in Context.
no code implementations • 4 May 2019 • Liang Xiong, Xi Chen, Tzu-Kuo Huang, Jeff Schneider, Jaime G. Carbonell
Motivated by our sales prediction problem, we propose a factor-based algorithm that is able to take time into account.
1 code implementation • EMNLP 2018 • Aditi Chaudhary, Chunting Zhou, Lori Levin, Graham Neubig, David R. Mortensen, Jaime G. Carbonell
Much work in Natural Language Processing (NLP) has been for resource-rich languages, making generalization to new, less-resourced languages challenging.
no code implementations • ICLR 2019 • George Philipp, Jaime G. Carbonell
Via an extensive empirical study, we show that the NLC is a powerful predictor of test error and that attaining a right-sized NLC is essential for optimal performance.
no code implementations • ICLR 2018 • George Philipp, Dawn Song, Jaime G. Carbonell
Whereas it is believed that techniques such as Adam, batch normalization and, more recently, SeLU nonlinearities ``solve'' the exploding gradient problem, we show that this is not the case and that in a range of popular MLP architectures, exploding gradients exist and that they limit the depth to which networks can be effectively trained, both in theory and in practice.
no code implementations • 15 Dec 2017 • George Philipp, Dawn Song, Jaime G. Carbonell
Whereas it is believed that techniques such as Adam, batch normalization and, more recently, SeLU nonlinearities "solve" the exploding gradient problem, we show that this is not the case in general and that in a range of popular MLP architectures, exploding gradients exist and that they limit the depth to which networks can be effectively trained, both in theory and in practice.
no code implementations • 14 Dec 2017 • George Philipp, Jaime G. Carbonell
Automatically determining the optimal size of a neural network for a given task without prior information currently requires an expensive global search and training many networks from scratch.