no code implementations • ACL (WebNLG, INLG) 2020 • Oshin Agarwal, Mihir Kale, Heming Ge, Siamak Shakeri, Rami Al-Rfou
We present a system for bilingual Data-ToText Generation and Semantic Parsing.
no code implementations • CVPR 2024 • Norman Mu, Jingwei Ji, Zhenpei Yang, Nate Harada, Haotian Tang, Kan Chen, Charles R. Qi, Runzhou Ge, Kratarth Goel, Zoey Yang, Scott Ettinger, Rami Al-Rfou, Dragomir Anguelov, Yin Zhou
This symbolic representation is a high-level abstraction of the real world, which may render the motion prediction model vulnerable to perception errors (e. g., failures in detecting open-vocabulary obstacles) while missing salient information from the scene context (e. g., poor road conditions).
no code implementations • 5 Apr 2024 • Scott Ettinger, Kratarth Goel, Avikalp Srivastava, Rami Al-Rfou
These experiments demonstrate distillation from ensembles as an effective method for improving accuracy of predictive models for robotic systems with limited compute budgets.
no code implementations • 8 Feb 2024 • Bryan Perozzi, Bahare Fatemi, Dustin Zelle, Anton Tsitsulin, Mehran Kazemi, Rami Al-Rfou, Jonathan Halcrow
How can we best encode structured data into sequential form for use in large language models (LLMs)?
1 code implementation • ICCV 2023 • Ari Seff, Brian Cera, Dian Chen, Mason Ng, Aurick Zhou, Nigamaa Nayakanti, Khaled S. Refaat, Rami Al-Rfou, Benjamin Sapp
Here, we represent continuous trajectories as sequences of discrete motion tokens and cast multi-agent motion prediction as a language modeling task over this domain.
no code implementations • 25 Mar 2023 • Bashar Al-Rfooh, Gheith Abandah, Rami Al-Rfou
Most of previous work on learning diacritization of the Arabic language relied on training models from scratch.
3 code implementations • 12 Jul 2022 • Nigamaa Nayakanti, Rami Al-Rfou, Aurick Zhou, Kratarth Goel, Khaled S. Refaat, Benjamin Sapp
In this paper, we present Wayformer, a family of attention based architectures for motion forecasting that are simple and homogeneous.
Ranked #6 on Motion Forecasting on Argoverse CVPR 2020
no code implementations • 8 Jun 2022 • Serge Assaad, Carlton Downey, Rami Al-Rfou, Nigamaa Nayakanti, Ben Sapp
Rotation equivariance is a desirable property in many practical applications such as motion forecasting and 3D perception, where it can offer benefits like sample efficiency, better generalization, and robustness to input perturbations.
no code implementations • 8 Jun 2022 • DiJia Su, Bertrand Douillard, Rami Al-Rfou, Cheolho Park, Benjamin Sapp
These models are intrinsically invariant to translation and rotation between scene elements, are best-performing on public leaderboards, but scale quadratically with the number of agents and scene elements.
no code implementations • ACL 2022 • Tu Vu, Brian Lester, Noah Constant, Rami Al-Rfou, Daniel Cer
Finally, we propose an efficient retrieval approach that interprets task prompts as task embeddings to identify similar tasks and predict the most transferable source tasks for a novel target task.
no code implementations • ACL 2021 • Mihir Kale, Aditya Siddhant, Rami Al-Rfou, Linting Xue, Noah Constant, Melvin Johnson
Recently, mT5 - a massively multilingual version of T5 - leveraged a unified text-to-text format to attain state-of-the-art results on a wide variety of multilingual NLP tasks.
no code implementations • 3 Jun 2021 • Mihir Kale, Aditya Siddhant, Noah Constant, Melvin Johnson, Rami Al-Rfou, Linting Xue
Recently, mT5 - a massively multilingual version of T5 - leveraged a unified text-to-text format to attain state-of-the-art results on a wide variety of multilingual NLP tasks.
5 code implementations • 28 May 2021 • Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel
Most widely-used pre-trained language models operate on sequences of tokens corresponding to word or subword units.
Ranked #1 on Cross-Lingual Natural Language Inference on XNLI
Cross-Lingual Natural Language Inference Cross-Lingual NER +3
10 code implementations • EMNLP 2021 • Brian Lester, Rami Al-Rfou, Noah Constant
More remarkably, through ablations on model size using T5, we show that prompt tuning becomes more competitive with scale: as models exceed billions of parameters, our method "closes the gap" and matches the strong performance of model tuning (where all model weights are tuned).
1 code implementation • NAACL 2021 • Oshin Agarwal, Heming Ge, Siamak Shakeri, Rami Al-Rfou
Prior work on Data-To-Text Generation, the task of converting knowledge graph (KG) triples into natural text, focused on domain-specific benchmark datasets.
8 code implementations • NAACL 2021 • Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, Colin Raffel
The recent "Text-to-Text Transfer Transformer" (T5) leveraged a unified text-to-text format and scale to attain state-of-the-art results on a wide variety of English-language NLP tasks.
Ranked #2 on Reading Comprehension on MuSeRC
no code implementations • LREC 2020 • M. Guo, y, Zihang Dai, Vr, Denny e{\v{c}}i{\'c}, Rami Al-Rfou
We released the cleaned-up text of 40+ Wikipedia language editions, the corresponding trained monolingual language models, and several multilingual language models with different fixed vocabulary sizes.
1 code implementation • EMNLP 2020 • Uma Roy, Noah Constant, Rami Al-Rfou, Aditya Barua, Aaron Phillips, Yinfei Yang
We present LAReQA, a challenging new benchmark for language-agnostic answer retrieval from a multilingual candidate pool.
no code implementations • 27 Aug 2019 • Dokook Choe, Rami Al-Rfou, Mandy Guo, Heeyoung Lee, Noah Constant
Purely character-based language models (LMs) have been lagging in quality on large scale datasets, and current state-of-the-art LMs rely on word tokenization.
1 code implementation • 21 Apr 2019 • Rami Al-Rfou, Dustin Zelle, Bryan Perozzi
Second, for each pair of graphs, we train a cross-graph attention network which uses the node representations of an anchor graph to reconstruct another graph.
Ranked #3 on Graph Classification on D&D
1 code implementation • 9 Aug 2018 • Rami Al-Rfou, Dokook Choe, Noah Constant, Mandy Guo, Llion Jones
LSTMs and other RNN variants have shown strong performance on character-level language modeling.
Ranked #8 on Language Modelling on Hutter Prize
2 code implementations • 8 Aug 2018 • Haochen Chen, Bryan Perozzi, Rami Al-Rfou, Steven Skiena
We further demonstrate the applications of network embeddings, and conclude the survey with future work in this area.
Social and Information Networks
2 code implementations • NeurIPS 2018 • Sami Abu-El-Haija, Bryan Perozzi, Rami Al-Rfou, Alex Alemi
Graph embedding methods represent nodes in a continuous vector space, preserving information from the graph (e. g. by sampling random walks).
Ranked #67 on Node Classification on Citeseer
2 code implementations • 7 Jun 2017 • Mustafa Mustafa, Deborah Bard, Wahid Bhimji, Zarija Lukić, Rami Al-Rfou, Jan M. Kratochvil
Inferring model parameters from experimental data is a grand challenge in many sciences, including cosmology.
1 code implementation • 16 May 2017 • Sami Abu-El-Haija, Bryan Perozzi, Rami Al-Rfou
Individually, both of these contributions improve the learned representations, especially when there are memory constraints on the total size of the embeddings.
no code implementations • 1 May 2017 • Matthew Henderson, Rami Al-Rfou, Brian Strope, Yun-Hsuan Sung, Laszlo Lukacs, Ruiqi Guo, Sanjiv Kumar, Balint Miklos, Ray Kurzweil
This paper presents a computationally efficient machine-learned method for natural language response suggestion.
no code implementations • 20 Nov 2016 • Salman Mahmood, Rami Al-Rfou, Klaus Mueller
Neural network based models are a very powerful tool for creating word embeddings, the objective of these models is to group similar words together.
no code implementations • 20 Oct 2016 • Marc Pickett, Rami Al-Rfou, Louis Shao, Chris Tar
The long-term memory of most connectionist systems lies entirely in the weights of the system.
no code implementations • 1 Jun 2016 • Rami Al-Rfou, Marc Pickett, Javier Snaider, Yun-Hsuan Sung, Brian Strope, Ray Kurzweil
Unlike previous efforts, which focused on modeling messages and responses, we extend the modeling to long context and participant's history.
1 code implementation • 9 May 2016 • The Theano Development Team, Rami Al-Rfou, Guillaume Alain, Amjad Almahairi, Christof Angermueller, Dzmitry Bahdanau, Nicolas Ballas, Frédéric Bastien, Justin Bayer, Anatoly Belikov, Alexander Belopolsky, Yoshua Bengio, Arnaud Bergeron, James Bergstra, Valentin Bisson, Josh Bleecher Snyder, Nicolas Bouchard, Nicolas Boulanger-Lewandowski, Xavier Bouthillier, Alexandre de Brébisson, Olivier Breuleux, Pierre-Luc Carrier, Kyunghyun Cho, Jan Chorowski, Paul Christiano, Tim Cooijmans, Marc-Alexandre Côté, Myriam Côté, Aaron Courville, Yann N. Dauphin, Olivier Delalleau, Julien Demouth, Guillaume Desjardins, Sander Dieleman, Laurent Dinh, Mélanie Ducoffe, Vincent Dumoulin, Samira Ebrahimi Kahou, Dumitru Erhan, Ziye Fan, Orhan Firat, Mathieu Germain, Xavier Glorot, Ian Goodfellow, Matt Graham, Caglar Gulcehre, Philippe Hamel, Iban Harlouchet, Jean-Philippe Heng, Balázs Hidasi, Sina Honari, Arjun Jain, Sébastien Jean, Kai Jia, Mikhail Korobov, Vivek Kulkarni, Alex Lamb, Pascal Lamblin, Eric Larsen, César Laurent, Sean Lee, Simon Lefrancois, Simon Lemieux, Nicholas Léonard, Zhouhan Lin, Jesse A. Livezey, Cory Lorenz, Jeremiah Lowin, Qianli Ma, Pierre-Antoine Manzagol, Olivier Mastropietro, Robert T. McGibbon, Roland Memisevic, Bart van Merriënboer, Vincent Michalski, Mehdi Mirza, Alberto Orlandi, Christopher Pal, Razvan Pascanu, Mohammad Pezeshki, Colin Raffel, Daniel Renshaw, Matthew Rocklin, Adriana Romero, Markus Roth, Peter Sadowski, John Salvatier, François Savard, Jan Schlüter, John Schulman, Gabriel Schwartz, Iulian Vlad Serban, Dmitriy Serdyuk, Samira Shabanian, Étienne Simon, Sigurd Spieckermann, S. Ramana Subramanyam, Jakub Sygnowski, Jérémie Tanguay, Gijs van Tulder, Joseph Turian, Sebastian Urban, Pascal Vincent, Francesco Visin, Harm de Vries, David Warde-Farley, Dustin J. Webb, Matthew Willson, Kelvin Xu, Lijun Xue, Li Yao, Saizheng Zhang, Ying Zhang
Since its introduction, it has been one of the most used CPU and GPU mathematical compilers - especially in the machine learning community - and has shown steady performance improvements.
no code implementations • 12 Nov 2014 • Vivek Kulkarni, Rami Al-Rfou, Bryan Perozzi, Steven Skiena
We propose a new computational approach for tracking and detecting statistically significant linguistic shifts in the meaning and usage of words.
no code implementations • 14 Oct 2014 • Rami Al-Rfou, Vivek Kulkarni, Bryan Perozzi, Steven Skiena
We describe a system that builds Named Entity Recognition (NER) annotators for 40 major languages using Wikipedia and Freebase.
14 code implementations • 26 Mar 2014 • Bryan Perozzi, Rami Al-Rfou, Steven Skiena
We present DeepWalk, a novel approach for learning latent representations of vertices in a network.
Ranked #1 on Link Property Prediction on ogbl-ppa
no code implementations • 6 Mar 2014 • Bryan Perozzi, Rami Al-Rfou, Vivek Kulkarni, Steven Skiena
Recent advancements in unsupervised feature learning have developed powerful latent representations of words.
no code implementations • WS 2013 • Rami Al-Rfou, Bryan Perozzi, Steven Skiena
We quantitatively demonstrate the utility of our word embeddings by using them as the sole features for training a part of speech tagger for a subset of these languages.
no code implementations • 15 Jan 2013 • Yanqing Chen, Bryan Perozzi, Rami Al-Rfou, Steven Skiena
We seek to better understand the difference in quality of the several publicly released embeddings.