Search Results for author: Alexander Rush

Found 9 papers, 3 papers with code

Sequence-to-Lattice Models for Fast Translation

no code implementations Findings (EMNLP) 2021 Yuntian Deng, Alexander Rush

Non-autoregressive machine translation (NAT) approaches enable fast generation by utilizing parallelizable generative processes.

Machine Translation Translation

Torch-Struct: Deep Structured Prediction Library

1 code implementation ACL 2020 Alexander Rush

The literature on structured prediction for NLP describes a rich collection of distributions and algorithms over sequences, segmentations, alignments, and trees; however, these algorithms are difficult to utilize in deep learning frameworks.

Structured Prediction

Template Filling with Generative Transformers

1 code implementation NAACL 2021 Xinya Du, Alexander Rush, Claire Cardie

Template filling is generally tackled by a pipeline of two separate supervised systems {--} one for role-filler extraction and another for template/event recognition.

AdaptivFloat: A Floating-point based Data Type for Resilient Deep Learning Inference

no code implementations29 Sep 2019 Thierry Tambe, En-Yu Yang, Zishen Wan, Yuntian Deng, Vijay Janapa Reddi, Alexander Rush, David Brooks, Gu-Yeon Wei

Conventional hardware-friendly quantization methods, such as fixed-point or integer, tend to perform poorly at very low word sizes as their shrinking dynamic ranges cannot adequately capture the wide data distributions commonly seen in sequence transduction models.


Tensor Variable Elimination for Plated Factor Graphs

no code implementations8 Feb 2019 Fritz Obermeyer, Eli Bingham, Martin Jankowiak, Justin Chiu, Neeraj Pradhan, Alexander Rush, Noah Goodman

To exploit efficient tensor algebra in graphs with plates of variables, we generalize undirected factor graphs to plated factor graphs and variable elimination to a tensor variable elimination algorithm that operates directly on plated factor graphs.

Music Modeling Probabilistic Programming +1

Cannot find the paper you are looking for? You can Submit a new open access paper.