no code implementations • • Lei Yu, Laurent Sartran, Po-Sen Huang, Wojciech Stokowiec, Domenic Donato, Srivatsan Srinivasan, Alek Andreev, Wang Ling, Sona Mokra, Agustin Dal Lago, Yotam Doron, Susannah Young, Phil Blunsom, Chris Dyer
This paper describes the DeepMind submission to the Chinese\rightarrowEnglish constrained data track of the WMT2020 Shared Task on News Translation.
After just a few hundred training updates, a standard probabilistic model for language generation has likely not yet learnt many semantic or syntactic rules of natural language, making it difficult to estimate the probability distribution over next tokens.
In this work, we aim to capitalize on the unique few-shot capabilities of large-scale language models (LSLMs) to overcome some of their challenges with respect to grounding to factual and up-to-date information.
When applied to autoregressive models, our algorithm has different biases than beam search has, which enables a new analysis of the role of decoding bias in autoregressive models.
We show that Bayes' rule provides an effective mechanism for creating document translation models that can be learned from only parallel sentences and monolingual documents---a compelling benefit as parallel documents are not always available.
We show that Bayes' rule provides a compelling mechanism for controlling unconditional document language models, using the long-standing challenge of effectively leveraging document context in machine translation.
Deep generative architectures provide a way to model not only images but also complex, 3-dimensional objects, such as point clouds.
We introduce a new dataset of 293, 008 high definition (1360 x 1360 pixels) fashion images paired with item descriptions provided by professional stylists.
In this paper we propose a new method of speaker diarization that employs a deep learning architecture to learn speaker embeddings.
In this paper, we propose a multi-modal search engine for interior design that combines visual and textual queries.
The web data contains immense amount of data, hundreds of billion words are waiting to be extracted and used for language research.