no code implementations • WMT (EMNLP) 2020 • Lei Yu, Laurent Sartran, Po-Sen Huang, Wojciech Stokowiec, Domenic Donato, Srivatsan Srinivasan, Alek Andreev, Wang Ling, Sona Mokra, Agustin Dal Lago, Yotam Doron, Susannah Young, Phil Blunsom, Chris Dyer
This paper describes the DeepMind submission to the Chinese\rightarrowEnglish constrained data track of the WMT2020 Shared Task on News Translation.
no code implementations • ICLR 2019 • Tiago Ramalho, Tomas Kocisky, Frederic Besse, S. M. Ali Eslami, Gabor Melis, Fabio Viola, Phil Blunsom, Karl Moritz Hermann
Natural language processing has made significant inroads into learning the semantics of words through distributional approaches, however representations learnt via these methods fail to capture certain kinds of information implicit in the real world.
no code implementations • EMNLP 2021 • Yishu Miao, Phil Blunsom, Lucia Specia
We propose a generative framework for simultaneous machine translation.
1 code implementation • 15 Aug 2024 • Qizhen Zhang, Nikolas Gritsch, Dwaraknath Gnaneshwar, Simon Guo, David Cairuz, Bharat Venkitesh, Jakob Foerster, Phil Blunsom, Sebastian Ruder, Ahmet Ustun, Acyr Locatelli
BAM makes full use of specialized dense models by not only using their FFN to initialize the MoE layers but also leveraging experts' attention parameters fully by initializing them into a soft-variant of Mixture of Attention (MoA) layers.
no code implementations • 13 Jun 2024 • Satwik Bhattamishra, Michael Hahn, Phil Blunsom, Varun Kanade
Furthermore, we show that two-layer Transformers of logarithmic size can perform decision tasks such as string equality or disjointness, whereas both one-layer Transformers and recurrent models require linear size for these tasks.
no code implementations • 31 May 2024 • Zihuiwen Ye, Fraser Greenlee-Scott, Max Bartolo, Phil Blunsom, Jon Ander Campos, Matthias Gallé
This offers richer signals and more robust features for RMs to assess and score on.
no code implementations • 23 May 2024 • Viraat Aryabumi, John Dang, Dwarak Talupuru, Saurabh Dash, David Cairuz, Hangyu Lin, Bharat Venkitesh, Madeline Smith, Jon Ander Campos, Yi Chern Tan, Kelly Marchisio, Max Bartolo, Sebastian Ruder, Acyr Locatelli, Julia Kreutzer, Nick Frosst, Aidan Gomez, Phil Blunsom, Marzieh Fadaee, Ahmet Üstün, Sara Hooker
This technical report introduces Aya 23, a family of multilingual language models.
no code implementations • 12 Feb 2024 • Ahmet Üstün, Viraat Aryabumi, Zheng-Xin Yong, Wei-Yin Ko, Daniel D'souza, Gbemileke Onilude, Neel Bhandari, Shivalika Singh, Hui-Lee Ooi, Amr Kayid, Freddie Vargus, Phil Blunsom, Shayne Longpre, Niklas Muennighoff, Marzieh Fadaee, Julia Kreutzer, Sara Hooker
Recent breakthroughs in large language models (LLMs) have centered around a handful of data-rich languages.
no code implementations • 4 Oct 2023 • Satwik Bhattamishra, Arkil Patel, Phil Blunsom, Varun Kanade
In this work, we take a step towards answering these questions by demonstrating the following: (a) On a test-bed with a variety of Boolean function classes, we find that Transformers can nearly match the optimal learning algorithm for 'simpler' tasks, while their performance deteriorates on more 'complex' tasks.
1 code implementation • 28 Sep 2023 • Tom Hosking, Phil Blunsom, Max Bartolo
We critically analyse the use of human feedback for both training and evaluation, to verify whether it fully captures a range of crucial error criteria.
no code implementations • 31 Jul 2023 • Kyle Duffy, Satwik Bhattamishra, Phil Blunsom
Large-scale pre-training has made progress in many fields of natural language processing, though little is understood about the design of pre-training datasets.
no code implementations • 5 Jun 2023 • Made Nindyatama Nityasya, Haryo Akbarianto Wibowo, Alham Fikri Aji, Genta Indra Winata, Radityo Eko Prasojo, Phil Blunsom, Adhiguna Kuncoro
This evidence-based position paper critiques current research practices within the language model pre-training literature.
1 code implementation • 22 Nov 2022 • Satwik Bhattamishra, Arkil Patel, Varun Kanade, Phil Blunsom
(ii) When trained on Boolean functions, both Transformers and LSTMs prioritize learning functions of low sensitivity, with Transformers ultimately converging to functions of lower sensitivity.
1 code implementation • 21 Oct 2022 • Qi Liu, Zihuiwen Ye, Tao Yu, Phil Blunsom, Linfeng Song
We first design a SQL-to-text model conditioned on a sampled goal query, which represents a user's intent, that then converses with a text-to-SQL semantic parser to generate new interactions.
no code implementations • 24 May 2022 • Aishwarya Agrawal, Ivana Kajić, Emanuele Bugliarello, Elnaz Davoodi, Anita Gergely, Phil Blunsom, Aida Nematzadeh
Vision-and-language (V&L) models pretrained on large-scale multimodal data have demonstrated strong performance on various tasks such as image captioning and visual question answering (VQA).
2 code implementations • 23 May 2022 • Adam Liška, Tomáš Kočiský, Elena Gribovskaya, Tayfun Terzi, Eren Sezener, Devang Agrawal, Cyprien de Masson d'Autume, Tim Scholtes, Manzil Zaheer, Susannah Young, Ellen Gilsenan-McMahon, Sophia Austin, Phil Blunsom, Angeliki Lazaridou
Knowledge and language understanding of models evaluated through question answering (QA) has been usually studied on static snapshots of knowledge, like Wikipedia.
1 code implementation • ACL 2022 • Arkil Patel, Satwik Bhattamishra, Phil Blunsom, Navin Goyal
Compositional generalization is a fundamental trait in humans, allowing us to effortlessly combine known phrases to form novel sentences.
no code implementations • 1 Mar 2022 • Laurent Sartran, Samuel Barrett, Adhiguna Kuncoro, Miloš Stanojević, Phil Blunsom, Chris Dyer
We find that TGs outperform various strong baselines on sentence-level language modeling perplexity, as well as on multiple syntax-sensitive language modeling evaluation metrics.
no code implementations • 24 Jan 2022 • Qi Liu, Dani Yogatama, Phil Blunsom
We present a memory-augmented approach to condition an autoregressive language model on a knowledge graph.
no code implementations • 31 Oct 2021 • Xiang Lorraine Li, Adhiguna Kuncoro, Jordan Hoffmann, Cyprien de Masson d'Autume, Phil Blunsom, Aida Nematzadeh
Language models (LMs) trained on large amounts of data have shown impressive performance on many NLP tasks under the zero-shot and few-shot setup.
1 code implementation • NAACL 2021 • Qi Liu, Matt Kusner, Phil Blunsom
We propose a data augmentation method for neural machine translation.
no code implementations • 18 Mar 2021 • Qi Liu, Lei Yu, Laura Rimell, Phil Blunsom
Direct decoding for task-oriented dialogue is known to suffer from the explaining-away effect, manifested in models that prefer short and generic responses.
Ranked #2 on End-To-End Dialogue Modelling on MULTIWOZ 2.0
1 code implementation • NeurIPS 2021 • Angeliki Lazaridou, Adhiguna Kuncoro, Elena Gribovskaya, Devang Agrawal, Adam Liska, Tayfun Terzi, Mai Gimenez, Cyprien de Masson d'Autume, Tomas Kocisky, Sebastian Ruder, Dani Yogatama, Kris Cao, Susannah Young, Phil Blunsom
Hence, given the compilation of ever-larger language modelling datasets, combined with the growing list of language-model-based NLP applications that require up-to-date factual knowledge about the world, we argue that now is the right time to rethink the static way in which we currently train and evaluate our language models, and develop adaptive language models that can remain up-to-date with respect to our ever-changing and non-stationary world.
no code implementations • 1 Dec 2020 • Gábor Melis, András György, Phil Blunsom
A common failure mode of density models trained as variational autoencoders is to model the data without relying on their latent variables, rendering these variables useless.
1 code implementation • 23 Sep 2020 • Oana-Maria Camburu, Eleonora Giunchiglia, Jakob Foerster, Thomas Lukasiewicz, Phil Blunsom
For neural models to garner widespread public trust and ensure fairness, we must have human-intelligible explanations for their predictions.
no code implementations • 27 May 2020 • Adhiguna Kuncoro, Lingpeng Kong, Daniel Fried, Dani Yogatama, Laura Rimell, Chris Dyer, Phil Blunsom
Textual representation learners trained on large amounts of data have achieved notable success on downstream tasks; intriguingly, they have also performed well on challenging tests of syntactic competence.
1 code implementation • ACL 2020 • Daniel Fried, Jean-Baptiste Alayrac, Phil Blunsom, Chris Dyer, Stephen Clark, Aida Nematzadeh
We apply a generative segmental model of task structure, guided by narration, to action segmentation in video.
no code implementations • 16 Mar 2020 • Qi Liu, Matt J. Kusner, Phil Blunsom
Contextual embeddings, such as ELMo and BERT, move beyond global word representations like Word2Vec and achieve ground-breaking performance on a wide range of natural language processing tasks.
1 code implementation • CVPR 2020 • Gunnar A. Sigurdsson, Jean-Baptiste Alayrac, Aida Nematzadeh, Lucas Smaira, Mateusz Malinowski, João Carreira, Phil Blunsom, Andrew Zisserman
Given this shared embedding we demonstrate that (i) we can map words between the languages, particularly the 'visual' words; (ii) that the shared embedding provides a good initialization for existing unsupervised text-based word translation techniques, forming the basis for our proposed hybrid visual-text mapping algorithm, MUVE; and (iii) our approach achieves superior performance by addressing the shortcomings of text-based methods -- it is more robust, handles datasets with less commonality, and is applicable to low-resource languages.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Kazuya Kawakami, Luyu Wang, Chris Dyer, Phil Blunsom, Aaron van den Oord
Unsupervised speech representation learning has shown remarkable success at finding representations that correlate with phonetic structures and improve downstream speech recognition performance.
1 code implementation • ACL 2020 • Oana-Maria Camburu, Brendan Shillingford, Pasquale Minervini, Thomas Lukasiewicz, Phil Blunsom
To increase trust in artificial intelligence systems, a promising research direction consists of designing neural models capable of generating natural language explanations for their predictions.
2 code implementations • 4 Oct 2019 • Oana-Maria Camburu, Eleonora Giunchiglia, Jakob Foerster, Thomas Lukasiewicz, Phil Blunsom
We aim for this framework to provide a publicly available, off-the-shelf evaluation when the feature-selection perspective on explanations is needed.
no code implementations • TACL 2020 • Lei Yu, Laurent Sartran, Wojciech Stokowiec, Wang Ling, Lingpeng Kong, Phil Blunsom, Chris Dyer
We show that Bayes' rule provides an effective mechanism for creating document translation models that can be learned from only parallel sentences and monolingual documents---a compelling benefit as parallel documents are not always available.
no code implementations • 25 Sep 2019 • Kazuya Kawakami, Luyu Wang, Chris Dyer, Phil Blunsom, Aaron van den Oord
We present an unsupervised method for learning speech representations based on a bidirectional contrastive predictive coding that implicitly discovers phonetic structure from large-scale corpora of unlabelled raw audio signals.
no code implementations • 25 Sep 2019 • Lei Yu, Laurent Sartran, Wojciech Stokowiec, Wang Ling, Lingpeng Kong, Phil Blunsom, Chris Dyer
We show that Bayes' rule provides a compelling mechanism for controlling unconditional document language models, using the long-standing challenge of effectively leveraging document context in machine translation.
1 code implementation • 20 Sep 2019 • Chris Dyer, Gábor Melis, Phil Blunsom
A series of recent papers has used a parsing algorithm due to Shen et al. (2018) to recover phrase-structure trees based on proxies for "syntactic depth."
3 code implementations • ICLR 2020 • Gábor Melis, Tomáš Kočiský, Phil Blunsom
Many advances in Natural Language Processing have been based upon more expressive models for how inputs interact with the context in which they occur.
1 code implementation • IJCNLP 2019 • Vid Kocijan, Oana-Maria Camburu, Ana-Maria Cretu, Yordan Yordanov, Phil Blunsom, Thomas Lukasiewicz
We use a language-model-based approach for pronoun resolution in combination with our WikiCREM dataset.
no code implementations • ACL 2019 • Adhiguna Kuncoro, Chris Dyer, Laura Rimell, Stephen Clark, Phil Blunsom
Prior work has shown that, on small amounts of training data, syntactic neural language models learn structurally sensitive generalisations more successfully than sequential language models.
no code implementations • 31 Jan 2019 • Dani Yogatama, Cyprien de Masson d'Autume, Jerome Connor, Tomas Kocisky, Mike Chrzanowski, Lingpeng Kong, Angeliki Lazaridou, Wang Ling, Lei Yu, Chris Dyer, Phil Blunsom
We define general linguistic intelligence as the ability to reuse previously acquired knowledge about a language's lexicon, syntax, semantics, and pragmatic conventions to adapt to new tasks quickly.
2 code implementations • NeurIPS 2018 • Oana-Maria Camburu, Tim Rocktäschel, Thomas Lukasiewicz, Phil Blunsom
In order for machine learning to garner widespread public adoption, models must be able to provide interpretable and robust explanations for their decisions, as well as learn from human-provided explanations at train time.
Ranked #1 on Natural Language Inference on e-SNLI
1 code implementation • 27 Nov 2018 • Linhai Xie, Yishu Miao, Sen Wang, Phil Blunsom, Zhihua Wang, Changhao Chen, Andrew Markham, Niki Trigoni
Due to the sparse rewards and high degree of environment variation, reinforcement learning approaches such as Deep Deterministic Policy Gradient (DDPG) are plagued by issues of high variance when applied in complex real world environments.
Robotics
no code implementations • 26 Nov 2018 • Lei Yu, Cyprien de Masson d'Autume, Chris Dyer, Phil Blunsom, Lingpeng Kong, Wang Ling
The meaning of a sentence is a function of the relations that hold between its words.
no code implementations • ACL 2019 • Kazuya Kawakami, Chris Dyer, Phil Blunsom
We propose a segmental neural language model that combines the generalization power of neural networks with the ability to discover word-like units that are latent in unsegmented character sequences.
no code implementations • 4 Oct 2018 • Changhao Chen, Yishu Miao, Chris Xiaoxuan Lu, Phil Blunsom, Andrew Markham, Niki Trigoni
Inertial information processing plays a pivotal role in ego-motion awareness for mobile agents, as inertial measurements are entirely egocentric and not environment dependent.
no code implementations • 27 Sep 2018 • Kazuya Kawakami, Chris Dyer, Phil Blunsom
We propose a segmental neural language model that combines the representational power of neural networks and the structure learning mechanism of Bayesian nonparametrics, and show that it learns to discover semantically meaningful units (e. g., morphemes and words) from unsegmented character sequences.
20 code implementations • NeurIPS 2018 • Andrew Trask, Felix Hill, Scott Reed, Jack Rae, Chris Dyer, Phil Blunsom
Neural networks can learn to represent and manipulate numerical information, but they seldom generalize well outside of the range of numerical values encountered during training.
1 code implementation • 4 Jul 2018 • Tiago Ramalho, Tomáš Kočiský, Frederic Besse, S. M. Ali Eslami, Gábor Melis, Fabio Viola, Phil Blunsom, Karl Moritz Hermann
Natural language processing has made significant inroads into learning the semantics of words through distributional approaches, however representations learnt via these methods fail to capture certain kinds of information implicit in the real world.
no code implementations • ACL 2018 • Adhiguna Kuncoro, Chris Dyer, John Hale, Dani Yogatama, Stephen Clark, Phil Blunsom
Language exhibits hierarchical structure, but recent work using a subject-verb agreement diagnostic argued that state-of-the-art language models, LSTMs, fail to learn long-range syntax sensitive dependencies.
no code implementations • NAACL 2018 • Jan Buys, Phil Blunsom
We present neural syntactic generative models with exact marginalization that support both dependency parsing and language modeling.
1 code implementation • ICLR 2019 • Gábor Melis, Charles Blundell, Tomáš Kočiský, Karl Moritz Hermann, Chris Dyer, Phil Blunsom
We show that dropout training is best understood as performing MAP estimation concurrently for a family of conditional models whose objectives are themselves lower bounded by the original dropout objective.
Ranked #24 on Language Modelling on Penn Treebank (Word Level)
no code implementations • ICLR 2018 • Felix Hill, Karl Moritz Hermann, Phil Blunsom, Stephen Clark
Neural network-based systems can now learn to locate the referents of words and phrases in images, answer questions about visual scenes, and even execute symbolic instructions as first-person actors in partially-observable worlds.
no code implementations • ICLR 2018 • Dani Yogatama, Yishu Miao, Gabor Melis, Wang Ling, Adhiguna Kuncoro, Chris Dyer, Phil Blunsom
We compare and analyze sequential, random access, and stack memory architectures for recurrent neural network language models.
2 code implementations • TACL 2018 • Tomáš Kočiský, Jonathan Schwarz, Phil Blunsom, Chris Dyer, Karl Moritz Hermann, Gábor Melis, Edward Grefenstette
Reading comprehension (RC)---in contrast to information retrieval---requires integrating information and reasoning about events, entities, and their relations across a full document.
Ranked #9 on Question Answering on NarrativeQA (BLEU-1 metric)
no code implementations • ICLR 2018 • Felix Hill, Stephen Clark, Karl Moritz Hermann, Phil Blunsom
Neural network-based systems can now learn to locate the referents of words and phrases in images, answer questions about visual scenes, and execute symbolic instructions as first-person actors in partially-observable worlds.
no code implementations • SEMEVAL 2017 • Jan Buys, Phil Blunsom
We present a neural encoder-decoder AMR parser that extends an attention-based model by predicting the alignment between graph nodes and sentence tokens explicitly with a pointer mechanism.
Ranked #27 on AMR Parsing on LDC2017T10
1 code implementation • ICLR 2018 • Gábor Melis, Chris Dyer, Phil Blunsom
Ongoing innovations in recurrent neural network architectures have provided a steady influx of apparently state-of-the-art results on language modelling benchmarks.
Ranked #32 on Language Modelling on WikiText-2
no code implementations • ACL 2017 • Wang Ling, Dani Yogatama, Chris Dyer, Phil Blunsom
Solving algebraic word problems requires executing a series of arithmetic operations{---}a program{---}to obtain a final answer.
1 code implementation • 20 Jun 2017 • Karl Moritz Hermann, Felix Hill, Simon Green, Fumin Wang, Ryan Faulkner, Hubert Soyer, David Szepesvari, Wojciech Marian Czarnecki, Max Jaderberg, Denis Teplyashin, Marcus Wainwright, Chris Apps, Demis Hassabis, Phil Blunsom
Trained via a combination of reinforcement and unsupervised learning, and beginning with minimal prior knowledge, the agent learns to relate linguistic symbols to emergent perceptual representations of its physical surroundings and to pertinent sequences of actions.
1 code implementation • ICML 2017 • Yishu Miao, Edward Grefenstette, Phil Blunsom
Topic models have been widely explored as probabilistic generative models of documents.
Ranked #2 on Topic Models on 20NewsGroups
1 code implementation • ICML 2017 • Tsung-Hsien Wen, Yishu Miao, Phil Blunsom, Steve Young
Developing a dialogue agent that is capable of making autonomous decisions and communicating by natural language is one of the long-term goals of machine learning research.
1 code implementation • 11 May 2017 • Wang Ling, Dani Yogatama, Chris Dyer, Phil Blunsom
Solving algebraic word problems requires executing a series of arithmetic operations---a program---to obtain a final answer.
1 code implementation • ACL 2017 • Jan Buys, Phil Blunsom
Parsing sentences to linguistically-expressive semantic representations is a key goal of Natural Language Processing.
no code implementations • ACL 2017 • Kazuya Kawakami, Chris Dyer, Phil Blunsom
Fixed-vocabulary language models fail to account for one of the most characteristic statistical facts of natural language: the frequent creation and reuse of new word types.
2 code implementations • 6 Mar 2017 • Dani Yogatama, Chris Dyer, Wang Ling, Phil Blunsom
We empirically characterize the performance of discriminative and generative LSTM models for text classification.
no code implementations • 28 Nov 2016 • Dani Yogatama, Phil Blunsom, Chris Dyer, Edward Grefenstette, Wang Ling
We use reinforcement learning to learn tree-structured neural networks for computing representations of natural language sentences.
no code implementations • 8 Nov 2016 • Lei Yu, Phil Blunsom, Chris Dyer, Edward Grefenstette, Tomas Kocisky
We formulate sequence to sequence transduction as a noisy channel decoding problem and use recurrent neural networks to parameterise the source and channel models.
no code implementations • EMNLP 2017 • Zichao Yang, Phil Blunsom, Chris Dyer, Wang Ling
We propose a general class of language models that treat reference as an explicit stochastic latent variable.
Ranked #1 on Recipe Generation on allrecipes.com
no code implementations • EMNLP 2016 • Tomáš Kočiský, Gábor Melis, Edward Grefenstette, Chris Dyer, Wang Ling, Phil Blunsom, Karl Moritz Hermann
We present a novel semi-supervised approach for sequence transduction and apply it to semantic parsing.
no code implementations • EMNLP 2016 • Lei Yu, Jan Buys, Phil Blunsom
We introduce an online neural sequence to sequence model that learns to alternate between encoding and decoding segments of the input as it is read.
no code implementations • EMNLP 2016 • Yishu Miao, Phil Blunsom
In this work we explore deep generative models of text in which the latent representation of a document is itself drawn from a discrete language model distribution.
1 code implementation • 7 Apr 2016 • Jeremy Appleyard, Tomas Kocisky, Phil Blunsom
As recurrent neural networks become larger and deeper, training times for single networks are rising into weeks or even months.
2 code implementations • ACL 2016 • Wang Ling, Edward Grefenstette, Karl Moritz Hermann, Tomáš Kočiský, Andrew Senior, Fumin Wang, Phil Blunsom
Many language generation tasks require the production of text conditioned on both structured and unstructured inputs.
Ranked #10 on Code Generation on Django
no code implementations • 5 Dec 2015 • Pengyu Wang, Phil Blunsom
Stochastic variational inference for collapsed models has recently been successfully applied to large scale topic modelling.
no code implementations • 5 Dec 2015 • Pengyu Wang, Phil Blunsom
In this paper, we propose a stochastic collapsed variational inference algorithm for hidden Markov models, in a sequential data setting.
6 code implementations • 19 Nov 2015 • Yishu Miao, Lei Yu, Phil Blunsom
We validate this framework on two very different text modelling applications, generative document modelling and supervised question answering.
Ranked #1 on Question Answering on QASent
7 code implementations • 22 Sep 2015 • Tim Rocktäschel, Edward Grefenstette, Karl Moritz Hermann, Tomáš Kočiský, Phil Blunsom
We extend this model with a word-by-word neural attention mechanism that encourages reasoning over entailments of pairs of words and phrases.
Ranked #81 on Natural Language Inference on SNLI
no code implementations • WS 2015 • Jan Buys, Phil Blunsom
We propose a simple, scalable, fully generative model for transition-based dependency parsing with high accuracy.
11 code implementations • NeurIPS 2015 • Karl Moritz Hermann, Tomáš Kočiský, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, Phil Blunsom
Teaching machines to read natural language documents remains an elusive challenge.
Ranked #13 on Question Answering on CNN / Daily Mail
4 code implementations • NeurIPS 2015 • Edward Grefenstette, Karl Moritz Hermann, Mustafa Suleyman, Phil Blunsom
Recently, strong results have been demonstrated by Deep Recurrent Neural Networks on natural language transduction problems.
no code implementations • HLT 2015 • Paul Baltescu, Phil Blunsom
This paper presents an in-depth investigation on integrating neural language models in translation systems.
no code implementations • 22 Dec 2014 • Yishu Miao, Ziyu Wang, Phil Blunsom
This paper presents novel Bayesian optimisation algorithms for minimum error rate training of statistical machine translation systems.
2 code implementations • 4 Dec 2014 • Lei Yu, Karl Moritz Hermann, Phil Blunsom, Stephen Pulman
Answer sentence selection is the task of identifying sentences that contain the answer to a given question.
Ranked #3 on Question Answering on QASent
no code implementations • 12 Nov 2014 • Dimitrios Kotzias, Misha Denil, Phil Blunsom, Nando de Freitas
We present a new approach for transferring knowledge from groups to individuals that comprise them.
no code implementations • 15 Jun 2014 • Misha Denil, Alban Demiraj, Nal Kalchbrenner, Phil Blunsom, Nando de Freitas
Capturing the compositional process which maps the meaning of words to that of documents is a central challenge for researchers in Natural Language Processing and Information Retrieval.
1 code implementation • 16 May 2014 • Jan A. Botha, Phil Blunsom
This paper presents a scalable method for integrating compositional morphological representations into a vector-based probabilistic language model.
no code implementations • ACL 2014 • Tomáš Kočiský, Karl Moritz Hermann, Phil Blunsom
We present a probabilistic model that simultaneously learns alignments and distributed representations for bilingual data.
no code implementations • WS 2014 • Edward Grefenstette, Phil Blunsom, Nando de Freitas, Karl Moritz Hermann
Many successful approaches to semantic parsing build on top of the syntactic analysis of text, and make use of distributional representations or statistical models to match parses to ontology-specific queries.
1 code implementation • ACL 2014 • Karl Moritz Hermann, Phil Blunsom
We present a novel technique for learning semantic representations, which extends the distributional hypothesis to multilingual data and joint-space embeddings.
Cross-Lingual Document Classification Document Classification +2
5 code implementations • ACL 2014 • Nal Kalchbrenner, Edward Grefenstette, Phil Blunsom
The ability to accurately represent sentences is central to language understanding.
no code implementations • EACL 2014 • Greg Dubbin, Phil Blunsom
Automatically inducing the syntactic part-of-speech categories for words in text is a fundamental task in Computational Linguistics.
1 code implementation • 20 Dec 2013 • Karl Moritz Hermann, Phil Blunsom
Distributed representations of meaning are a natural way to encode covariance relationships between words and phrases in NLP.
Cross-Lingual Document Classification Document Classification +3
no code implementations • WS 2013 • Nal Kalchbrenner, Phil Blunsom
The compositionality of meaning extends beyond the single sentence.
no code implementations • 10 Jun 2013 • Karl Moritz Hermann, Edward Grefenstette, Phil Blunsom
With the increasing empirical success of distributional models of compositional semantics, it is timely to consider the types of textual logic that such models are capable of capturing.
no code implementations • 1 Oct 2010 • Phil Blunsom, Trevor Cohn
Inducing a grammar directly from text is one of the oldest and most challenging tasks in Computational Linguistics.
Ranked #3 on Unsupervised Dependency Parsing on Penn Treebank
Dependency Grammar Induction Unsupervised Dependency Parsing
no code implementations • NeurIPS 2008 • Phil Blunsom, Trevor Cohn, Miles Osborne
We present a novel method for inducing synchronous context free grammars (SCFGs) from a corpus of parallel string pairs.