no code implementations • ACL (IWSLT) 2021 • Biao Zhang, Rico Sennrich
This paper describes Edinburgh’s submissions to the IWSLT2021 multilingual speech translation (ST) task.
no code implementations • WMT (EMNLP) 2020 • Rachel Bawden, Biao Zhang, Andre Tättar, Matt Post
We describe parBLEU, parCHRF++, and parESIM, which augment baseline metrics with automatically generated paraphrases produced by PRISM (Thompson and Post, 2020a), a multilingual neural machine translation system.
no code implementations • 9 Sep 2023 • Sneha Kudugunta, Isaac Caswell, Biao Zhang, Xavier Garcia, Christopher A. Choquette-Choo, Katherine Lee, Derrick Xin, Aditya Kusupati, Romi Stella, Ankur Bapna, Orhan Firat
We introduce MADLAD-400, a manually audited, general domain 3T token monolingual dataset based on CommonCrawl, spanning 419 languages.
1 code implementation • 29 May 2023 • Qian Wang, Biao Zhang, Michael Birsak, Peter Wonka
In this work, we propose a framework termed InstructEdit that can do fine-grained editing based on user instructions.
no code implementations • 23 May 2023 • Christos Baziotis, Biao Zhang, Alexandra Birch, Barry Haddow
Next, we analyze the impact of scale (from 90M to 1. 6B parameters) and find it is important for both methods, particularly DAE.
1 code implementation • International Conference on Learning Representations (ICLR) 2023 • Biao Zhang, Mathias Müller, Rico Sennrich
We propose SLTUNET, a simple unified neural model designed to support multiple SLTrelated tasks jointly, such as sign-to-gloss, gloss-to-text and sign-to-text translation.
2 code implementations • 29 Mar 2023 • Qian Wang, Biao Zhang, Michael Birsak, Peter Wonka
Image generation using diffusion can be controlled in multiple ways.
1 code implementation • 21 Feb 2023 • Biao Zhang, Barry Haddow, Rico Sennrich
For end-to-end speech translation, regularizing the encoder with the Connectionist Temporal Classification (CTC) objective using the source transcript or target translation as labels can greatly improve quality metrics.
no code implementations • 26 Jan 2023 • Biao Zhang, Jiapeng Tang, Matthias Niessner, Peter Wonka
We introduce 3DShape2VecSet, a novel shape representation for neural fields designed for generative diffusion models.
no code implementations • 17 Jan 2023 • Biao Zhang, Barry Haddow, Alexandra Birch
Research on prompting has shown excellent performance with little or even no supervised training across many tasks.
no code implementations • 24 Oct 2022 • Biao Zhang, Peng Xiao, Shuqin Zhang
Fusing deep learning models trained on separately located clients into a global model in a one-shot communication round is a straightforward implementation of Federated Learning.
no code implementations • 16 Sep 2022 • Zhuoran Liu, Leqi Zou, Xuan Zou, Caihua Wang, Biao Zhang, Da Tang, Bolin Zhu, Yijie Zhu, Peng Wu, Ke Wang, Youlong Cheng
In this paper, we present Monolith, a system tailored for online training.
1 code implementation • 9 Jun 2022 • Biao Zhang, Barry Haddow, Rico Sennrich
Finally, we discuss neural acoustic feature modeling, where a neural model is designed to extract acoustic features from raw speech signals directly, with the goal to simplify inductive biases and add freedom to the model in describing speech.
1 code implementation • 27 May 2022 • Biao Zhang, Matthias Nießner, Peter Wonka
All probabilistic experiments confirm that we are able to generate detailed and high quality shapes to yield the new state of the art in generative 3D shape modeling.
no code implementations • 4 Feb 2022 • Yamini Bansal, Behrooz Ghorbani, Ankush Garg, Biao Zhang, Maxim Krikun, Colin Cherry, Behnam Neyshabur, Orhan Firat
In this work, we study the effect of varying the architecture and training data quality on the data scaling properties of Neural Machine Translation (NMT).
no code implementations • 1 Feb 2022 • Biao Zhang, Behrooz Ghorbani, Ankur Bapna, Yong Cheng, Xavier Garcia, Jonathan Shen, Orhan Firat
Natural language understanding and generation models follow one of the two dominant architectural paradigms: language models (LMs) that process concatenated sequences in a single stack of layers, and encoder-decoder models (EncDec) that utilize separate layer stacks for input and output processing.
no code implementations • ACL 2022 • Biao Zhang, Ankur Bapna, Melvin Johnson, Ali Dabirmoghaddam, Naveen Arivazhagan, Orhan Firat
Using simple concatenation-based DocNMT, we explore the effect of 3 factors on the transfer: the number of teacher languages with document level data, the balance between document and sentence level data at training, and the data condition of parallel documents (genuine vs. backtranslated).
1 code implementation • ACL 2021 • Biao Zhang, Ivan Titov, Barry Haddow, Rico Sennrich
Document-level contextual information has shown benefits to text-based machine translation, but whether and how context helps end-to-end (E2E) speech translation (ST) is still under-studied.
1 code implementation • ACL 2021 • Hui Jiang, Chulun Zhou, Fandong Meng, Biao Zhang, Jie zhou, Degen Huang, Qingqiang Wu, Jinsong Su
Due to the great potential in facilitating software development, code generation has attracted increasing attention recently.
2 code implementations • EMNLP 2021 • Biao Zhang, Ivan Titov, Rico Sennrich
Recently, it has been argued that encoder-decoder models can be made more interpretable by replacing the softmax function in the attention with its sparse variants.
no code implementations • ICLR 2021 • Biao Zhang, Ankur Bapna, Rico Sennrich, Orhan Firat
Our study further verifies the trade-off between the shared capacity and LS capacity for multilingual translation.
1 code implementation • WMT (EMNLP) 2020 • Biao Zhang, Ivan Titov, Rico Sennrich
Instead of assuming independence between neighbouring tokens (semi-autoregressive decoding, SA), we take inspiration from bidirectional sequence generation and introduce a decoder that generates target words from the left-to-right and right-to-left directions simultaneously.
no code implementations • ICLR 2022 • Biao Zhang, Peter Wonka
We combine training data generating networks with bi-level optimization algorithms to obtain a complete framework for which all components can be jointly trained.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Biao Zhang, Ivan Titov, Barry Haddow, Rico Sennrich
Information in speech signals is not evenly distributed, making it an additional challenge for end-to-end (E2E) speech translation (ST) to learn to focus on informative features.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Rachel Bawden, Biao Zhang, Lisa Yankovskaya, Andre Tättar, Matt Post
We investigate a long-perceived shortcoming in the typical use of BLEU: its reliance on a single reference.
3 code implementations • ACL 2020 • Biao Zhang, Philip Williams, Ivan Titov, Rico Sennrich
Massively multilingual models for neural machine translation (NMT) are theoretically attractive, but often underperform bilingual models and deliver poor zero-shot translations.
1 code implementation • Findings (ACL) 2021 • Biao Zhang, Ivan Titov, Rico Sennrich
Inspired by these observations, we explore the feasibility of specifying rule-based patterns that mask out encoder outputs based on information such as part-of-speech tags, word frequency and word position.
no code implementations • CVPR 2021 • Biao Zhang, Peter Wonka
In this paper we propose a new framework for point cloud instance segmentation.
Ranked #1 on
Instance Segmentation
on PartNet
2 code implementations • NeurIPS 2019 • Biao Zhang, Rico Sennrich
RMSNorm regularizes the summed inputs to a neuron in one layer according to root mean square (RMS), giving the model re-scaling invariance property and implicit learning rate adaptation ability.
1 code implementation • IJCNLP 2019 • Biao Zhang, Ivan Titov, Rico Sennrich
The general trend in NLP is towards increasing model capacity and performance via deeper neural networks.
1 code implementation • ACL 2019 • Biao Zhang, Rico Sennrich
We apply LRN as a drop-in replacement of existing recurrent units in several neural sequential models.
2 code implementations • ACL 2019 • Rico Sennrich, Biao Zhang
It has been shown that the performance of neural machine translation (NMT) drops starkly in low-resource conditions, underperforming phrase-based statistical machine translation (PBSMT) and requiring large amounts of auxiliary data to achieve competitive results.
3 code implementations • EMNLP 2018 • Biao Zhang, Deyi Xiong, Jinsong Su, Qian Lin, Huiji Zhang
Experiments on WMT14 translation tasks demonstrate that ATR-based neural machine translation can yield competitive performance on English- German and English-French language pairs in terms of both translation quality and speed.
1 code implementation • 24 Jul 2018 • Jing Yang, Biao Zhang, Yue Qin, Xiangwen Zhang, Qian Lin, Jinsong Su
Although neural machine translation(NMT) yields promising translation performance, it unfortunately suffers from over- and under-translation is- sues [Tu et al., 2016], of which studies have become research hotspots in NMT.
1 code implementation • ACL 2018 • Biao Zhang, Deyi Xiong, Jinsong Su
To alleviate this issue, we propose an average attention network as an alternative to the self-attention network in the decoder of the neural Transformer.
Ranked #59 on
Machine Translation
on WMT2014 English-German
no code implementations • E2E NLG Challenge System Descriptions 2018 • Biao Zhang, Jing Yang, Qian Lin, Jinsong Su
This paper describes our system used for the end-to-end (E2E) natural language generation (NLG) challenge.
Ranked #7 on
Data-to-Text Generation
on E2E NLG Challenge
no code implementations • 16 Jan 2018 • Jinsong Su, Shan Wu, Deyi Xiong, Yaojie Lu, Xianpei Han, Biao Zhang
Partially inspired by successful applications of variational recurrent neural networks, we propose a novel variational recurrent neural machine translation (VRNMT) model in this paper.
no code implementations • 27 Apr 2017 • Biao Zhang, Deyi Xiong, Jinsong Su
In this paper, we propose a novel GRU-gated attention model (GAtt) for NMT which enhances the degree of discrimination of context vectors by enabling source representations to be sensitive to the partial translation generated by the decoder.
no code implementations • COLING 2016 • Jinsong Su, Biao Zhang, Deyi Xiong, Ruochen Li, Jianmin Yin
After that, we fully incorporate information of different linguistic units into a bilinear semantic similarity model.
no code implementations • COLING 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, Min Zhang
Parallel sentence representations are important for bilingual and cross-lingual tasks in natural language processing.
no code implementations • 29 Jul 2016 • Biao Zhang, Deyi Xiong, Jinsong Su
The vanilla sequence-to-sequence learning (seq2seq) reads and encodes a source sequence into a fixed-length vector only once, suffering from its insufficiency in modeling structural correspondence between the source and target sequence.
1 code implementation • 25 May 2016 • Biao Zhang, Deyi Xiong, Jinsong Su
In this paper, we propose a bidimensional attention based recursive autoencoder (BattRAE) to integrate clues and sourcetarget interactions at multiple levels of granularity into bilingual phrase representations.
1 code implementation • EMNLP 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, Min Zhang
Models of neural machine translation are often from a discriminative family of encoderdecoders that learn a conditional distribution of a target sentence given a source sentence.
no code implementations • 12 Mar 2016 • Biao Zhang, Deyi Xiong, Jinsong Su
Inspired by this, we propose a neural recognizer for implicit discourse relation analysis, which builds upon a semantic memory that stores knowledge in a distributed fashion.
1 code implementation • EMNLP 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Qun Liu, Rongrong Ji, Hong Duan, Min Zhang
In order to perform efficient inference and learning, we introduce neural discourse relation models to approximate the prior and posterior distributions of the latent variable, and employ these approximated distributions to optimize a reparameterized variational lower bound.
no code implementations • CVPR 2014 • Yi Peng, Deyu Meng, Zongben Xu, Chenqiang Gao, Yi Yang, Biao Zhang
As compared to the conventional RGB or gray-scale images, multispectral images (MSI) can deliver more faithful representation for real scenes, and enhance the performance of many computer vision tasks.