1 code implementation • 26 May 2023 • Shiyue Zhang, Shijie Wu, Ozan Irsoy, Steven Lu, Mohit Bansal, Mark Dredze, David Rosenberg
Autoregressive language models are trained by minimizing the cross-entropy of the model distribution Q relative to the data distribution P -- that is, minimizing the forward cross-entropy, which is equivalent to maximum likelihood estimation (MLE).
1 code implementation • 8 May 2023 • David Wan, Shiyue Zhang, Mohit Bansal
Cache-LMs, which augment LMs with a memory of recent history, can increase context dependency and have shown remarkable performance in diverse language generation tasks.
1 code implementation • 15 Nov 2022 • Derek Tam, Anisha Mascarenhas, Shiyue Zhang, Sarah Kwan, Mohit Bansal, Colin Raffel
To generate summaries that are factually inconsistent, we generate summaries from a suite of summarization models that we have manually annotated as factually inconsistent.
1 code implementation • 21 Sep 2022 • Swarnadeep Saha, Shiyue Zhang, Peter Hase, Mohit Bansal
We demonstrate that SP-Search effectively represents the generative process behind human summaries using modules that are typically faithful to their intended behavior.
1 code implementation • 8 Sep 2022 • Shiyue Zhang, David Wan, Mohit Bansal
Though extractive summarization is less prone to the common unfaithfulness issues of abstractive summaries, does that mean extractive is equal to faithful?
1 code implementation • NAACL (ACL) 2022 • Yinuo Hu, Shiyue Zhang, Viji Sathy, A. T. Panter, Mohit Bansal
Ten university professors from diverse departments serve as evaluators of the system and all agree that SETSum helps them interpret SET results more efficiently; and 6 out of 10 instructors prefer our system over the standard static PDF report (while the remaining 4 would like to have both).
1 code implementation • NAACL 2022 • Xiang Zhou, Shiyue Zhang, Mohit Bansal
MPoSM can model arbitrary tag dependency and perform POS induction through the objective of masked POS reconstruction.
no code implementations • AMTA 2022 • Shiyue Zhang, Vishrav Chaudhary, Naman Goyal, James Cross, Guillaume Wenzek, Mohit Bansal, Francisco Guzman
Since a skewed data distribution is considered to be harmful, a sampling strategy is usually used to balance languages in the corpus.
1 code implementation • ACL 2022 • Shiyue Zhang, Ben Frey, Mohit Bansal
We hope that our work serves not only to inform the NLP community about Cherokee, but also to provide inspiration for future work on endangered languages in general.
1 code implementation • EMNLP 2021 • Shiyue Zhang, Mohit Bansal
In this work, we propose flexible semiautomatic to automatic summary evaluation metrics, following the Pyramid human evaluation method.
1 code implementation • ACL 2021 • Zineng Tang, Shiyue Zhang, Hyounghun Kim, Mohit Bansal
Recent years have witnessed various types of generative models for natural language generation (NLG), especially RNNs or transformer based sequence-to-sequence models, as well as variational autoencoder (VAE) and generative adversarial network (GAN) based models.
1 code implementation • ACL 2021 • Shiyue Zhang, Asli Celikyilmaz, Jianfeng Gao, Mohit Bansal
Furthermore, we find that widely used automatic evaluation metrics (ROUGE, BERTScore) are weakly correlated with human judgments on this email thread summarization task.
Ranked #1 on
Email Thread Summarization
on EmailSum (short)
2 code implementations • ACL 2021 • Shiyue Zhang, Benjamin Frey, Mohit Bansal
The quantitative evaluation demonstrates that our backbone translation models achieve state-of-the-art translation performance and our quality estimation well correlates with both BLEU and human judgment.
1 code implementation • EMNLP 2020 • Shiyue Zhang, Benjamin Frey, Mohit Bansal
To help save this endangered language, we introduce ChrEn, a Cherokee-English parallel dataset, to facilitate machine translation research between Cherokee and English.
Cultural Vocal Bursts Intensity Prediction
Language Modelling
+4
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Peter Hase, Shiyue Zhang, Harry Xie, Mohit Bansal
We provide code for the experiments in this paper at https://github. com/peterbhase/LAS-NL-Explanations
1 code implementation • IJCNLP 2019 • Shiyue Zhang, Mohit Bansal
Second, since the traditional evaluation metrics (e. g., BLEU) often fall short in evaluating the quality of generated questions, we propose a QA-based evaluation method which measures the QG model's ability to mimic human annotators in generating QA training data.
no code implementations • 12 Nov 2017 • Shiyue Zhang, Pengtao Xie, Dong Wang, Eric P. Xing
In hospital, physicians rely on massive clinical data to make diagnosis decisions, among which laboratory tests are one of the most important resources.
2 code implementations • 4 Oct 2017 • Aodong Li, Shiyue Zhang, Dong Wang, Thomas Fang Zheng
Neural machine translation (NMT) has recently achieved impressive results.
no code implementations • EMNLP 2017 • Yang Feng, Shiyue Zhang, Andi Zhang, Dong Wang, Andrew Abel
Neural machine translation (NMT) has achieved notable success in recent times, however it is also widely recognized that this approach has limitations with handling infrequent words and word pairs.
no code implementations • 27 Jun 2017 • Shiyue Zhang, Gulnigar Mahmut, Dong Wang, Askar Hamdulla
Neural machine translation (NMT) has achieved notable performance recently.
no code implementations • ACL 2017 • Jiyuan Zhang, Yang Feng, Dong Wang, Yang Wang, Andrew Abel, Shiyue Zhang, Andi Zhang
It has been shown that Chinese poems can be successfully generated by sequence-to-sequence neural models, particularly with the attention mechanism.
no code implementations • 28 Sep 2016 • Zhiyuan Tang, Ying Shi, Dong Wang, Yang Feng, Shiyue Zhang
Recurrent neural networks (RNNs) have shown clear superiority in sequence modeling, particularly the ones with gated units, such as long short-term memory (LSTM) and gated recurrent unit (GRU).
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 27 Sep 2016 • Lantian Li, Zhiyuan Tang, Dong Wang, Andrew Abel, Yang Feng, Shiyue Zhang
This paper presents a unified model to perform language and speaker recognition simultaneously and altogether.