Search Results for author: Zijun Sun

Found 9 papers, 4 papers with code

ChineseBERT: Chinese Pretraining Enhanced by Glyph and Pinyin Information

3 code implementations ACL 2021 Zijun Sun, Xiaoya Li, Xiaofei Sun, Yuxian Meng, Xiang Ao, Qing He, Fei Wu, Jiwei Li

Recent pretraining models in Chinese neglect two important aspects specific to the Chinese language: glyph and pinyin, which carry significant syntax and semantic information for language understanding.

Language Modelling Machine Reading Comprehension +5

Self-Explaining Structures Improve NLP Models

1 code implementation3 Dec 2020 Zijun Sun, Chun Fan, Qinghong Han, Xiaofei Sun, Yuxian Meng, Fei Wu, Jiwei Li

The proposed model comes with the following merits: (1) span weights make the model self-explainable and do not require an additional probing model for interpretation; (2) the proposed model is general and can be adapted to any existing deep learning structures in NLP; (3) the weight associated with each text span provides direct importance scores for higher-level text units such as phrases and sentences.

Natural Language Inference Paraphrase Identification +1

Neural Semi-supervised Learning for Text Classification Under Large-Scale Pretraining

1 code implementation17 Nov 2020 Zijun Sun, Chun Fan, Xiaofei Sun, Yuxian Meng, Fei Wu, Jiwei Li

The goal of semi-supervised learning is to utilize the unlabeled, in-domain dataset U to improve models trained on the labeled dataset D. Under the context of large-scale language-model (LM) pretraining, how we can make the best use of U is poorly understood: is semi-supervised learning still beneficial with the presence of large-scale pretraining?

Ranked #1000000000 on Text Classification on IMDb

General Classification Language Modelling +3

Summarize, Outline, and Elaborate: Long-Text Generation via Hierarchical Supervision from Extractive Summaries

no code implementations COLING 2022 Xiaofei Sun, Zijun Sun, Yuxian Meng, Jiwei Li, Chun Fan

The difficulty of generating coherent long texts lies in the fact that existing models overwhelmingly focus on predicting local words, and cannot make high level plans on what to generate or capture the high-level discourse dependencies between chunks of texts.

Text Generation

Pair the Dots: Jointly Examining Training History and Test Stimuli for Model Interpretability

no code implementations14 Oct 2020 Yuxian Meng, Chun Fan, Zijun Sun, Eduard Hovy, Fei Wu, Jiwei Li

Any prediction from a model is made by a combination of learning history and test stimuli.

Large-scale Pretraining for Neural Machine Translation with Tens of Billions of Sentence Pairs

no code implementations26 Sep 2019 Yuxian Meng, Xiangyuan Ren, Zijun Sun, Xiaoya Li, Arianna Yuan, Fei Wu, Jiwei Li

In this paper, we investigate the problem of training neural machine translation (NMT) systems with a dataset of more than 40 billion bilingual sentence pairs, which is larger than the largest dataset to date by orders of magnitude.

Machine Translation NMT +2

Cannot find the paper you are looking for? You can Submit a new open access paper.