3 code implementations • ACL 2021 • Zijun Sun, Xiaoya Li, Xiaofei Sun, Yuxian Meng, Xiang Ao, Qing He, Fei Wu, Jiwei Li
Recent pretraining models in Chinese neglect two important aspects specific to the Chinese language: glyph and pinyin, which carry significant syntax and semantic information for language understanding.
1 code implementation • 3 Dec 2020 • Zijun Sun, Chun Fan, Qinghong Han, Xiaofei Sun, Yuxian Meng, Fei Wu, Jiwei Li
The proposed model comes with the following merits: (1) span weights make the model self-explainable and do not require an additional probing model for interpretation; (2) the proposed model is general and can be adapted to any existing deep learning structures in NLP; (3) the weight associated with each text span provides direct importance scores for higher-level text units such as phrases and sentences.
1 code implementation • 17 Nov 2020 • Zijun Sun, Chun Fan, Xiaofei Sun, Yuxian Meng, Fei Wu, Jiwei Li
The goal of semi-supervised learning is to utilize the unlabeled, in-domain dataset U to improve models trained on the labeled dataset D. Under the context of large-scale language-model (LM) pretraining, how we can make the best use of U is poorly understood: is semi-supervised learning still beneficial with the presence of large-scale pretraining?
Ranked #1000000000 on Text Classification on IMDb
no code implementations • 14 Oct 2020 • Yuxian Meng, Chun Fan, Zijun Sun, Eduard Hovy, Fei Wu, Jiwei Li
Any prediction from a model is made by a combination of learning history and test stimuli.
no code implementations • COLING 2022 • Xiaofei Sun, Zijun Sun, Yuxian Meng, Jiwei Li, Chun Fan
The difficulty of generating coherent long texts lies in the fact that existing models overwhelmingly focus on predicting local words, and cannot make high level plans on what to generate or capture the high-level discourse dependencies between chunks of texts.
no code implementations • 26 Sep 2019 • Yuxian Meng, Xiangyuan Ren, Zijun Sun, Xiaoya Li, Arianna Yuan, Fei Wu, Jiwei Li
In this paper, we investigate the problem of training neural machine translation (NMT) systems with a dataset of more than 40 billion bilingual sentence pairs, which is larger than the largest dataset to date by orders of magnitude.
no code implementations • 24 Aug 2019 • Yuxian Meng, Xiaoya Li, Zijun Sun, Jiwei Li
In this paper, we propose a new strategy for the task of named entity recognition (NER).
Entity Extraction using GAN Machine Reading Comprehension +3
1 code implementation • ACL 2019 • Xiaoya Li, Fan Yin, Zijun Sun, Xiayu Li, Arianna Yuan, Duo Chai, Mingxin Zhou, Jiwei Li
In this paper, we propose a new paradigm for the task of entity-relation extraction.
Ranked #1 on Relation Extraction on ACE 2005 (Sentence Encoder metric)
no code implementations • 8 Mar 2018 • Shuqing Bian, Zhenpeng Deng, Fei Li, Will Monroe, Peng Shi, Zijun Sun, Wei Wu, Sikuang Wang, William Yang Wang, Arianna Yuan, Tianwei Zhang, Jiwei Li
For the best setting, the proposed system is able to identify scam ICO projects with 0. 83 precision.