no code implementations • ACL 2022 • Yi Chen, Jiayang Cheng, Haiyun Jiang, Lemao Liu, Haisong Zhang, Shuming Shi, Ruifeng Xu
In this paper, we firstly empirically find that existing models struggle to handle hard mentions due to their insufficient contexts, which consequently limits their overall typing performance.
no code implementations • ACL 2022 • Yanling Xiao, Lemao Liu, Guoping Huang, Qu Cui, ShuJian Huang, Shuming Shi, Jiajun Chen
In this work, we propose a novel BiTIIMT system, Bilingual Text-Infilling for Interactive Neural Machine Translation.
1 code implementation • EMNLP 2021 • Jing Qian, Yibin Liu, Lemao Liu, Yangming Li, Haiyun Jiang, Haisong Zhang, Shuming Shi
Existing work on Fine-grained Entity Typing (FET) typically trains automatic models on the datasets obtained by using Knowledge Bases (KB) as distant supervision.
no code implementations • EMNLP 2021 • Yi Chen, Haiyun Jiang, Lemao Liu, Shuming Shi, Chuang Fan, Min Yang, Ruifeng Xu
Auxiliary information from multiple sources has been demonstrated to be effective in zero-shot fine-grained entity typing (ZFET).
no code implementations • Findings (EMNLP) 2021 • Yu Feng, Jing Zhang, Gaole He, Wayne Xin Zhao, Lemao Liu, Quan Liu, Cuiping Li, Hong Chen
Knowledge Base Question Answering (KBQA) is to answer natural language questions posed over knowledge bases (KBs).
no code implementations • Xintong Li, Lemao Liu, Guanlin Li, Max Meng, Shuming Shi
We find that although NMT models are difficult to capture word alignment for CFT words but these words do not sacrifice translation quality significantly, which provides an explanation why NMT is more successful for translation yet worse for word alignment compared to statistical machine translation.
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Yahui Liu, Lemao Liu, Guoping Huang, Defu Lian, Shuming Shi
Current practices in metric evaluation focus on one single dataset, e. g., Newstest dataset in each year's WMT Metrics Shared Task.
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Defu Lian, Guoping Huang, Taro Watanabe, Lemao Liu
To this end, we study the dynamic relationship between the encoded linguistic information and task performance from the viewpoint of Pareto Optimality.
1 code implementation • 9 Mar 2022 • Wenye Lin, Yangming Li, Lemao Liu, Shuming Shi, Hai-Tao Zheng
Specifically, we transfer the knowledge from a teacher model to its student model by locally matching their predictions on all sub-structures, instead of the whole output space.
no code implementations • 17 Feb 2022 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
(2) reference-free metrics outperform reference-based metrics, indicating that the standard references are unnecessary to evaluate the paraphrase's quality.
no code implementations • 2 Feb 2022 • Huayang Li, Yixuan Su, Deng Cai, Yan Wang, Lemao Liu
Recently, retrieval-augmented text generation attracted increasing attention of the computational linguistics community.
no code implementations • 9 Jan 2022 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
It has been shown that natural language processing (NLP) models are vulnerable to a kind of security threat called the Backdoor Attack, which utilizes a `backdoor trigger' paradigm to mislead the models.
1 code implementation • 12 Dec 2021 • Yu Feng, Jing Zhang, Xiaokang Zhang, Lemao Liu, Cuiping Li, Hong Chen
Embedding-based methods are popular for Knowledge Base Question Answering (KBQA), but few current models have numerical reasoning skills and thus struggle to answer ordinal constrained questions.
no code implementations • ACL 2022 • Yangming Li, Lemao Liu, Shuming Shi
Negative sampling is highly effective in handling missing annotations for named entity recognition (NER).
1 code implementation • ACL 2021 • Zexin Lu, Keyang Ding, Yuji Zhang, Jing Li, Baolin Peng, Lemao Liu
This paper presents a novel task to generate poll questions for social media posts.
no code implementations • ACL 2021 • Lemao Liu, Haisong Zhang, Haiyun Jiang, Yangming Li, Enbo Zhao, Kun Xu, Linfeng Song, Suncong Zheng, Botong Zhou, Dick Zhu, Xiao Feng, Tao Chen, Tao Yang, Dong Yu, Feng Zhang, Zhanhui Kang, Shuming Shi
This paper introduces TexSmart, a text understanding system that supports fine-grained named entity recognition (NER) and enhanced semantic analysis functionalities.
no code implementations • ACL 2021 • Qiuxiang He, Guoping Huang, Qu Cui, Li Li, Lemao Liu
It is generally believed that a translation memory (TM) should be beneficial for machine translation tasks.
no code implementations • ACL 2021 • Huayang Li, Lemao Liu, Guoping Huang, Shuming Shi
In this paper, we propose the task of general word-level autocompletion (GWLAN) from a real-world CAT scenario, and construct the first public benchmark to facilitate research in this topic.
no code implementations • 27 May 2021 • Guoping Huang, Lemao Liu, Xing Wang, Longyue Wang, Huayang Li, Zhaopeng Tu, Chengyan Huang, Shuming Shi
Automatic machine translation is super efficient to produce translations yet their quality is not guaranteed.
1 code implementation • ACL 2021 • Deng Cai, Yan Wang, Huayang Li, Wai Lam, Lemao Liu
Second, the memory retriever and NMT model can be jointly optimized for the ultimate translation goal.
1 code implementation • Findings (ACL) 2021 • Jiannan Xiang, Yahui Liu, Deng Cai, Huayang Li, Defu Lian, Lemao Liu
An important aspect of developing dialogue systems is how to evaluate and compare the performance of different systems.
1 code implementation • NAACL 2021 • Yangming Li, Lemao Liu, Kaisheng Yao
Prior methods to text segmentation are mostly at token level.
no code implementations • 1 Jan 2021 • Guanlin Li, Lemao Liu, Taro Watanabe, Conghui Zhu, Tiejun Zhao
Unsupervised Neural Machine Translation or UNMT has received great attention in recent years.
no code implementations • 31 Dec 2020 • Haisong Zhang, Lemao Liu, Haiyun Jiang, Yangming Li, Enbo Zhao, Kun Xu, Linfeng Song, Suncong Zheng, Botong Zhou, Jianchen Zhu, Xiao Feng, Tao Chen, Tao Yang, Dong Yu, Feng Zhang, Zhanhui Kang, Shuming Shi
This technique report introduces TexSmart, a text understanding system that supports fine-grained named entity recognition (NER) and enhanced semantic analysis functionalities.
no code implementations • Findings (EMNLP) 2021 • Yangming Li, Lemao Liu, Shuming Shi
In this work, we present Lexical Unit Analysis (LUA), a framework for general sequence segmentation tasks.
1 code implementation • ICLR 2021 • Yangming Li, Lemao Liu, Shuming Shi
Experiments on synthetic datasets and real-world datasets show that our model is robust to unlabeled entity problem and surpasses prior baselines.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Qian Wang, Jiajun Zhang, Lemao Liu, Guoping Huang, Chengqing Zong
We propose a touch-based editing method for translation, which is more flexible than traditional keyboard-mouse-based translation postediting.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Huayang Li, Lemao Liu, Guoping Huang, Shuming Shi
Many efforts have been devoted to extracting constituency trees from pre-trained language models, often proceeding in two stages: feature definition and parsing.
no code implementations • ACL 2020 • Jierui Li, Lemao Liu, Huayang Li, Guanlin Li, Guoping Huang, Shuming Shi
Recently many efforts have been devoted to interpreting the black-box NMT models, but little progress has been made on metrics to evaluate explanation methods.
no code implementations • 5 Apr 2020 • Guanlin Li, Lemao Liu, Conghui Zhu, Tiejun Zhao, Shuming Shi
Generalization to unseen instances is our eternal pursuit for all data-driven models.
no code implementations • 5 Apr 2020 • Conghui Zhu, Guanlin Li, Lemao Liu, Tiejun Zhao, Shuming Shi
Despite the great success of NMT, there still remains a severe challenge: it is hard to interpret the internal dynamics during its training process.
no code implementations • IJCNLP 2019 • Guanlin Li, Lemao Liu, Guoping Huang, Conghui Zhu, Tiejun Zhao
Many Data Augmentation (DA) methods have been proposed for neural machine translation.
no code implementations • ACL 2020 • Xintong Li, Lemao Liu, Rui Wang, Guoping Huang, Max Meng
This paper first provides a method to identify source and target contexts and then introduce a gate mechanism to control the source and target contributions in Transformer.
no code implementations • 13 Aug 2019 • Huayang Li, Guoping Huang, Deng Cai, Lemao Liu
Experiments show that our approach can indeed improve the translation quality with the automatically generated constraints.
no code implementations • ACL 2019 • Xintong Li, Guanlin Li, Lemao Liu, Max Meng, Shuming Shi
Prior researches suggest that neural machine translation (NMT) captures word alignment through its attention mechanism, however, this paper finds attention may almost fail to capture word alignment for some NMT models.
no code implementations • NAACL 2019 • Guanlin Li, Lemao Liu, Xintong Li, Conghui Zhu, Tiejun Zhao, Shuming Shi
Multilayer architectures are currently the gold standard for large-scale neural machine translation.
no code implementations • NAACL 2018 • Xintong Li, Lemao Liu, Zhaopeng Tu, Shuming Shi, Max Meng
In neural machine translation, an attention model is used to identify the aligned source words for a target word (target foresight word) in order to select translation context, but it does not make use of any information of this target foresight word at all.
no code implementations • ACL 2018 • Lianhui Qin, Lemao Liu, Victoria Bi, Yan Wang, Xiaojiang Liu, Zhiting Hu, Hai Zhao, Shuming Shi
Comments of online articles provide extended views and improve user engagement.
no code implementations • EMNLP 2017 • Kehai Chen, Rui Wang, Masao Utiyama, Lemao Liu, Akihiro Tamura, Eiichiro Sumita, Tiejun Zhao
Source dependency information has been successfully introduced into statistical machine translation.
1 code implementation • EMNLP 2017 • Rui Wang, Masao Utiyama, Lemao Liu, Kehai Chen, Eiichiro Sumita
Instance weighting has been widely applied to phrase-based machine translation domain adaptation.
no code implementations • COLING 2016 • Lemao Liu, Masao Utiyama, Andrew Finch, Eiichiro Sumita
The attention mechanisim is appealing for neural machine translation, since it is able to dynam- ically encode a source sentence by generating a alignment between a target word and source words.