no code implementations • Findings (EMNLP) 2021 • Yu Feng, Jing Zhang, Gaole He, Wayne Xin Zhao, Lemao Liu, Quan Liu, Cuiping Li, Hong Chen
Knowledge Base Question Answering (KBQA) is to answer natural language questions posed over knowledge bases (KBs).
no code implementations • EMNLP 2021 • Yi Chen, Haiyun Jiang, Lemao Liu, Shuming Shi, Chuang Fan, Min Yang, Ruifeng Xu
Auxiliary information from multiple sources has been demonstrated to be effective in zero-shot fine-grained entity typing (ZFET).
no code implementations • Xintong Li, Lemao Liu, Guanlin Li, Max Meng, Shuming Shi
We find that although NMT models are difficult to capture word alignment for CFT words but these words do not sacrifice translation quality significantly, which provides an explanation why NMT is more successful for translation yet worse for word alignment compared to statistical machine translation.
no code implementations • ACL 2022 • Yi Chen, Jiayang Cheng, Haiyun Jiang, Lemao Liu, Haisong Zhang, Shuming Shi, Ruifeng Xu
In this paper, we firstly empirically find that existing models struggle to handle hard mentions due to their insufficient contexts, which consequently limits their overall typing performance.
no code implementations • ACL 2022 • Yanling Xiao, Lemao Liu, Guoping Huang, Qu Cui, ShuJian Huang, Shuming Shi, Jiajun Chen
In this work, we propose a novel BiTIIMT system, Bilingual Text-Infilling for Interactive Neural Machine Translation.
1 code implementation • EMNLP 2021 • Jing Qian, Yibin Liu, Lemao Liu, Yangming Li, Haiyun Jiang, Haisong Zhang, Shuming Shi
Existing work on Fine-grained Entity Typing (FET) typically trains automatic models on the datasets obtained by using Knowledge Bases (KB) as distant supervision.
no code implementations • 24 Feb 2025 • Xuanfan Ni, Liyan Xu, Chenyang Lyu, Longyue Wang, Mo Yu, Lemao Liu, Fandong Meng, Jie zhou, Piji Li
To alleviate memory burden during inference of large language models (LLMs), numerous studies have focused on compressing the KV cache by exploring aspects such as attention sparsity.
1 code implementation • 13 Feb 2025 • Mo Yu, Lemao Liu, Junjie Wu, Tsz Ting Chung, Shunchi Zhang, Jiangnan Li, Dit-yan Yeung, Jie zhou
In a systematic way, we investigate a widely asked question: Do LLMs really understand what they say?, which relates to the more familiar term Stochastic Parrot.
no code implementations • 11 Feb 2025 • Junjie Wu, Mo Yu, Lemao Liu, Dit-yan Yeung, Jie zhou
While LLMs have exhibited strong performance on various NLP tasks, it is noteworthy that most of these tasks rely on utilizing the vast amount of knowledge encoded in LLMs' parameters, rather than solving new problems without prior knowledge.
1 code implementation • 12 Nov 2024 • Siheng Li, Cheng Yang, Zesen Cheng, Lemao Liu, Mo Yu, Yujiu Yang, Wai Lam
Large language models (LLMs) have achieved substantial progress in processing long contexts but still struggle with long-context reasoning.
no code implementations • 15 Oct 2024 • Tsz Ting Chung, Leyang Cui, Lemao Liu, Xinting Huang, Shuming Shi, Dit-yan Yeung
Large Language Models (LLMs) have demonstrated impressive capabilities in a wide range of natural language processing tasks when leveraging in-context learning.
2 code implementations • 27 Sep 2024 • Siheng Li, Cheng Yang, Taiqiang Wu, Chufan Shi, Yuji Zhang, Xinyu Zhu, Zesen Cheng, Deng Cai, Mo Yu, Lemao Liu, Jie zhou, Yujiu Yang, Ngai Wong, Xixin Wu, Wai Lam
Honesty is a fundamental principle for aligning large language models (LLMs) with human values, requiring these models to recognize what they know and don't know and be able to faithfully express their knowledge.
no code implementations • 26 Aug 2024 • Zelin Li, Kehai Chen, Lemao Liu, Xuefeng Bai, Mingming Yang, Yang Xiang, Min Zhang
In this paper, we analyze the core mechanisms of previous predominant adversarial attack methods, revealing that 1) the distributions of importance score differ markedly among victim models, restricting the transferability; 2) the sequential attack processes induces substantial time overheads.
1 code implementation • 29 Jul 2024 • Cheng Yang, Guoping Huang, Mo Yu, Zhirui Zhang, Siheng Li, Mingming Yang, Shuming Shi, Yujiu Yang, Lemao Liu
Existing work addresses this task through a classification model based on a neural network that maps the hidden vector of the input context into its corresponding label (i. e., the candidate target word is treated as a label).
1 code implementation • 7 Jul 2024 • Junjie Wu, Lemao Liu, Wei Bi, Dit-yan Yeung
To this end, this paper presents a new setting for NMT targeted adversarial attacks that could lead to reliable attacking results.
no code implementations • 24 Jun 2024 • Deng Cai, Huayang Li, Tingchen Fu, Siheng Li, Weiwen Xu, Shuaiyi Li, Bowen Cao, Zhisong Zhang, Xinting Huang, Leyang Cui, Yan Wang, Lemao Liu, Taro Watanabe, Shuming Shi
Despite the general capabilities of pre-trained large language models (LLMs), they still need further adaptation to better serve practical applications.
1 code implementation • 11 Jun 2024 • Meizhi Zhong, Kehai Chen, Zhengshan Xue, Lemao Liu, Mingming Yang, Min Zhang
It is widely known that hallucination is a critical issue in Simultaneous Machine Translation (SiMT) due to the absence of source-side information.
1 code implementation • 22 May 2024 • Tingchen Fu, Deng Cai, Lemao Liu, Shuming Shi, Rui Yan
However, the performance of LLMs on standard knowledge and reasoning benchmarks tends to suffer from deterioration at the latter stage of the SFT process, echoing the phenomenon of alignment tax.
1 code implementation • 25 Mar 2024 • Huayang Li, Deng Cai, Zhi Qu, Qu Cui, Hidetaka Kamigaito, Lemao Liu, Taro Watanabe
In our work, we propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval, which aims to augment cross-lingual applications by addressing polysemy using context information.
no code implementations • 21 Feb 2024 • Xueliang Zhao, Xinting Huang, Tingchen Fu, Qintong Li, Shansan Gong, Lemao Liu, Wei Bi, Lingpeng Kong
Multimodal reasoning stands as a pivotal capability for large vision-language models (LVLMs).
1 code implementation • 16 Dec 2023 • Qihang Ai, Jianwu Zhou, Haiyun Jiang, Lemao Liu, Shuming Shi
Graph data is ubiquitous in the physical world, and it has always been a challenge to efficiently model graph structures using a unified paradigm for the understanding and reasoning on various graphs.
1 code implementation • 13 Nov 2023 • Meizhi Zhong, Lemao Liu, Kehai Chen, Mingming Yang, Min Zhang
Simultaneous Machine Translation (SiMT) aims to yield a real-time partial translation with a monotonically growing the source-side context.
no code implementations • 3 Nov 2023 • Yifan Wang, Qingyan Guo, Xinzhe Ni, Chufan Shi, Lemao Liu, Haiyun Jiang, Yujiu Yang
In-context learning (ICL) ability has emerged with the increasing scale of large language models (LLMs), enabling them to learn input-label mappings from demonstrations and perform well on downstream tasks.
1 code implementation • 23 Oct 2023 • Xingyu Chen, Lemao Liu, Guoping Huang, Zhirui Zhang, Mingming Yang, Shuming Shi, Rui Wang
Word-Level Auto-Completion (WLAC) plays a crucial role in Computer-Assisted Translation.
1 code implementation • NAACL 2022 • Jiahao Xu, Yubin Ruan, Wei Bi, Guoping Huang, Shuming Shi, Lihui Chen, Lemao Liu
Back translation (BT) is one of the most significant technologies in NMT research fields.
1 code implementation • 20 Oct 2023 • Jiahao Xu, Wei Shao, Lihui Chen, Lemao Liu
This paper proposes the DistillCSE framework, which performs contrastive learning under the self-training paradigm with knowledge distillation.
1 code implementation • 20 Oct 2023 • Junjie Wu, Lemao Liu, Dit-yan Yeung
Behavioral testing offers a crucial means of diagnosing linguistic errors and assessing capabilities of NLP models.
1 code implementation • 17 Oct 2023 • Xu Huang, Zhirui Zhang, Ruize Gao, Yichao Du, Lemao Liu, Gouping Huang, Shuming Shi, Jiajun Chen, ShuJian Huang
We present IMTLab, an open-source end-to-end interactive machine translation (IMT) system platform that enables researchers to quickly build IMT systems with state-of-the-art models, perform an end-to-end evaluation, and diagnose the weakness of systems.
1 code implementation • 14 Sep 2023 • Huayang Li, Siheng Li, Deng Cai, Longyue Wang, Lemao Liu, Taro Watanabe, Yujiu Yang, Shuming Shi
We release our dataset, model, and demo to foster future research in the area of multimodal instruction following.
Ranked #230 on
Visual Question Answering
on MM-Vet
1 code implementation • 3 Sep 2023 • Yue Zhang, Yafu Li, Leyang Cui, Deng Cai, Lemao Liu, Tingchen Fu, Xinting Huang, Enbo Zhao, Yu Zhang, Yulong Chen, Longyue Wang, Anh Tuan Luu, Wei Bi, Freda Shi, Shuming Shi
While large language models (LLMs) have demonstrated remarkable capabilities across a range of downstream tasks, a significant concern revolves around their propensity to exhibit hallucinations: LLMs occasionally generate content that diverges from the user input, contradicts previously generated context, or misaligns with established world knowledge.
no code implementations • 12 Jun 2023 • Hongkun Hao, Guoping Huang, Lemao Liu, Zhirui Zhang, Shuming Shi, Rui Wang
The finding demonstrates that TM-augmented NMT is good at the ability of fitting data (i. e., lower bias) but is more sensitive to the fluctuations in the training data (i. e., higher variance), which provides an explanation to a recently reported contradictory phenomenon on the same translation task: TM-augmented NMT substantially advances vanilla NMT under the high-resource scenario whereas it fails under the low-resource scenario.
no code implementations • 4 Jun 2023 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
Sentence embedding is one of the most fundamental tasks in Natural Language Processing and plays an important role in various tasks.
1 code implementation • 29 May 2023 • Zhen Zhang, Mengting Hu, Shiwan Zhao, Minlie Huang, Haotian Wang, Lemao Liu, Zhirui Zhang, Zhe Liu, Bingzhe Wu
Most named entity recognition (NER) systems focus on improving model performance, ignoring the need to quantify model uncertainty, which is critical to the reliability of NER systems in open environments.
1 code implementation • 22 May 2023 • Ruize Gao, Zhirui Zhang, Yichao Du, Lemao Liu, Rui Wang
Nearest Neighbor Machine Translation ($k$NN-MT) has achieved great success in domain adaptation tasks by integrating pre-trained Neural Machine Translation (NMT) models with domain-specific token-level retrieval.
no code implementations • 22 May 2023 • Jiahao Xu, Wei Shao, Lihui Chen, Lemao Liu
This paper improves contrastive learning for sentence embeddings from two perspectives: handling dropout noise and addressing feature corruption.
no code implementations • 13 May 2023 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Ying Chen
Static word embedding is still useful, particularly for context-unavailable tasks, because in the case of no context available, pre-trained language models often perform worse than static word embeddings.
no code implementations • 13 May 2023 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
Generating proper embedding of sentences through an unsupervised way is beneficial to semantic matching and retrieval problems in real-world scenarios.
1 code implementation • 3 May 2023 • Xin Cheng, Di Luo, Xiuying Chen, Lemao Liu, Dongyan Zhao, Rui Yan
In this paper, by exploring the duality of the primal problem: better generation also prompts better memory, we propose a novel framework, selfmem, which addresses this limitation by iteratively employing a retrieval-augmented generator to create an unbounded memory pool and using a memory selector to choose one output as memory for the subsequent generation round.
Ranked #1 on
Text Summarization
on X-Sum
no code implementations • 23 Feb 2023 • Yichao Du, Zhirui Zhang, Bingzhe Wu, Lemao Liu, Tong Xu, Enhong Chen
To protect user privacy and meet legal regulations, federated learning (FL) is attracting significant attention.
1 code implementation • 6 Dec 2022 • Xin Cheng, Shen Gao, Lemao Liu, Dongyan Zhao, Rui Yan
Retrieval-augmented Neural Machine Translation models have been successful in many translation scenarios.
no code implementations • 22 Oct 2022 • Xueliang Zhao, Lemao Liu, Tingchen Fu, Shuming Shi, Dongyan Zhao, Rui Yan
With the availability of massive general-domain dialogue data, pre-trained dialogue generation appears to be super appealing to transfer knowledge from the general domain to downstream applications.
no code implementations • 4 Jul 2022 • Yinya Huang, Lemao Liu, Kun Xu, Meng Fang, Liang Lin, Xiaodan Liang
In this work, we propose logic structural-constraint modeling to solve the logical reasoning QA and introduce discourse-aware graph networks (DAGNs).
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Yahui Liu, Lemao Liu, Guoping Huang, Defu Lian, Shuming Shi
Current practices in metric evaluation focus on one single dataset, e. g., Newstest dataset in each year's WMT Metrics Shared Task.
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Defu Lian, Guoping Huang, Taro Watanabe, Lemao Liu
To this end, we study the dynamic relationship between the encoded linguistic information and task performance from the viewpoint of Pareto Optimality.
1 code implementation • 9 Mar 2022 • Wenye Lin, Yangming Li, Lemao Liu, Shuming Shi, Hai-Tao Zheng
Specifically, we transfer the knowledge from a teacher model to its student model by locally matching their predictions on all sub-structures, instead of the whole output space.
1 code implementation • 17 Feb 2022 • Lingfeng Shen, Lemao Liu, Haiyun Jiang, Shuming Shi
In this paper we revisit automatic metrics for paraphrase evaluation and obtain two findings that disobey conventional wisdom: (1) Reference-free metrics achieve better performance than their reference-based counterparts.
no code implementations • 2 Feb 2022 • Huayang Li, Yixuan Su, Deng Cai, Yan Wang, Lemao Liu
Recently, retrieval-augmented text generation attracted increasing attention of the computational linguistics community.
no code implementations • 9 Jan 2022 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
It has been shown that natural language processing (NLP) models are vulnerable to a kind of security threat called the Backdoor Attack, which utilizes a `backdoor trigger' paradigm to mislead the models.
1 code implementation • 12 Dec 2021 • Yu Feng, Jing Zhang, Xiaokang Zhang, Lemao Liu, Cuiping Li, Hong Chen
Embedding-based methods are popular for Knowledge Base Question Answering (KBQA), but few current models have numerical reasoning skills and thus struggle to answer ordinal constrained questions.
no code implementations • ACL 2022 • Yangming Li, Lemao Liu, Shuming Shi
Negative sampling is highly effective in handling missing annotations for named entity recognition (NER).
1 code implementation • ACL 2021 • Zexin Lu, Keyang Ding, Yuji Zhang, Jing Li, Baolin Peng, Lemao Liu
This paper presents a novel task to generate poll questions for social media posts.
Ranked #3 on
Answer Generation
on WeiboPolls
no code implementations • ACL 2021 • Qiuxiang He, Guoping Huang, Qu Cui, Li Li, Lemao Liu
It is generally believed that a translation memory (TM) should be beneficial for machine translation tasks.
no code implementations • ACL 2021 • Lemao Liu, Haisong Zhang, Haiyun Jiang, Yangming Li, Enbo Zhao, Kun Xu, Linfeng Song, Suncong Zheng, Botong Zhou, Dick Zhu, Xiao Feng, Tao Chen, Tao Yang, Dong Yu, Feng Zhang, Zhanhui Kang, Shuming Shi
This paper introduces TexSmart, a text understanding system that supports fine-grained named entity recognition (NER) and enhanced semantic analysis functionalities.
no code implementations • ACL 2021 • Huayang Li, Lemao Liu, Guoping Huang, Shuming Shi
In this paper, we propose the task of general word-level autocompletion (GWLAN) from a real-world CAT scenario, and construct the first public benchmark to facilitate research in this topic.
no code implementations • 27 May 2021 • Guoping Huang, Lemao Liu, Xing Wang, Longyue Wang, Huayang Li, Zhaopeng Tu, Chengyan Huang, Shuming Shi
Automatic machine translation is super efficient to produce translations yet their quality is not guaranteed.
1 code implementation • ACL 2021 • Deng Cai, Yan Wang, Huayang Li, Wai Lam, Lemao Liu
Second, the memory retriever and NMT model can be jointly optimized for the ultimate translation goal.
1 code implementation • Findings (ACL) 2021 • Jiannan Xiang, Yahui Liu, Deng Cai, Huayang Li, Defu Lian, Lemao Liu
An important aspect of developing dialogue systems is how to evaluate and compare the performance of different systems.
1 code implementation • NAACL 2021 • Yangming Li, Lemao Liu, Kaisheng Yao
Prior methods to text segmentation are mostly at token level.
no code implementations • 1 Jan 2021 • Guanlin Li, Lemao Liu, Taro Watanabe, Conghui Zhu, Tiejun Zhao
Unsupervised Neural Machine Translation or UNMT has received great attention in recent years.
no code implementations • 31 Dec 2020 • Haisong Zhang, Lemao Liu, Haiyun Jiang, Yangming Li, Enbo Zhao, Kun Xu, Linfeng Song, Suncong Zheng, Botong Zhou, Jianchen Zhu, Xiao Feng, Tao Chen, Tao Yang, Dong Yu, Feng Zhang, Zhanhui Kang, Shuming Shi
This technique report introduces TexSmart, a text understanding system that supports fine-grained named entity recognition (NER) and enhanced semantic analysis functionalities.
no code implementations • Findings (EMNLP) 2021 • Yangming Li, Lemao Liu, Shuming Shi
In this work, we present Lexical Unit Analysis (LUA), a framework for general sequence segmentation tasks.
1 code implementation • ICLR 2021 • Yangming Li, Lemao Liu, Shuming Shi
Experiments on synthetic datasets and real-world datasets show that our model is robust to unlabeled entity problem and surpasses prior baselines.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Qian Wang, Jiajun Zhang, Lemao Liu, Guoping Huang, Chengqing Zong
We propose a touch-based editing method for translation, which is more flexible than traditional keyboard-mouse-based translation postediting.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Huayang Li, Lemao Liu, Guoping Huang, Shuming Shi
Many efforts have been devoted to extracting constituency trees from pre-trained language models, often proceeding in two stages: feature definition and parsing.
no code implementations • ACL 2020 • Jierui Li, Lemao Liu, Huayang Li, Guanlin Li, Guoping Huang, Shuming Shi
Recently many efforts have been devoted to interpreting the black-box NMT models, but little progress has been made on metrics to evaluate explanation methods.
no code implementations • 5 Apr 2020 • Conghui Zhu, Guanlin Li, Lemao Liu, Tiejun Zhao, Shuming Shi
Despite the great success of NMT, there still remains a severe challenge: it is hard to interpret the internal dynamics during its training process.
no code implementations • 5 Apr 2020 • Guanlin Li, Lemao Liu, Conghui Zhu, Tiejun Zhao, Shuming Shi
Generalization to unseen instances is our eternal pursuit for all data-driven models.
no code implementations • IJCNLP 2019 • Guanlin Li, Lemao Liu, Guoping Huang, Conghui Zhu, Tiejun Zhao
Many Data Augmentation (DA) methods have been proposed for neural machine translation.
no code implementations • ACL 2020 • Xintong Li, Lemao Liu, Rui Wang, Guoping Huang, Max Meng
This paper first provides a method to identify source and target contexts and then introduce a gate mechanism to control the source and target contributions in Transformer.
no code implementations • 13 Aug 2019 • Huayang Li, Guoping Huang, Deng Cai, Lemao Liu
Experiments show that our approach can indeed improve the translation quality with the automatically generated constraints.
no code implementations • ACL 2019 • Xintong Li, Guanlin Li, Lemao Liu, Max Meng, Shuming Shi
Prior researches suggest that neural machine translation (NMT) captures word alignment through its attention mechanism, however, this paper finds attention may almost fail to capture word alignment for some NMT models.
no code implementations • NAACL 2019 • Guanlin Li, Lemao Liu, Xintong Li, Conghui Zhu, Tiejun Zhao, Shuming Shi
Multilayer architectures are currently the gold standard for large-scale neural machine translation.
no code implementations • NAACL 2018 • Xintong Li, Lemao Liu, Zhaopeng Tu, Shuming Shi, Max Meng
In neural machine translation, an attention model is used to identify the aligned source words for a target word (target foresight word) in order to select translation context, but it does not make use of any information of this target foresight word at all.
no code implementations • ACL 2018 • Lianhui Qin, Lemao Liu, Victoria Bi, Yan Wang, Xiaojiang Liu, Zhiting Hu, Hai Zhao, Shuming Shi
Comments of online articles provide extended views and improve user engagement.
no code implementations • EMNLP 2017 • Kehai Chen, Rui Wang, Masao Utiyama, Lemao Liu, Akihiro Tamura, Eiichiro Sumita, Tiejun Zhao
Source dependency information has been successfully introduced into statistical machine translation.
1 code implementation • EMNLP 2017 • Rui Wang, Masao Utiyama, Lemao Liu, Kehai Chen, Eiichiro Sumita
Instance weighting has been widely applied to phrase-based machine translation domain adaptation.
no code implementations • COLING 2016 • Lemao Liu, Masao Utiyama, Andrew Finch, Eiichiro Sumita
The attention mechanisim is appealing for neural machine translation, since it is able to dynam- ically encode a source sentence by generating a alignment between a target word and source words.