no code implementations • EMNLP 2020 • Jian Liu, Yubo Chen, Kang Liu, Wei Bi, Xiaojiang Liu
ii) Our model is excelled in the data-scarce scenario, for example, obtaining 49. 8{\%} in F1 for event argument extraction with only 1{\%} data, compared with 2. 2{\%} of the previous method.
1 code implementation • NAACL 2022 • Jiahao Xu, Yubin Ruan, Wei Bi, Guoping Huang, Shuming Shi, Lihui Chen, Lemao Liu
Back translation (BT) is one of the most significant technologies in NMT research fields.
no code implementations • EMNLP 2021 • Dianbo Sui, Chenhao Wang, Yubo Chen, Kang Liu, Jun Zhao, Wei Bi
In this paper, we formulate end-to-end KBP as a direct set generation problem, avoiding considering the order of multiple facts.
no code implementations • 17 Sep 2023 • Yi Chen, Haiyun Jiang, Wei Bi, Rui Wang, Longyue Wang, Shuming Shi, Ruifeng Xu
This work presents a new task of Text Expansion (TE), which aims to insert fine-grained modifiers into proper locations of the plain text to concretize or vivify human writings.
1 code implementation • 3 Sep 2023 • Yue Zhang, Yafu Li, Leyang Cui, Deng Cai, Lemao Liu, Tingchen Fu, Xinting Huang, Enbo Zhao, Yu Zhang, Yulong Chen, Longyue Wang, Anh Tuan Luu, Wei Bi, Freda Shi, Shuming Shi
While large language models (LLMs) have demonstrated remarkable capabilities across a range of downstream tasks, a significant concern revolves around their propensity to exhibit hallucinations: LLMs occasionally generate content that diverges from the user input, contradicts previously generated context, or misaligns with established world knowledge.
1 code implementation • 20 Jun 2023 • Yafu Li, Leyang Cui, Jianhao Yan, Yongjing Yin, Wei Bi, Shuming Shi, Yue Zhang
Most existing text generation models follow the sequence-to-sequence paradigm.
1 code implementation • 24 May 2023 • Yiyang Li, Xinting Huang, Wei Bi, Hai Zhao
Multi-party dialogues are more difficult for models to understand than one-to-one two-party dialogues, since they involve multiple interlocutors, resulting in interweaving reply-to relations and information flows.
no code implementations • 22 May 2023 • Haoran Yang, Deng Cai, Huayang Li, Wei Bi, Wai Lam, Shuming Shi
We introduce a frustratingly simple, super efficient and surprisingly effective decoding method, which we call Frustratingly Simple Decoding (FSD), for neural text generation.
1 code implementation • 22 May 2023 • Yafu Li, Qintong Li, Leyang Cui, Wei Bi, Longyue Wang, Linyi Yang, Shuming Shi, Yue Zhang
In practical scenarios, the detector faces texts from various domains or LLMs without knowing their sources.
no code implementations • 22 May 2023 • Yue Zhang, Leyang Cui, Deng Cai, Xinting Huang, Tao Fang, Wei Bi
ChatGPT and GPT-4 have attracted substantial interest from both academic and industrial circles, owing to their remarkable few-shot (or even zero-shot) ability to handle various tasks.
1 code implementation • 17 May 2023 • Fanqi Wan, Weizhou Shen, Ke Yang, Xiaojun Quan, Wei Bi
Retrieving proper domain knowledge from an external database lies at the heart of end-to-end task-oriented dialog systems to generate informative responses.
1 code implementation • 19 Dec 2022 • Qintong Li, Zhiyong Wu, Lingpeng Kong, Wei Bi
Explaining the black-box predictions of NLP models naturally and accurately is an important open problem in natural language generation.
no code implementations • 3 Aug 2022 • Shuming Shi, Enbo Zhao, Duyu Tang, Yan Wang, Piji Li, Wei Bi, Haiyun Jiang, Guoping Huang, Leyang Cui, Xinting Huang, Cong Zhou, Yong Dai, Dongyang Ma
In Effidit, we significantly expand the capacities of a writing assistant by providing functions in five categories: text completion, error checking, text polishing, keywords to sentences (K2S), and cloud input methods (cloud IME).
1 code implementation • 9 Jun 2022 • Elia Peruzzo, Enver Sangineto, Yahui Liu, Marco De Nadai, Wei Bi, Bruno Lepri, Nicu Sebe
In this work, we propose a different and complementary direction, in which a local bias is introduced using an auxiliary self-supervised task, performed jointly with standard supervised training.
1 code implementation • CVPR 2023 • Bin Ren, Yahui Liu, Yue Song, Wei Bi, Rita Cucchiara, Nicu Sebe, Wei Wang
In particular, MJP first shuffles the selected patches via our block-wise random jigsaw puzzle shuffle algorithm, and their corresponding PEs are occluded.
1 code implementation • ACL 2022 • Zhiyong Wu, Wei Bi, Xiang Li, Lingpeng Kong, Ben Kao
We propose knowledge internalization (KI), which aims to complement the lexical knowledge into neural dialog models.
1 code implementation • ACL 2022 • Yu Cao, Wei Bi, Meng Fang, Shuming Shi, DaCheng Tao
To alleviate the above data issues, we propose a data manipulation method, which is model-agnostic to be packed with any persona-based dialogue generation model to improve its performance.
1 code implementation • Findings (ACL) 2022 • Qintong Li, Piji Li, Wei Bi, Zhaochun Ren, Yuxuan Lai, Lingpeng Kong
Open-ended text generation tasks, such as dialogue generation and story completion, require models to generate a coherent continuation given limited preceding context.
1 code implementation • NeurIPS 2021 • Yahui Liu, Enver Sangineto, Wei Bi, Nicu Sebe, Bruno Lepri, Marco De Nadai
This task encourages the VTs to learn spatial relations within an image and makes the VT training much more robust when training data are scarce.
no code implementations • ACL 2021 • Zhiyong Wu, Lingpeng Kong, Wei Bi, Xiang Li, Ben Kao
A neural multimodal machine translation (MMT) system is one that aims to perform better translation by extending conventional text-only translation models with multimodal information.
no code implementations • 30 May 2021 • Jun Gao, Wei Bi, Ruifeng Xu, Shuming Shi
We first clarify an assumption on reference-based metrics that, if more high-quality references are added into the reference set, the reliability of the metric will increase.
no code implementations • ACL 2021 • Wei Bi, Huayang Li, Jiacheng Huang
Data augmentation is an effective way to improve the performance of many neural text generation models.
no code implementations • 21 May 2021 • Zhiliang Tian, Wei Bi, Zihan Zhang, Dongkyu Lee, Yiping Song, Nevin L. Zhang
The task requires models to generate personalized responses for a speaker given a few conversations from the speaker and a social network.
no code implementations • 17 Dec 2020 • Zelong Yang, Yan Wang, Piji Li, Shaobin Lin, Shuming Shi, Shao-Lun Huang, Wei Bi
The multiplayer online battle arena (MOBA) games have become increasingly popular in recent years.
1 code implementation • COLING 2020 • Jian Wang, Junhao Liu, Wei Bi, Xiaojiang Liu, Kejing He, Ruifeng Xu, Min Yang
To conquer these limitations, we propose a Dual Dynamic Memory Network (DDMN) for multi-turn dialog generation, which maintains two core components: dialog memory manager and KB memory manager.
1 code implementation • COLING 2020 • Heng Gong, Yawei Sun, Xiaocheng Feng, Bing Qin, Wei Bi, Xiaojiang Liu, Ting Liu
Although neural table-to-text models have achieved remarkable progress with the help of large-scale datasets, they suffer insufficient learning problem with limited training data.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Heng Gong, Wei Bi, Xiaocheng Feng, Bing Qin, Xiaojiang Liu, Ting Liu
Neural table-to-text models, which select and order salient data, as well as verbalizing them fluently via surface realization, have achieved promising progress.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Yu Cao, Wei Bi, Meng Fang, DaCheng Tao
In this work, we study dialogue models with multiple input sources adapted from the pretrained language model GPT2.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yifan Gao, Piji Li, Wei Bi, Xiaojiang Liu, Michael R. Lyu, Irwin King
Besides a small number of high-resource sentence functions, a large portion of sentence functions is infrequent.
no code implementations • ACL 2020 • Zhiliang Tian, Wei Bi, Dongkyu Lee, Lanqing Xue, Yiping Song, Xiaojiang Liu, Nevin L. Zhang
In previous work, the external document is utilized by (1) creating a context-aware document memory that integrates information from the document and the conversational context, and then (2) generating responses referring to the memory.
1 code implementation • ACL 2020 • Qile Zhu, Jianlin Su, Wei Bi, Xiaojiang Liu, Xiyao Ma, Xiaolin Li, Dapeng Wu
Variational Autoencoder (VAE) is widely used as a generative model to approximate a model's posterior on latent variables by combining the amortized variational inference and deep neural networks.
1 code implementation • 24 Feb 2020 • Xiaocheng Feng, Yawei Sun, Bing Qin, Heng Gong, Yibo Sun, Wei Bi, Xiaojiang Liu, Ting Liu
In this paper, we focus on a new practical task, document-scale text content manipulation, which is the opposite of text style transfer and aims to preserve text styles while altering the content.
1 code implementation • 16 Dec 2019 • Jian Wang, Junhao Liu, Wei Bi, Xiaojiang Liu, Kejing He, Ruifeng Xu, Min Yang
In this paper, we propose a novel knowledge-aware dialogue generation model (called TransDG), which transfers question representation and knowledge matching abilities from knowledge base question answering (KBQA) task to facilitate the utterance understanding and factual knowledge selection for dialogue generation.
no code implementations • 26 Nov 2019 • Xin Li, Piji Li, Wei Bi, Xiaojiang Liu, Wai Lam
In this paper, we propose to formulate the STC task as a language modeling problem and tailor-make a training strategy to adapt a language model for response generation.
no code implementations • IJCNLP 2019 • Jun Gao, Wei Bi, Xiaojiang Liu, Junhui Li, Guodong Zhou, Shuming Shi
In this paper, we introduce a discrete latent variable with an explicit semantic meaning to improve the CVAE on short-text conversation.
no code implementations • IJCNLP 2019 • Deng Cai, Yan Wang, Wei Bi, Zhaopeng Tu, Xiaojiang Liu, Shuming Shi
End-to-end sequence generation is a popular technique for developing open domain dialogue systems, though they suffer from the \textit{safe response problem}.
1 code implementation • ACL 2020 • Yiping Song, Zequn Liu, Wei Bi, Rui Yan, Ming Zhang
Training the generative models with minimal corpus is one of the critical challenges for building open-domain dialogue systems.
no code implementations • 18 Sep 2019 • Jiangtong Li, Hai Zhao, Zuchao Li, Wei Bi, Xiaojiang Liu
Embedding from Language Models (ELMo) has shown to be effective for improving many natural language processing (NLP) tasks, and ELMo takes character information to compose word representation to train language models. However, the character is an insufficient and unnatural linguistic unit for word representation. Thus we introduce Embedding from Subword-aware Language Models (ESuLMo) which learns word representation from subwords using unsupervised segmentation over words. We show that ESuLMo can enhance four benchmark NLP tasks more effectively than ELMo, including syntactic dependency parsing, semantic role labeling, implicit discourse relation recognition and textual entailment, which brings a meaningful improvement over ELMo.
no code implementations • ACL 2019 • Wei Bi, Jun Gao, Xiaojiang Liu, Shuming Shi
Classification models are trained on this dataset to (i) recognize the sentence function of new data in a large corpus of short-text conversations; (ii) estimate a proper sentence function of the response given a test query.
no code implementations • ACL 2019 • Lisong Qiu, Juntao Li, Wei Bi, Dongyan Zhao, Rui Yan
Due to its potential applications, open-domain dialogue generation has become popular and achieved remarkable progress in recent years, but sometimes suffers from generic responses.
no code implementations • ACL 2019 • Yang Zhao, Xiaoyu Shen, Wei Bi, Akiko Aizawa
First, the word graph approach that simply concatenates fragments from multiple sentences may yield non-fluent or ungrammatical compression.
1 code implementation • ACL 2019 • Zhiliang Tian, Wei Bi, Xiaopeng Li, Nevin L. Zhang
In this work, we propose a memory-augmented generative model, which learns to abstract from the training corpus and saves the useful information to the memory to assist the response generation.
no code implementations • 14 Nov 2018 • Jun Gao, Wei Bi, Xiaojiang Liu, Junhui Li, Shuming Shi
In this paper, we propose a novel response generation model, which considers a set of responses jointly and generates multiple diverse responses simultaneously.
1 code implementation • EMNLP 2018 • Yahui Liu, Wei Bi, Jun Gao, Xiaojiang Liu, Jian Yao, Shuming Shi
We observe that in the conversation tasks, each query could have multiple responses, which forms a 1-to-n or m-to-n relationship in the view of the total corpus.
no code implementations • 13 Aug 2018 • Yanpeng Zhao, Wei Bi, Deng Cai, Xiaojiang Liu, Kewei Tu, Shuming Shi
Then, by recombining the content with the target style, we decode a sentence aligned in the target domain.
no code implementations • NeurIPS 2012 • Wei Bi, James T. Kwok
However, while there have been a lot of MLNP methods in hierarchical multiclass classification, performing MLNP in hierarchical multilabel classification is much more difficult.