no code implementations • EMNLP 2021 • Yi Chen, Haiyun Jiang, Lemao Liu, Shuming Shi, Chuang Fan, Min Yang, Ruifeng Xu
Auxiliary information from multiple sources has been demonstrated to be effective in zero-shot fine-grained entity typing (ZFET).
1 code implementation • EMNLP 2021 • Jing Qian, Yibin Liu, Lemao Liu, Yangming Li, Haiyun Jiang, Haisong Zhang, Shuming Shi
Existing work on Fine-grained Entity Typing (FET) typically trains automatic models on the datasets obtained by using Knowledge Bases (KB) as distant supervision.
no code implementations • ACL 2022 • Yi Chen, Jiayang Cheng, Haiyun Jiang, Lemao Liu, Haisong Zhang, Shuming Shi, Ruifeng Xu
In this paper, we firstly empirically find that existing models struggle to handle hard mentions due to their insufficient contexts, which consequently limits their overall typing performance.
no code implementations • 15 Oct 2024 • Bolun Sun, Yifan Zhou, Haiyun Jiang
This paper presents a novel application of large language models (LLMs) to enhance user comprehension of privacy policies through an interactive dialogue agent.
1 code implementation • 10 Jul 2024 • Yuhan Li, Peisong Wang, Xiao Zhu, Aochuan Chen, Haiyun Jiang, Deng Cai, Victor Wai Kin Chan, Jia Li
To bridge this gap, we introduce GLBench, the first comprehensive benchmark for evaluating GraphLLM methods in both supervised and zero-shot scenarios.
no code implementations • 24 Apr 2024 • Jincheng Dai, Zhuowei Huang, Haiyun Jiang, Chen Chen, Deng Cai, Wei Bi, Shuming Shi
Our validation shows that CORM reduces the inference memory usage of KV cache by up to 70\% with negligible performance degradation across six tasks in LongBench.
no code implementations • 18 Feb 2024 • Jiaqi Li, Miaozeng Du, Chuanyi Zhang, Yongrui Chen, Nan Hu, Guilin Qi, Haiyun Jiang, Siyuan Cheng, Bozhong Tian
Multimodal knowledge editing represents a critical advancement in enhancing the capabilities of Multimodal Large Language Models (MLLMs).
2 code implementations • 28 Dec 2023 • Zhongshen Zeng, Pengguang Chen, Shu Liu, Haiyun Jiang, Jiaya Jia
In this work, we introduce a novel evaluation paradigm for Large Language Models (LLMs) that compels them to transition from a traditional question-answering role, akin to a student, to a solution-scoring role, akin to a teacher.
no code implementations • 16 Dec 2023 • Qihang Ai, Jianwu Zhou, Haiyun Jiang, Lemao Liu, Shuming Shi
Graph data is ubiquitous in the physical world, and it has always been a challenge to efficiently model graph structures using a unified paradigm for the understanding and reasoning on various graphs.
1 code implementation • 15 Nov 2023 • Chang Gao, Haiyun Jiang, Deng Cai, Shuming Shi, Wai Lam
Most existing prompting methods suffer from the issues of generalizability and consistency, as they often rely on instance-specific solutions that may not be applicable to other instances and lack task-level consistency across the selected few-shot examples.
no code implementations • 3 Nov 2023 • Yifan Wang, Qingyan Guo, Xinzhe Ni, Chufan Shi, Lemao Liu, Haiyun Jiang, Yujiu Yang
In-context learning (ICL) ability has emerged with the increasing scale of large language models (LLMs), enabling them to learn input-label mappings from demonstrations and perform well on downstream tasks.
no code implementations • 17 Sep 2023 • Yi Chen, Haiyun Jiang, Wei Bi, Rui Wang, Longyue Wang, Shuming Shi, Ruifeng Xu
This work presents a new task of Text Expansion (TE), which aims to insert fine-grained modifiers into proper locations of the plain text to concretize or vivify human writings.
1 code implementation • 12 Sep 2023 • Tinghui Zhu, Jingping Liu, Jiaqing Liang, Haiyun Jiang, Yanghua Xiao, ZongYu Wang, Rui Xie, Yunsen Xian
Specifically, on the Chinese taxonomy dataset, our method significantly improves accuracy by 8. 75 %.
1 code implementation • 11 Sep 2023 • Yongrui Chen, Haiyun Jiang, Xinting Huang, Shuming Shi, Guilin Qi
In particular, compared to the best-performing baseline, the LLM trained using our generated dataset exhibits a 10\% relative improvement in performance on AlpacaEval, despite utilizing only 1/5 of its training data.
1 code implementation • 16 Jul 2023 • Longyue Wang, Zefeng Du, Donghuai Liu, Deng Cai, Dian Yu, Haiyun Jiang, Yan Wang, Leyang Cui, Shuming Shi, Zhaopeng Tu
Modeling discourse -- the linguistic phenomena that go beyond individual sentences, is a fundamental yet challenging aspect of natural language processing (NLP).
no code implementations • 4 Jun 2023 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
Sentence embedding is one of the most fundamental tasks in Natural Language Processing and plays an important role in various tasks.
no code implementations • 13 May 2023 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Ying Chen
Static word embedding is still useful, particularly for context-unavailable tasks, because in the case of no context available, pre-trained language models often perform worse than static word embeddings.
no code implementations • 13 May 2023 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
Generating proper embedding of sentences through an unsupervised way is beneficial to semantic matching and retrieval problems in real-world scenarios.
1 code implementation • 3 Apr 2023 • Yi Chen, Rui Wang, Haiyun Jiang, Shuming Shi, Ruifeng Xu
Evaluating the quality of generated text is a challenging task in NLP, due to the inherent complexity and diversity of text.
1 code implementation • 23 Mar 2023 • Mingyang Song, Haiyun Jiang, Shuming Shi, Songfang Yao, Shilong Lu, Yi Feng, Huafeng Liu, Liping Jing
Based on our findings, we conclude that ChatGPT has great potential for keyphrase generation.
no code implementations • 3 Feb 2023 • Lingfeng Shen, Ze Zhang, Haiyun Jiang, Ying Chen
A recent line of work, detection-based defense, aims to distinguish adversarial sentences from benign ones.
1 code implementation • 2 Dec 2022 • Hongzhan Lin, Pengyao Yi, Jing Ma, Haiyun Jiang, Ziyang Luo, Shuming Shi, Ruifang Liu
The spread of rumors along with breaking events seriously hinders the truth in the era of social media.
1 code implementation • 13 Nov 2022 • Nuo Chen, Yan Wang, Haiyun Jiang, Deng Cai, Yuhan Li, Ziyang Chen, Longyue Wang, Jia Li
In this paper, we introduce the Harry Potter Dialogue (HPD) dataset, designed to advance the study of dialogue agents and character alignment.
no code implementations • 3 Aug 2022 • Shuming Shi, Enbo Zhao, Duyu Tang, Yan Wang, Piji Li, Wei Bi, Haiyun Jiang, Guoping Huang, Leyang Cui, Xinting Huang, Cong Zhou, Yong Dai, Dongyang Ma
In Effidit, we significantly expand the capacities of a writing assistant by providing functions in five categories: text completion, error checking, text polishing, keywords to sentences (K2S), and cloud input methods (cloud IME).
1 code implementation • 30 Jun 2022 • Jingping Liu, Yuqiu Song, Kui Xue, Hongli Sun, Chao Wang, Lihan Chen, Haiyun Jiang, Jiaqing Liang, Tong Ruan
Specifically, we focus on layer tuning for feed-forward network in the Transformer, namely FL-tuning.
no code implementations • 3 Mar 2022 • Mao Yan Chen, Haiyun Jiang, Yujiu Yang
The short text matching task employs a model to determine whether two short texts have the same semantic meaning or intent.
1 code implementation • 17 Feb 2022 • Lingfeng Shen, Lemao Liu, Haiyun Jiang, Shuming Shi
In this paper we revisit automatic metrics for paraphrase evaluation and obtain two findings that disobey conventional wisdom: (1) Reference-free metrics achieve better performance than their reference-based counterparts.
no code implementations • 9 Jan 2022 • Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi
It has been shown that natural language processing (NLP) models are vulnerable to a kind of security threat called the Backdoor Attack, which utilizes a `backdoor trigger' paradigm to mislead the models.
no code implementations • ACL 2021 • Lemao Liu, Haisong Zhang, Haiyun Jiang, Yangming Li, Enbo Zhao, Kun Xu, Linfeng Song, Suncong Zheng, Botong Zhou, Dick Zhu, Xiao Feng, Tao Chen, Tao Yang, Dong Yu, Feng Zhang, Zhanhui Kang, Shuming Shi
This paper introduces TexSmart, a text understanding system that supports fine-grained named entity recognition (NER) and enhanced semantic analysis functionalities.
no code implementations • 7 Apr 2021 • Jiayang Cheng, Haiyun Jiang, Deqing Yang, Yanghua Xiao
However, few works have focused on how to validate and correct the results generated by the existing relation extraction models.
no code implementations • 31 Dec 2020 • Haisong Zhang, Lemao Liu, Haiyun Jiang, Yangming Li, Enbo Zhao, Kun Xu, Linfeng Song, Suncong Zheng, Botong Zhou, Jianchen Zhu, Xiao Feng, Tao Chen, Tao Yang, Dong Yu, Feng Zhang, Zhanhui Kang, Shuming Shi
This technique report introduces TexSmart, a text understanding system that supports fine-grained named entity recognition (NER) and enhanced semantic analysis functionalities.
no code implementations • 17 Dec 2020 • Zhendong Chu, Haiyun Jiang, Yanghua Xiao, Wei Wang
We see information sources as multiple views and fusing them to construct an intact space with sufficient information.
no code implementations • 9 Dec 2020 • Haiyun Jiang, Qiaoben Bao, Qiao Cheng, Deqing Yang, Li Wang, Yanghua Xiao
In recent years, many complex relation extraction tasks, i. e., the variants of simple binary relation extraction, are proposed to meet the complex applications in practice.
no code implementations • ACL 2019 • Jiangjie Chen, Ao Wang, Haiyun Jiang, Suo Feng, Chenguang Li, Yanghua Xiao
A type description is a succinct noun compound which helps human and machines to quickly grasp the informative and distinctive information of an entity.
1 code implementation • 21 Feb 2019 • Jindong Chen, Yizhou Hu, Jingping Liu, Yanghua Xiao, Haiyun Jiang
For the purpose of measuring the importance of knowledge, we introduce attention mechanisms and propose deep Short Text Classification with Knowledge powered Attention (STCKA).