1 code implementation • 10 Mar 2025 • Ruoxi Xu, Hongyu Lin, Xianpei Han, Jia Zheng, Weixiang Zhou, Le Sun, Yingfei Sun
As large language models (LLMs) increasingly become central to various applications and interact with diverse user populations, ensuring their reliable and consistent performance is becoming more important.
no code implementations • 5 Mar 2025 • Zichao Li, Xueru Wen, Jie Lou, Yuqiu Ji, Yaojie Lu, Xianpei Han, Debing Zhang, Le Sun
Multimodal Reward Models (MM-RMs) are crucial for aligning Large Language Models (LLMs) with human preferences, particularly as LLMs increasingly interact with multimodal data.
1 code implementation • 28 Feb 2025 • Zhuoqun Li, Haiyang Yu, Xuanang Chen, Hongyu Lin, Yaojie Lu, Fei Huang, Xianpei Han, Yongbin Li, Le Sun
Designing solutions for complex engineering challenges is crucial in human production activities.
no code implementations • 24 Feb 2025 • Xueru Wen, Jie Lou, Zichao Li, Yaojie Lu, Xing Yu, Yuqiu Ji, Guohai Xu, Hongyu Lin, Ben He, Xianpei Han, Le Sun, Debing Zhang
Reward models (RMs) are crucial for aligning large language models (LLMs) with human preferences.
no code implementations • 4 Feb 2025 • Qianhao Yuan, Yanjiang Liu, Yaojie Lu, Hongyu Lin, Ben He, Xianpei Han, Le Sun
To investigate whether attention among visual tokens is necessary, we propose a new self-attention mechanism, NAAViT (\textbf{N}o \textbf{A}ttention \textbf{A}mong \textbf{Vi}sual \textbf{T}okens), which eliminates this type of attention.
1 code implementation • 3 Feb 2025 • Xinyan Guan, Jiali Zeng, Fandong Meng, Chunlei Xin, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Jie zhou
Large Language Models (LLMs) have shown remarkable potential in reasoning while they still suffer from severe factual hallucinations due to timeliness, accuracy, and coverage of parametric knowledge.
1 code implementation • 7 Jan 2025 • Hao Zheng, Xinyan Guan, Hao Kong, Jia Zheng, Hongyu Lin, Yaojie Lu, Ben He, Xianpei Han, Le Sun
Automatically generating presentations from documents is a challenging task that requires balancing content quality, visual design, and structural coherence.
no code implementations • 3 Jan 2025 • Yanjiang Liu, Shuhen Zhou, Yaojie Lu, Huijia Zhu, Weiqiang Wang, Hongyu Lin, Ben He, Xianpei Han, Le Sun
Automated red-teaming has become a crucial approach for uncovering vulnerabilities in large language models (LLMs).
no code implementations • 10 Dec 2024 • Jiawei Chen, Wentao Chen, Jing Su, Jingjing Xu, Hongyu Lin, Mengjie Ren, Yaojie Lu, Xianpei Han, Le Sun
Based on our observations, we propose the Babel Tower Hypothesis, which describes the entire process of LLMs acquiring new language capabilities.
1 code implementation • 18 Nov 2024 • Xinyan Guan, Yanjiang Liu, Xinyu Lu, Boxi Cao, Ben He, Xianpei Han, Le Sun, Jie Lou, Bowen Yu, Yaojie Lu, Hongyu Lin
The evolution of machine learning has increasingly prioritized the development of powerful models and more scalable supervision signals.
no code implementations • 28 Oct 2024 • Xinyu Lu, Xueru Wen, Yaojie Lu, Bowen Yu, Hongyu Lin, Haiyang Yu, Le Sun, Xianpei Han, Yongbin Li
After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference, enables them to achieve similar capability enhancements.
1 code implementation • 22 Oct 2024 • Hao Xiang, Bowen Yu, Hongyu Lin, Keming Lu, Yaojie Lu, Xianpei Han, Le Sun, Jingren Zhou, Junyang Lin
The key to automated alignment lies in providing learnable and accurate preference signals for preference learning without human annotation.
no code implementations • 17 Oct 2024 • Qiaoyu Tang, Le Yu, Bowen Yu, Hongyu Lin, Keming Lu, Yaojie Lu, Xianpei Han, Le Sun
Post-training has emerged as a crucial paradigm for adapting large-scale pre-trained models to various tasks, whose effects are fully reflected by delta parameters (i. e., the disparity between post-trained and pre-trained parameters).
1 code implementation • 11 Oct 2024 • Zhuoqun Li, Xuanang Chen, Haiyang Yu, Hongyu Lin, Yaojie Lu, Qiaoyu Tang, Fei Huang, Xianpei Han, Le Sun, Yongbin Li
Retrieval-augmented generation (RAG) is a key means to effectively enhance large language models (LLMs) in many knowledge-based tasks.
no code implementations • 10 Oct 2024 • Jiasheng Zheng, Hongyu Lin, Boxi Cao, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun
Evaluating the quality of documents is essential for filtering valuable content from the current massive amount of information.
1 code implementation • 9 Oct 2024 • Zichao Li, Shaojie He, Meng Liao, Xuanang Chen, Yaojie Lu, Hongyu Lin, Yanxiong Lu, Xianpei Han, Le Sun
Document logical structuring aims to extract the underlying hierarchical structure of documents, which is crucial for document intelligence.
no code implementations • 8 Oct 2024 • Xueru Wen, Jie Lou, Yaojie Lu, Hongyu Lin, Xing Yu, Xinyu Lu, Ben He, Xianpei Han, Debing Zhang, Le Sun
Reward Models (RMs) are crucial for aligning language models with human preferences.
1 code implementation • 8 Sep 2024 • Zichao Li, Aizier Abulaiti, Yaojie Lu, Xuanang Chen, Jia Zheng, Hongyu Lin, Xianpei Han, Le Sun
Document Structured Extraction (DSE) aims to extract structured content from raw documents.
no code implementations • 29 Aug 2024 • Xin Zheng, Jie Lou, Boxi Cao, Xueru Wen, Yuqiu Ji, Hongyu Lin, Yaojie Lu, Xianpei Han, Debing Zhang, Le Sun
Self-critic has become a crucial mechanism for enhancing the reasoning performance of LLMs.
no code implementations • 23 Aug 2024 • Ruiyang Xu, Jialun Cao, Yaojie Lu, Hongyu Lin, Xianpei Han, Ben He, Shing-Chi Cheung, Le Sun
However, there is an unignorable programming language bias in existing code benchmarks -- over 95% code generation benchmarks are dominated by Python, leaving the LLMs' capabilities in other programming languages such as Java and C/C++ unknown.
no code implementations • 23 Aug 2024 • Qiming Zhu, Jialun Cao, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Shing-Chi Cheung
We notice that LLMs are generally good at computation tasks while falling short on cryptography and system coding tasks.
1 code implementation • 20 Aug 2024 • Shu Chen, Xinyan Guan, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun
Manually annotating instruction data for large language models is difficult, costly, and hard to scale.
1 code implementation • 6 Aug 2024 • Boxi Cao, Mengjie Ren, Hongyu Lin, Xianpei Han, Feng Zhang, Junfeng Zhan, Le Sun
Evaluation is the baton for the development of large language models.
2 code implementations • 16 Jul 2024 • Jiasheng Zheng, Boxi Cao, Zhengzhao Ma, Ruotong Pan, Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
In recent years, researchers have proposed numerous benchmarks to evaluate the impressive coding capabilities of large language models (LLMs).
no code implementations • 18 Jun 2024 • Xueru Wen, Xinyu Lu, Xinyan Guan, Yaojie Lu, Hongyu Lin, Ben He, Xianpei Han, Le Sun
Previous learning-based methods focus on detecting knowledge boundaries and finetuning models with instance-level feedback, but they suffer from inaccurate signals due to off-policy data sampling and coarse-grained feedback.
1 code implementation • 5 Jun 2024 • Shiguang Guo, Ziliang Deng, Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
In this paper, we propose a new planning task--open grounded planning.
1 code implementation • 3 Jun 2024 • Boxi Cao, Keming Lu, Xinyu Lu, Jiawei Chen, Mengjie Ren, Hao Xiang, Peilin Liu, Yaojie Lu, Ben He, Xianpei Han, Le Sun, Hongyu Lin, Bowen Yu
Alignment is the most critical step in building large language models (LLMs) that meet human needs.
2 code implementations • 27 May 2024 • Tianshu Wang, Xiaoyang Chen, Hongyu Lin, Xuanang Chen, Xianpei Han, Hao Wang, Zhenyu Zeng, Le Sun
Based on our findings, we further design a compound entity matching framework (ComEM) that leverages the composition of multiple strategies and LLMs.
no code implementations • 23 May 2024 • Xin Men, Mingyu Xu, Bingning Wang, Qingyu Zhang, Hongyu Lin, Xianpei Han, WeiPeng Chen
We revisit the role of RoPE in LLMs and propose a novel property of long-term decay, we derive that the \textit{base of RoPE bounds context length}: there is an absolute lower bound for the base value to obtain certain context length capability.
1 code implementation • 24 Apr 2024 • Zhuoqun Li, Hongyu Lin, Tianshu Wang, Boxi Cao, Yaojie Lu, Weixiang Zhou, Hao Wang, Zhenyu Zeng, Le Sun, Xianpei Han
Linking a claim to grounded references is a critical ability to fulfill human demands for authentic and reliable information.
2 code implementations • 23 Apr 2024 • Tianshu Wang, Hongyu Lin, Xianpei Han, Xiaoyang Chen, Boxi Cao, Le Sun
Blocking is a critical step in entity resolution, and the emergence of neural network-based representation models has led to the development of dense blocking as a promising approach for exploring deep semantics in blocking.
1 code implementation • 16 Apr 2024 • Xiaoyang Chen, Ben He, Hongyu Lin, Xianpei Han, Tianshu Wang, Boxi Cao, Le Sun, Yingfei Sun
The practice of Retrieval-Augmented Generation (RAG), which integrates Large Language Models (LLMs) with retrieval systems, has become increasingly prevalent.
1 code implementation • 10 Apr 2024 • Ruotong Pan, Boxi Cao, Hongyu Lin, Xianpei Han, Jia Zheng, Sirui Wang, Xunliang Cai, Le Sun
In this paper, we propose Credibility-aware Generation (CAG), a universally applicable framework designed to mitigate the impact of flawed information in RAG.
1 code implementation • 25 Mar 2024 • Jiawei Chen, Hongyu Lin, Xianpei Han, Yaojie Lu, Shanshan Jiang, Bin Dong, Le Sun
Then a superposition instance retriever is applied to retrieve corresponding instances of these superposition concepts from large-scale text corpus.
1 code implementation • 14 Mar 2024 • Zhuoqun Li, Hongyu Lin, Yaojie Lu, Hao Xiang, Xianpei Han, Le Sun
Declarative knowledge and procedural knowledge are two key parts in meta-cognitive theory, and these two hold significant importance in pre-training and inference of LLMs.
1 code implementation • 11 Mar 2024 • Ruoxi Xu, Hongyu Lin, Xianpei Han, Le Sun, Yingfei Sun
The academic intelligence of large language models (LLMs) has made remarkable progress in recent times, but their social intelligence performance remains unclear.
1 code implementation • 6 Mar 2024 • Xin Men, Mingyu Xu, Qingyu Zhang, Bingning Wang, Hongyu Lin, Yaojie Lu, Xianpei Han, WeiPeng Chen
As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters.
1 code implementation • 28 Feb 2024 • Mengjie Ren, Boxi Cao, Hongyu Lin, Cao Liu, Xianpei Han, Ke Zeng, Guanglu Wan, Xunliang Cai, Le Sun
Instruction Fine-tuning~(IFT) is a critical phase in building large language models~(LLMs).
1 code implementation • 27 Feb 2024 • Xinyu Lu, Bowen Yu, Yaojie Lu, Hongyu Lin, Haiyang Yu, Le Sun, Xianpei Han, Yongbin Li
The alignment problem in Large Language Models (LLMs) involves adapting them to the broad spectrum of human values.
2 code implementations • 23 Feb 2024 • Qiaoyu Tang, Jiawei Chen, Zhuoqun Li, Bowen Yu, Yaojie Lu, Cheng Fu, Haiyang Yu, Hongyu Lin, Fei Huang, Ben He, Xianpei Han, Le Sun, Yongbin Li
However, current interactions between IR systems and LLMs remain limited, with LLMs merely serving as part of components within IR systems, and IR systems being constructed independently of LLMs.
1 code implementation • 23 Feb 2024 • Xin Zheng, Qiming Zhu, Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
In this paper, we seek to examine the capacity of present-day LLMs to comprehend and execute algorithms outlined in natural language.
no code implementations • 22 Feb 2024 • Ning Bian, Xianpei Han, Hongyu Lin, Yaojie Lu, Ben He, Le Sun
Building machines with commonsense has been a longstanding challenge in NLP due to the reporting bias of commonsense rules and the exposure bias of rule-based commonsense reasoning.
no code implementations • 22 Jan 2024 • Ruoxi Xu, Yingfei Sun, Mengjie Ren, Shiguang Guo, Ruotong Pan, Hongyu Lin, Le Sun, Xianpei Han
Recent advancements in artificial intelligence, particularly with the emergence of large language models (LLMs), have sparked a rethinking of artificial general intelligence possibilities.
1 code implementation • 6 Dec 2023 • Tianshu Wang, Xiaoyang Chen, Hongyu Lin, Xianpei Han, Le Sun, Hao Wang, Zhenyu Zeng
The development of Natural Language Interfaces to Databases (NLIDBs) has been greatly advanced by the advent of large language models (LLMs), which provide an intuitive way to translate natural language (NL) questions into Structured Query Language (SQL) queries.
no code implementations • 22 Nov 2023 • Xinyan Guan, Yanjiang Liu, Hongyu Lin, Yaojie Lu, Ben He, Xianpei Han, Le Sun
Incorporating factual knowledge in knowledge graph is regarded as a promising approach for mitigating the hallucination of large language models (LLMs).
1 code implementation • 19 Sep 2023 • Xin Zheng, Hongyu Lin, Xianpei Han, Le Sun
Controllable text generation is a fundamental aspect of natural language generation, with numerous methods proposed for different constraint types.
1 code implementation • 4 Sep 2023 • Jiawei Chen, Hongyu Lin, Xianpei Han, Le Sun
In this paper, we systematically investigate the impact of Retrieval-Augmented Generation on large language models.
3 code implementations • 8 Jun 2023 • Qiaoyu Tang, Ziliang Deng, Hongyu Lin, Xianpei Han, Qiao Liang, Boxi Cao, Le Sun
Existing approaches to tool learning have either primarily relied on extremely large language models, such as GPT-4, to attain generalized tool-use abilities in a zero-shot manner, or utilized supervised learning to train limited scopes of tools on compact models.
2 code implementations • 18 May 2023 • Jiawei Chen, Yaojie Lu, Hongyu Lin, Jie Lou, Wei Jia, Dai Dai, Hua Wu, Boxi Cao, Xianpei Han, Le Sun
M}$, and a new entity extractor can be implicitly constructed by applying new instruction and demonstrations to PLMs, i. e., $\mathcal{ (\lambda .
no code implementations • 16 May 2023 • Ruoxi Xu, Hongyu Lin, Xinyan Guan, Xianpei Han, Yingfei Sun, Le Sun
Understanding documents is central to many real-world tasks but remains a challenging topic.
no code implementations • 16 May 2023 • Boxi Cao, Qiaoyu Tang, Hongyu Lin, Shanshan Jiang, Bin Dong, Xianpei Han, Jiawei Chen, Tianshu Wang, Le Sun
Memory is one of the most essential cognitive functions serving as a repository of world knowledge and episodes of activities.
1 code implementation • 12 May 2023 • Jialong Tang, Hongyu Lin, Zhuoqun Li, Yaojie Lu, Xianpei Han, Le Sun
Event schema provides a conceptual, structural and formal language to represent events and model the world event knowledge.
no code implementations • 8 May 2023 • Ning Bian, Hongyu Lin, Peilin Liu, Yaojie Lu, Chunkang Zhang, Ben He, Xianpei Han, Le Sun
LLMs, as AI agents, can observe external information, which shapes their cognition and behaviors.
no code implementations • 29 Mar 2023 • Ning Bian, Xianpei Han, Le Sun, Hongyu Lin, Yaojie Lu, Ben He, Shanshan Jiang, Bin Dong
(4) Can ChatGPT effectively leverage commonsense for answering questions?
1 code implementation • 14 Mar 2023 • Boxi Cao, Hongyu Lin, Xianpei Han, Le Sun
Knowledge plays a critical role in artificial intelligence.
no code implementations • 19 Jan 2023 • Shan Wu, Chunlei Xin, Bo Chen, Xianpei Han, Le Sun
Since the meaning representations are detailed and accurate annotations which express fine-grained sequence-level semtantics, it is usually hard to train discriminative semantic parsers via Maximum Likelihood Estimation (MLE) in an autoregressive fashion.
no code implementations • 9 Jan 2023 • Jie Lou, Yaojie Lu, Dai Dai, Wei Jia, Hongyu Lin, Xianpei Han, Le Sun, Hua Wu
Based on this paradigm, we propose to universally model various IE tasks with Unified Semantic Matching (USM) framework, which introduces three unified token linking operations to model the abilities of structuring and conceptualizing.
no code implementations • 12 May 2022 • Tianshu Wang, Hongyu Lin, Cheng Fu, Xianpei Han, Le Sun, Feiyu Xiong, Hui Chen, Minlong Lu, Xiuwen Zhu
Experimental results demonstrate that the assumptions made in the previous benchmark construction process are not coincidental with the open environment, which conceal the main challenges of the task and therefore significantly overestimate the current progress of entity matching.
no code implementations • 26 Mar 2022 • Sha Yuan, Hanyu Zhao, Shuai Zhao, Jiahong Leng, Yangxiao Liang, Xiaozhi Wang, Jifan Yu, Xin Lv, Zhou Shao, Jiaao He, Yankai Lin, Xu Han, Zhenghao Liu, Ning Ding, Yongming Rao, Yizhao Gao, Liang Zhang, Ming Ding, Cong Fang, Yisen Wang, Mingsheng Long, Jing Zhang, Yinpeng Dong, Tianyu Pang, Peng Cui, Lingxiao Huang, Zheng Liang, HuaWei Shen, HUI ZHANG, Quanshi Zhang, Qingxiu Dong, Zhixing Tan, Mingxuan Wang, Shuo Wang, Long Zhou, Haoran Li, Junwei Bao, Yingwei Pan, Weinan Zhang, Zhou Yu, Rui Yan, Chence Shi, Minghao Xu, Zuobai Zhang, Guoqiang Wang, Xiang Pan, Mengjie Li, Xiaoyu Chu, Zijun Yao, Fangwei Zhu, Shulin Cao, Weicheng Xue, Zixuan Ma, Zhengyan Zhang, Shengding Hu, Yujia Qin, Chaojun Xiao, Zheni Zeng, Ganqu Cui, Weize Chen, Weilin Zhao, Yuan YAO, Peng Li, Wenzhao Zheng, Wenliang Zhao, Ziyi Wang, Borui Zhang, Nanyi Fei, Anwen Hu, Zenan Ling, Haoyang Li, Boxi Cao, Xianpei Han, Weidong Zhan, Baobao Chang, Hao Sun, Jiawen Deng, Chujie Zheng, Juanzi Li, Lei Hou, Xigang Cao, Jidong Zhai, Zhiyuan Liu, Maosong Sun, Jiwen Lu, Zhiwu Lu, Qin Jin, Ruihua Song, Ji-Rong Wen, Zhouchen Lin, LiWei Wang, Hang Su, Jun Zhu, Zhifang Sui, Jiajun Zhang, Yang Liu, Xiaodong He, Minlie Huang, Jian Tang, Jie Tang
With the rapid development of deep learning, training Big Models (BMs) for multiple downstream tasks becomes a popular paradigm.
1 code implementation • ACL 2022 • Jiawei Chen, Qing Liu, Hongyu Lin, Xianpei Han, Le Sun
In this paper, we propose a self-describing mechanism for few-shot NER, which can effectively leverage illustrative instances and precisely transfer knowledge from external resources by describing both entity types and mentions using a universal concept set.
1 code implementation • ACL 2022 • Boxi Cao, Hongyu Lin, Xianpei Han, Fangchao Liu, Le Sun
Prompt-based probing has been widely used in evaluating the abilities of pretrained language models (PLMs).
2 code implementations • ACL 2022 • Yaojie Lu, Qing Liu, Dai Dai, Xinyan Xiao, Hongyu Lin, Xianpei Han, Le Sun, Hua Wu
Information extraction suffers from its varying targets, heterogeneous structures, and demand-specific schemas.
Ranked #5 on
Aspect-Based Sentiment Analysis (ABSA)
on ASTE
(using extra training data)
no code implementations • Findings (ACL) 2022 • Ruoxi Xu, Hongyu Lin, Meng Liao, Xianpei Han, Jin Xu, Wei Tan, Yingfei Sun, Le Sun
Events are considered as the fundamental building blocks of the world.
1 code implementation • ACL 2022 • Fangchao Liu, Hongyu Lin, Xianpei Han, Boxi Cao, Le Sun
Low-shot relation extraction~(RE) aims to recognize novel relations with very few or even no samples, which is critical in real scenario application.
no code implementations • 15 Mar 2022 • Jialong Tang, Hongyu Lin, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun, Weijian Xie, Jin Xu
In this paper, we propose a new \textbf{scene-wise} paradigm for procedural text understanding, which jointly tracks states of all entities in a scene-by-scene manner.
no code implementations • 27 Dec 2021 • Yuan YAO, Qingxiu Dong, Jian Guan, Boxi Cao, Zhengyan Zhang, Chaojun Xiao, Xiaozhi Wang, Fanchao Qi, Junwei Bao, Jinran Nie, Zheni Zeng, Yuxian Gu, Kun Zhou, Xuancheng Huang, Wenhao Li, Shuhuai Ren, Jinliang Lu, Chengqiang Xu, Huadong Wang, Guoyang Zeng, Zile Zhou, Jiajun Zhang, Juanzi Li, Minlie Huang, Rui Yan, Xiaodong He, Xiaojun Wan, Xin Zhao, Xu sun, Yang Liu, Zhiyuan Liu, Xianpei Han, Erhong Yang, Zhifang Sui, Maosong Sun
We argue that for general-purpose language intelligence evaluation, the benchmark itself needs to be comprehensive and systematic.
1 code implementation • EMNLP 2021 • Lingyong Yan, Xianpei Han, Le Sun
Bootstrapping has become the mainstream method for entity set expansion.
1 code implementation • EMNLP 2021 • Jiawei Chen, Hongyu Lin, Xianpei Han, Le Sun
In this paper, we identify and solve the trigger curse problem in few-shot event detection (FSED) from a causal view.
no code implementations • EMNLP 2021 • Qing Liu, Hongyu Lin, Xinyan Xiao, Xianpei Han, Le Sun, Hua Wu
Conventional entity typing approaches are based on independent classification paradigms, which make them difficult to recognize inter-dependent, long-tailed and fine-grained entity types.
Ranked #8 on
Entity Typing
on Open Entity
no code implementations • 19 Jul 2021 • Ning Bian, Xianpei Han, Bo Chen, Hongyu Lin, Ben He, Le Sun
In this paper, we propose a new framework for unsupervised MRC.
1 code implementation • ACL 2021 • Boxi Cao, Hongyu Lin, Xianpei Han, Le Sun, Lingyong Yan, Meng Liao, Tong Xue, Jin Xu
Previous literatures show that pre-trained masked language models (MLMs) such as BERT can achieve competitive factual knowledge extraction performance on some datasets, indicating that MLMs can potentially be a reliable knowledge source.
1 code implementation • ACL 2021 • Yaojie Lu, Hongyu Lin, Jin Xu, Xianpei Han, Jialong Tang, Annan Li, Le Sun, Meng Liao, Shaoyi Chen
Event extraction is challenging due to the complex structure of event records and the semantic gap between text and event.
Ranked #3 on
Event Extraction
on ACE2005
1 code implementation • ACL 2021 • Wenkai Zhang, Hongyu Lin, Xianpei Han, Le Sun
Distant supervision tackles the data bottleneck in NER by automatically generating training instances via dictionary matching.
1 code implementation • 17 Jun 2021 • Wenkai Zhang, Hongyu Lin, Xianpei Han, Le Sun, Huidan Liu, Zhicheng Wei, Nicholas Jing Yuan
Specifically, during neural network training, we naturally model the noise samples in each batch following a hypergeometric distribution parameterized by the noise-rate.
no code implementations • ACL 2021 • Fangchao Liu, Lingyong Yan, Hongyu Lin, Xianpei Han, Le Sun
Open relation extraction aims to cluster relation instances referring to the same underlying relation, which is a critical step for general relation extraction.
1 code implementation • ACL 2021 • Jialong Tang, Hongyu Lin, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun, Weijian Xie, Jin Xu
Current event-centric knowledge graphs highly rely on explicit connectives to mine relations between events.
no code implementations • ACL 2021 • Shan Wu, Bo Chen, Chunlei Xin, Xianpei Han, Le Sun, Weipeng Zhang, Jiansong Chen, Fan Yang, Xunliang Cai
During synchronous decoding: the utterance paraphrasing is constrained by the structure of the logical form, therefore the canonical utterance can be paraphrased controlledly; the semantic decoding is guided by the semantics of the canonical utterance, therefore its logical form can be generated unsupervisedly.
no code implementations • 17 Apr 2021 • Xiaoyang Chen, Kai Hui, Ben He, Xianpei Han, Le Sun, Zheng Ye
BERT-based text ranking models have dramatically advanced the state-of-the-art in ad-hoc retrieval, wherein most models tend to consider individual query-document pairs independently.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jialong Tang, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Xinyan Xiao, Hua Wu
One of the biggest bottlenecks in building accurate, high coverage neural open IE systems is the need for large labelled corpora.
no code implementations • 4 Jan 2021 • Ning Bian, Xianpei Han, Bo Chen, Le Sun
Experiments show that: (1) Our knowledge-to-text framework is effective and achieves state-of-the-art performance on CommonsenseQA dataset, providing a simple and strong knowledge-enhanced baseline for CQA; (2) The potential of knowledge is still far from being fully exploited in CQA -- there is a significant performance gap from current models to our models with golden knowledge; and (3) Context-sensitive knowledge selection, heterogeneous knowledge exploitation, and commonsense-rich language models are promising CQA directions.
1 code implementation • 8 Dec 2020 • Lingyong Yan, Xianpei Han, Le Sun, Fangchao Liu, Ning Bian
By re-organizing all sentences about an entity as a document and extracting relations via querying the document with relation-specific questions, the document-based DS paradigm can simultaneously encode and exploit all sentence-level, inter-sentence-level, and entity-level evidence.
Ranked #1 on
Relationship Extraction (Distant Supervised)
on NYT
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Lingyong Yan, Xianpei Han, Ben He, Le Sun
Bootstrapping for entity set expansion (ESE) has been studied for a long period, which expands new entities using only a few seed entities as supervision.
1 code implementation • 17 Sep 2020 • Yaojie Lu, Hongyu Lin, Jialong Tang, Xianpei Han, Le Sun
Traditional event coreference systems usually rely on pipeline framework and hand-crafted features, which often face error propagation problem and have poor generalization ability.
1 code implementation • SEMEVAL 2020 • Yaojie Lu, Annan Li, Hongyu Lin, Xianpei Han, Le Sun
ISCAS participated in two subtasks of SemEval 2020 Task 5: detecting counterfactual statements and detecting antecedent and consequence.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Zhi Zheng, Kai Hui, Ben He, Xianpei Han, Le Sun, Andrew Yates
Query expansion aims to mitigate the mismatch between the language used in a query and in a document.
no code implementations • EMNLP 2020 • Hongyu Lin, Yaojie Lu, Jialong Tang, Xianpei Han, Le Sun, Zhicheng Wei, Nicholas Jing Yuan
Specifically, we erase name regularity, mention coverage and context diversity respectively from the benchmarks, in order to explore their impact on the generalization ability of models.
no code implementations • 9 Mar 2020 • Xianpei Han, Zhichun Wang, Jiangtao Zhang, Qinghua Wen, Wenqi Li, Buzhou Tang, Qi. Wang, Zhifan Feng, Yang Zhang, Yajuan Lu, Haitao Wang, Wenliang Chen, Hao Shao, Yubo Chen, Kang Liu, Jun Zhao, Taifeng Wang, Kezun Zhang, Meng Wang, Yinlin Jiang, Guilin Qi, Lei Zou, Sen Hu, Minhao Zhang, Yinnian Lin
Knowledge graph models world knowledge as concepts, entities, and the relationships between them, which has been widely used in many real-world tasks.
2 code implementations • 20 Nov 2019 • Chaojun Xiao, Haoxi Zhong, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Tianyang Zhang, Xianpei Han, Zhen Hu, Heng Wang, Jianfeng Xu
In this paper, we introduce CAIL2019-SCM, Chinese AI and Law 2019 Similar Case Matching dataset.
no code implementations • IJCNLP 2019 • Bo An, Chen Bo, Xianpei Han, Le Sun
Semantic parsing aims to map natural language utterances into structured meaning representations.
no code implementations • IJCNLP 2019 • Lingyong Yan, Xianpei Han, Le Sun, Ben He
Bootstrapping for Entity Set Expansion (ESE) aims at iteratively acquiring new instances of a specific target category.
no code implementations • IJCNLP 2019 • Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun, Bin Dong, Shanshan Jiang
Current region-based NER models only rely on fully-annotated training data to learn effective region encoder, which often face the training data bottleneck.
1 code implementation • ACL 2019 • Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun
Event detection systems rely on discrimination knowledge to distinguish ambiguous trigger words and generalization knowledge to detect unseen/sparse trigger words.
1 code implementation • ACL 2019 • Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
In supervised event detection, most of the mislabeling occurs between a small number of confusing type pairs, including trigger-NIL pairs and sibling sub-types of the same coarse type.
1 code implementation • ACL 2019 • Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
In this paper, we propose to resolve this problem by modeling and leveraging the head-driven phrase structures of entity mentions, i. e., although a mention can nest other mentions, they will not share the same head word.
Ranked #7 on
Nested Mention Recognition
on ACE 2005
no code implementations • ACL 2016 • Bo Chen, Le Sun, Xianpei Han, Bo An
A major challenge of semantic parsing is the vocabulary mismatch problem between natural language and target ontology.
2 code implementations • 13 Oct 2018 • Haoxi Zhong, Chaojun Xiao, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Yansong Feng, Xianpei Han, Zhen Hu, Heng Wang, Jianfeng Xu
In this paper, we give an overview of the Legal Judgment Prediction (LJP) competition at Chinese AI and Law challenge (CAIL2018).
1 code implementation • ACL 2018 • Bo Chen, Le Sun, Xianpei Han
This paper proposes a neural semantic parsing approach -- Sequence-to-Action, which models semantic parsing as an end-to-end semantic graph generation process.
no code implementations • COLING 2018 • Bo Chen, Bo An, Le Sun, Xianpei Han
Semantic parsers critically rely on accurate and high-coverage lexicons.
no code implementations • COLING 2018 • Bo An, Xianpei Han, Le Sun
Word composition is a promising technique for representation learning of large linguistic units (e. g., phrases, sentences and documents).
3 code implementations • 4 Jul 2018 • Chaojun Xiao, Haoxi Zhong, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Yansong Feng, Xianpei Han, Zhen Hu, Heng Wang, Jianfeng Xu
In this paper, we introduce the \textbf{C}hinese \textbf{AI} and \textbf{L}aw challenge dataset (CAIL2018), the first large-scale Chinese legal dataset for judgment prediction.
no code implementations • NAACL 2018 • Bo An, Bo Chen, Xianpei Han, Le Sun
Previous representation learning techniques for knowledge graph representation usually represent the same entity or relation in different triples with the same representation, without considering the ambiguity of relations and entities.
1 code implementation • ACL 2018 • Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
Neural network based models commonly regard event detection as a word-wise classification task, which suffer from the mismatch problem between words and event triggers, especially in languages without natural word delimiters such as Chinese.
1 code implementation • ACL 2018 • Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun
This paper focuses on detection tasks in information extraction, where positive instances are sparsely distributed and models are usually evaluated using F-measure on positive classes.
no code implementations • 16 Jan 2018 • Jinsong Su, Shan Wu, Deyi Xiong, Yaojie Lu, Xianpei Han, Biao Zhang
Partially inspired by successful applications of variational recurrent neural networks, we propose a novel variational recurrent neural machine translation (VRNMT) model in this paper.
no code implementations • EMNLP 2017 • Hongyu Lin, Le Sun, Xianpei Han
Then we propose a multi-knowledge reasoning model, which selects inference rules for a specific reasoning context using attention mechanism, and reasons by summarizing all valid inference rules.
no code implementations • COLING 2016 • Xianpei Han, Le Sun
Firstly, these methods are mostly context-insensitive, cannot accurately measure the similarity between two predicates in a specific context.