Search Results for author: Xianpei Han

Found 116 papers, 60 papers with code

Large Language Models Often Say One Thing and Do Another

1 code implementation10 Mar 2025 Ruoxi Xu, Hongyu Lin, Xianpei Han, Jia Zheng, Weixiang Zhou, Le Sun, Yingfei Sun

As large language models (LLMs) increasingly become central to various applications and interact with diverse user populations, ensuring their reliable and consistent performance is becoming more important.

Word Alignment

The Devil Is in the Details: Tackling Unimodal Spurious Correlations for Generalizable Multimodal Reward Models

no code implementations5 Mar 2025 Zichao Li, Xueru Wen, Jie Lou, Yuqiu Ji, Yaojie Lu, Xianpei Han, Debing Zhang, Le Sun

Multimodal Reward Models (MM-RMs) are crucial for aligning Large Language Models (LLMs) with human preferences, particularly as LLMs increasingly interact with multimodal data.

SAISA: Towards Multimodal Large Language Models with Both Training and Inference Efficiency

no code implementations4 Feb 2025 Qianhao Yuan, Yanjiang Liu, Yaojie Lu, Hongyu Lin, Ben He, Xianpei Han, Le Sun

To investigate whether attention among visual tokens is necessary, we propose a new self-attention mechanism, NAAViT (\textbf{N}o \textbf{A}ttention \textbf{A}mong \textbf{Vi}sual \textbf{T}okens), which eliminates this type of attention.

DeepRAG: Thinking to Retrieval Step by Step for Large Language Models

1 code implementation3 Feb 2025 Xinyan Guan, Jiali Zeng, Fandong Meng, Chunlei Xin, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Jie zhou

Large Language Models (LLMs) have shown remarkable potential in reasoning while they still suffer from severe factual hallucinations due to timeliness, accuracy, and coverage of parametric knowledge.

RAG Retrieval

PPTAgent: Generating and Evaluating Presentations Beyond Text-to-Slides

1 code implementation7 Jan 2025 Hao Zheng, Xinyan Guan, Hao Kong, Jia Zheng, Hongyu Lin, Yaojie Lu, Ben He, Xianpei Han, Le Sun

Automatically generating presentations from documents is a challenging task that requires balancing content quality, visual design, and structural coherence.

Auto-RT: Automatic Jailbreak Strategy Exploration for Red-Teaming Large Language Models

no code implementations3 Jan 2025 Yanjiang Liu, Shuhen Zhou, Yaojie Lu, Huijia Zhu, Weiqiang Wang, Hongyu Lin, Ben He, Xianpei Han, Le Sun

Automated red-teaming has become a crucial approach for uncovering vulnerabilities in large language models (LLMs).

Red Teaming

Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering

1 code implementation18 Nov 2024 Xinyan Guan, Yanjiang Liu, Xinyu Lu, Boxi Cao, Ben He, Xianpei Han, Le Sun, Jie Lou, Bowen Yu, Yaojie Lu, Hongyu Lin

The evolution of machine learning has increasingly prioritized the development of powerful models and more scalable supervision signals.

Transferable Post-training via Inverse Value Learning

no code implementations28 Oct 2024 Xinyu Lu, Xueru Wen, Yaojie Lu, Bowen Yu, Hongyu Lin, Haiyang Yu, Le Sun, Xianpei Han, Yongbin Li

After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference, enables them to achieve similar capability enhancements.

Aligning Large Language Models via Self-Steering Optimization

1 code implementation22 Oct 2024 Hao Xiang, Bowen Yu, Hongyu Lin, Keming Lu, Yaojie Lu, Xianpei Han, Le Sun, Jingren Zhou, Junyang Lin

The key to automated alignment lies in providing learnable and accurate preference signals for preference learning without human annotation.

A Unified View of Delta Parameter Editing in Post-Trained Large-Scale Models

no code implementations17 Oct 2024 Qiaoyu Tang, Le Yu, Bowen Yu, Hongyu Lin, Keming Lu, Yaojie Lu, Xianpei Han, Le Sun

Post-training has emerged as a crucial paradigm for adapting large-scale pre-trained models to various tasks, whose effects are fully reflected by delta parameters (i. e., the disparity between post-trained and pre-trained parameters).

Quantization

StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization

1 code implementation11 Oct 2024 Zhuoqun Li, Xuanang Chen, Haiyang Yu, Hongyu Lin, Yaojie Lu, Qiaoyu Tang, Fei Huang, Xianpei Han, Le Sun, Yongbin Li

Retrieval-augmented generation (RAG) is a key means to effectively enhance large language models (LLMs) in many knowledge-based tasks.

RAG

Multi-Facet Counterfactual Learning for Content Quality Evaluation

no code implementations10 Oct 2024 Jiasheng Zheng, Hongyu Lin, Boxi Cao, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun

Evaluating the quality of documents is essential for filtering valuable content from the current massive amount of information.

Contrastive Learning counterfactual

Seg2Act: Global Context-aware Action Generation for Document Logical Structuring

1 code implementation9 Oct 2024 Zichao Li, Shaojie He, Meng Liao, Xuanang Chen, Yaojie Lu, Hongyu Lin, Yanxiong Lu, Xianpei Han, Le Sun

Document logical structuring aims to extract the underlying hierarchical structure of documents, which is crucial for document intelligence.

Action Generation Transfer Learning

READoc: A Unified Benchmark for Realistic Document Structured Extraction

1 code implementation8 Sep 2024 Zichao Li, Aizier Abulaiti, Yaojie Lu, Xuanang Chen, Jia Zheng, Hongyu Lin, Xianpei Han, Le Sun

Document Structured Extraction (DSE) aims to extract structured content from raw documents.

CRUXEval-X: A Benchmark for Multilingual Code Reasoning, Understanding and Execution

no code implementations23 Aug 2024 Ruiyang Xu, Jialun Cao, Yaojie Lu, Hongyu Lin, Xianpei Han, Ben He, Shing-Chi Cheung, Le Sun

However, there is an unignorable programming language bias in existing code benchmarks -- over 95% code generation benchmarks are dominated by Python, leaving the LLMs' capabilities in other programming languages such as Java and C/C++ unknown.

Code Generation HumanEval

DOMAINEVAL: An Auto-Constructed Benchmark for Multi-Domain Code Generation

no code implementations23 Aug 2024 Qiming Zhu, Jialun Cao, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Shing-Chi Cheung

We notice that LLMs are generally good at computation tasks while falling short on cryptography and system coding tasks.

Code Generation HumanEval

REInstruct: Building Instruction Data from Unlabeled Corpus

1 code implementation20 Aug 2024 Shu Chen, Xinyan Guan, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun

Manually annotating instruction data for large language models is difficult, costly, and hard to scale.

Beyond Correctness: Benchmarking Multi-dimensional Code Generation for Large Language Models

2 code implementations16 Jul 2024 Jiasheng Zheng, Boxi Cao, Zhengzhao Ma, Ruotong Pan, Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun

In recent years, researchers have proposed numerous benchmarks to evaluate the impressive coding capabilities of large language models (LLMs).

Benchmarking Code Generation

On-Policy Fine-grained Knowledge Feedback for Hallucination Mitigation

no code implementations18 Jun 2024 Xueru Wen, Xinyu Lu, Xinyan Guan, Yaojie Lu, Hongyu Lin, Ben He, Xianpei Han, Le Sun

Previous learning-based methods focus on detecting knowledge boundaries and finetuning models with instance-level feedback, but they suffer from inaccurate signals due to off-policy data sampling and coarse-grained feedback.

Hallucination Response Generation

Towards Scalable Automated Alignment of LLMs: A Survey

1 code implementation3 Jun 2024 Boxi Cao, Keming Lu, Xinyu Lu, Jiawei Chen, Mengjie Ren, Hao Xiang, Peilin Liu, Yaojie Lu, Ben He, Xianpei Han, Le Sun, Hongyu Lin, Bowen Yu

Alignment is the most critical step in building large language models (LLMs) that meet human needs.

Survey

Match, Compare, or Select? An Investigation of Large Language Models for Entity Matching

2 code implementations27 May 2024 Tianshu Wang, Xiaoyang Chen, Hongyu Lin, Xuanang Chen, Xianpei Han, Hao Wang, Zhenyu Zeng, Le Sun

Based on our findings, we further design a compound entity matching framework (ComEM) that leverages the composition of multiple strategies and LLMs.

Entity Resolution

Base of RoPE Bounds Context Length

no code implementations23 May 2024 Xin Men, Mingyu Xu, Bingning Wang, Qingyu Zhang, Hongyu Lin, Xianpei Han, WeiPeng Chen

We revisit the role of RoPE in LLMs and propose a novel property of long-term decay, we derive that the \textit{base of RoPE bounds context length}: there is an absolute lower bound for the base value to obtain certain context length capability.

Position

Towards Universal Dense Blocking for Entity Resolution

2 code implementations23 Apr 2024 Tianshu Wang, Hongyu Lin, Xianpei Han, Xiaoyang Chen, Boxi Cao, Le Sun

Blocking is a critical step in entity resolution, and the emergence of neural network-based representation models has led to the development of dense blocking as a promising approach for exploring deep semantics in blocking.

Blocking Contrastive Learning

Spiral of Silence: How is Large Language Model Killing Information Retrieval? -- A Case Study on Open Domain Question Answering

1 code implementation16 Apr 2024 Xiaoyang Chen, Ben He, Hongyu Lin, Xianpei Han, Tianshu Wang, Boxi Cao, Le Sun, Yingfei Sun

The practice of Retrieval-Augmented Generation (RAG), which integrates Large Language Models (LLMs) with retrieval systems, has become increasingly prevalent.

Information Retrieval Language Modeling +4

Not All Contexts Are Equal: Teaching LLMs Credibility-aware Generation

1 code implementation10 Apr 2024 Ruotong Pan, Boxi Cao, Hongyu Lin, Xianpei Han, Jia Zheng, Sirui Wang, Xunliang Cai, Le Sun

In this paper, we propose Credibility-aware Generation (CAG), a universally applicable framework designed to mitigate the impact of flawed information in RAG.

All RAG +1

Few-shot Named Entity Recognition via Superposition Concept Discrimination

1 code implementation25 Mar 2024 Jiawei Chen, Hongyu Lin, Xianpei Han, Yaojie Lu, Shanshan Jiang, Bin Dong, Le Sun

Then a superposition instance retriever is applied to retrieve corresponding instances of these superposition concepts from large-scale text corpus.

Active Learning few-shot-ner +4

Meta-Cognitive Analysis: Evaluating Declarative and Procedural Knowledge in Datasets and Large Language Models

1 code implementation14 Mar 2024 Zhuoqun Li, Hongyu Lin, Yaojie Lu, Hao Xiang, Xianpei Han, Le Sun

Declarative knowledge and procedural knowledge are two key parts in meta-cognitive theory, and these two hold significant importance in pre-training and inference of LLMs.

Academically intelligent LLMs are not necessarily socially intelligent

1 code implementation11 Mar 2024 Ruoxi Xu, Hongyu Lin, Xianpei Han, Le Sun, Yingfei Sun

The academic intelligence of large language models (LLMs) has made remarkable progress in recent times, but their social intelligence performance remains unclear.

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect

1 code implementation6 Mar 2024 Xin Men, Mingyu Xu, Qingyu Zhang, Bingning Wang, Hongyu Lin, Yaojie Lu, Xianpei Han, WeiPeng Chen

As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters.

Quantization

SoFA: Shielded On-the-fly Alignment via Priority Rule Following

1 code implementation27 Feb 2024 Xinyu Lu, Bowen Yu, Yaojie Lu, Hongyu Lin, Haiyang Yu, Le Sun, Xianpei Han, Yongbin Li

The alignment problem in Large Language Models (LLMs) involves adapting them to the broad spectrum of human values.

Diversity

Self-Retrieval: End-to-End Information Retrieval with One Large Language Model

2 code implementations23 Feb 2024 Qiaoyu Tang, Jiawei Chen, Zhuoqun Li, Bowen Yu, Yaojie Lu, Cheng Fu, Haiyang Yu, Hongyu Lin, Fei Huang, Ben He, Xianpei Han, Le Sun, Yongbin Li

However, current interactions between IR systems and LLMs remain limited, with LLMs merely serving as part of components within IR systems, and IR systems being constructed independently of LLMs.

Information Retrieval Language Modeling +4

Executing Natural Language-Described Algorithms with Large Language Models: An Investigation

1 code implementation23 Feb 2024 Xin Zheng, Qiming Zhu, Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun

In this paper, we seek to examine the capacity of present-day LLMs to comprehend and execute algorithms outlined in natural language.

Natural Language Understanding

Rule or Story, Which is a Better Commonsense Expression for Talking with Large Language Models?

no code implementations22 Feb 2024 Ning Bian, Xianpei Han, Hongyu Lin, Yaojie Lu, Ben He, Le Sun

Building machines with commonsense has been a longstanding challenge in NLP due to the reporting bias of commonsense rules and the exposure bias of rule-based commonsense reasoning.

AI for social science and social science of AI: A Survey

no code implementations22 Jan 2024 Ruoxi Xu, Yingfei Sun, Mengjie Ren, Shiguang Guo, Ruotong Pan, Hongyu Lin, Le Sun, Xianpei Han

Recent advancements in artificial intelligence, particularly with the emergence of large language models (LLMs), have sparked a rethinking of artificial general intelligence possibilities.

DBCopilot: Natural Language Querying over Massive Databases via Schema Routing

1 code implementation6 Dec 2023 Tianshu Wang, Xiaoyang Chen, Hongyu Lin, Xianpei Han, Le Sun, Hao Wang, Zhenyu Zeng

The development of Natural Language Interfaces to Databases (NLIDBs) has been greatly advanced by the advent of large language models (LLMs), which provide an intuitive way to translate natural language (NL) questions into Structured Query Language (SQL) queries.

Navigate Question Generation +2

Mitigating Large Language Model Hallucinations via Autonomous Knowledge Graph-based Retrofitting

no code implementations22 Nov 2023 Xinyan Guan, Yanjiang Liu, Hongyu Lin, Yaojie Lu, Ben He, Xianpei Han, Le Sun

Incorporating factual knowledge in knowledge graph is regarded as a promising approach for mitigating the hallucination of large language models (LLMs).

Hallucination Language Modeling +2

Toward Unified Controllable Text Generation via Regular Expression Instruction

1 code implementation19 Sep 2023 Xin Zheng, Hongyu Lin, Xianpei Han, Le Sun

Controllable text generation is a fundamental aspect of natural language generation, with numerous methods proposed for different constraint types.

In-Context Learning Text Generation

Benchmarking Large Language Models in Retrieval-Augmented Generation

1 code implementation4 Sep 2023 Jiawei Chen, Hongyu Lin, Xianpei Han, Le Sun

In this paper, we systematically investigate the impact of Retrieval-Augmented Generation on large language models.

Benchmarking counterfactual +3

ToolAlpaca: Generalized Tool Learning for Language Models with 3000 Simulated Cases

3 code implementations8 Jun 2023 Qiaoyu Tang, Ziliang Deng, Hongyu Lin, Xianpei Han, Qiao Liang, Boxi Cao, Le Sun

Existing approaches to tool learning have either primarily relied on extremely large language models, such as GPT-4, to attain generalized tool-use abilities in a zero-shot manner, or utilized supervised learning to train limited scopes of tools on compact models.

Learning In-context Learning for Named Entity Recognition

2 code implementations18 May 2023 Jiawei Chen, Yaojie Lu, Hongyu Lin, Jie Lou, Wei Jia, Dai Dai, Hua Wu, Boxi Cao, Xianpei Han, Le Sun

M}$, and a new entity extractor can be implicitly constructed by applying new instruction and demonstrations to PLMs, i. e., $\mathcal{ (\lambda .

Diversity few-shot-ner +5

Retentive or Forgetful? Diving into the Knowledge Memorizing Mechanism of Language Models

no code implementations16 May 2023 Boxi Cao, Qiaoyu Tang, Hongyu Lin, Shanshan Jiang, Bin Dong, Xianpei Han, Jiawei Chen, Tianshu Wang, Le Sun

Memory is one of the most essential cognitive functions serving as a repository of world knowledge and episodes of activities.

World Knowledge

Harvesting Event Schemas from Large Language Models

1 code implementation12 May 2023 Jialong Tang, Hongyu Lin, Zhuoqun Li, Yaojie Lu, Xianpei Han, Le Sun

Event schema provides a conceptual, structural and formal language to represent events and model the world event knowledge.

Diversity

The Life Cycle of Knowledge in Big Language Models: A Survey

1 code implementation14 Mar 2023 Boxi Cao, Hongyu Lin, Xianpei Han, Le Sun

Knowledge plays a critical role in artificial intelligence.

Semantic-aware Contrastive Learning for More Accurate Semantic Parsing

no code implementations19 Jan 2023 Shan Wu, Chunlei Xin, Bo Chen, Xianpei Han, Le Sun

Since the meaning representations are detailed and accurate annotations which express fine-grained sequence-level semtantics, it is usually hard to train discriminative semantic parsers via Maximum Likelihood Estimation (MLE) in an autoregressive fashion.

Contrastive Learning Semantic Parsing

Universal Information Extraction as Unified Semantic Matching

no code implementations9 Jan 2023 Jie Lou, Yaojie Lu, Dai Dai, Wei Jia, Hongyu Lin, Xianpei Han, Le Sun, Hua Wu

Based on this paradigm, we propose to universally model various IE tasks with Unified Semantic Matching (USM) framework, which introduces three unified token linking operations to model the abilities of structuring and conceptualizing.

Diversity

Bridging the Gap between Reality and Ideality of Entity Matching: A Revisiting and Benchmark Re-Construction

no code implementations12 May 2022 Tianshu Wang, Hongyu Lin, Cheng Fu, Xianpei Han, Le Sun, Feiyu Xiong, Hui Chen, Minlong Lu, Xiuwen Zhu

Experimental results demonstrate that the assumptions made in the previous benchmark construction process are not coincidental with the open environment, which conceal the main challenges of the task and therefore significantly overestimate the current progress of entity matching.

Entity Resolution

Few-shot Named Entity Recognition with Self-describing Networks

1 code implementation ACL 2022 Jiawei Chen, Qing Liu, Hongyu Lin, Xianpei Han, Le Sun

In this paper, we propose a self-describing mechanism for few-shot NER, which can effectively leverage illustrative instances and precisely transfer knowledge from external resources by describing both entity types and mentions using a universal concept set.

Few-shot NER Named Entity Recognition

Can Prompt Probe Pretrained Language Models? Understanding the Invisible Risks from a Causal View

1 code implementation ACL 2022 Boxi Cao, Hongyu Lin, Xianpei Han, Fangchao Liu, Le Sun

Prompt-based probing has been widely used in evaluating the abilities of pretrained language models (PLMs).

Pre-training to Match for Unified Low-shot Relation Extraction

1 code implementation ACL 2022 Fangchao Liu, Hongyu Lin, Xianpei Han, Boxi Cao, Le Sun

Low-shot relation extraction~(RE) aims to recognize novel relations with very few or even no samples, which is critical in real scenario application.

Meta-Learning Relation +2

Procedural Text Understanding via Scene-Wise Evolution

no code implementations15 Mar 2022 Jialong Tang, Hongyu Lin, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun, Weijian Xie, Jin Xu

In this paper, we propose a new \textbf{scene-wise} paradigm for procedural text understanding, which jointly tracks states of all entities in a scene-by-scene manner.

Procedural Text Understanding

Honey or Poison? Solving the Trigger Curse in Few-shot Event Detection via Causal Intervention

1 code implementation EMNLP 2021 Jiawei Chen, Hongyu Lin, Xianpei Han, Le Sun

In this paper, we identify and solve the trigger curse problem in few-shot event detection (FSED) from a causal view.

Event Detection

Fine-grained Entity Typing via Label Reasoning

no code implementations EMNLP 2021 Qing Liu, Hongyu Lin, Xinyan Xiao, Xianpei Han, Le Sun, Hua Wu

Conventional entity typing approaches are based on independent classification paradigms, which make them difficult to recognize inter-dependent, long-tailed and fine-grained entity types.

Attribute Entity Typing

Knowledgeable or Educated Guess? Revisiting Language Models as Knowledge Bases

1 code implementation ACL 2021 Boxi Cao, Hongyu Lin, Xianpei Han, Le Sun, Lingyong Yan, Meng Liao, Tong Xue, Jin Xu

Previous literatures show that pre-trained masked language models (MLMs) such as BERT can achieve competitive factual knowledge extraction performance on some datasets, indicating that MLMs can potentially be a reliable knowledge source.

Denoising Distantly Supervised Named Entity Recognition via a Hypergeometric Probabilistic Model

1 code implementation17 Jun 2021 Wenkai Zhang, Hongyu Lin, Xianpei Han, Le Sun, Huidan Liu, Zhicheng Wei, Nicholas Jing Yuan

Specifically, during neural network training, we naturally model the noise samples in each batch following a hypergeometric distribution parameterized by the noise-rate.

Denoising named-entity-recognition +2

Element Intervention for Open Relation Extraction

no code implementations ACL 2021 Fangchao Liu, Lingyong Yan, Hongyu Lin, Xianpei Han, Le Sun

Open relation extraction aims to cluster relation instances referring to the same underlying relation, which is a critical step for general relation extraction.

Relation Relation Extraction

From Paraphrasing to Semantic Parsing: Unsupervised Semantic Parsing via Synchronous Semantic Decoding

no code implementations ACL 2021 Shan Wu, Bo Chen, Chunlei Xin, Xianpei Han, Le Sun, Weipeng Zhang, Jiansong Chen, Fan Yang, Xunliang Cai

During synchronous decoding: the utterance paraphrasing is constrained by the structure of the logical form, therefore the canonical utterance can be paraphrased controlledly; the semantic decoding is guided by the semantics of the canonical utterance, therefore its logical form can be generated unsupervisedly.

Form Unsupervised semantic parsing

Co-BERT: A Context-Aware BERT Retrieval Model Incorporating Local and Query-specific Context

no code implementations17 Apr 2021 Xiaoyang Chen, Kai Hui, Ben He, Xianpei Han, Le Sun, Zheng Ye

BERT-based text ranking models have dramatically advanced the state-of-the-art in ad-hoc retrieval, wherein most models tend to consider individual query-document pairs independently.

Learning-To-Rank Re-Ranking +1

Benchmarking Knowledge-Enhanced Commonsense Question Answering via Knowledge-to-Text Transformation

no code implementations4 Jan 2021 Ning Bian, Xianpei Han, Bo Chen, Le Sun

Experiments show that: (1) Our knowledge-to-text framework is effective and achieves state-of-the-art performance on CommonsenseQA dataset, providing a simple and strong knowledge-enhanced baseline for CQA; (2) The potential of knowledge is still far from being fully exploited in CQA -- there is a significant performance gap from current models to our models with golden knowledge; and (3) Context-sensitive knowledge selection, heterogeneous knowledge exploitation, and commonsense-rich language models are promising CQA directions.

Benchmarking Question Answering

From Bag of Sentences to Document: Distantly Supervised Relation Extraction via Machine Reading Comprehension

1 code implementation8 Dec 2020 Lingyong Yan, Xianpei Han, Le Sun, Fangchao Liu, Ning Bian

By re-organizing all sentences about an entity as a document and extracting relations via querying the document with relation-specific questions, the document-based DS paradigm can simultaneously encode and exploit all sentence-level, inter-sentence-level, and entity-level evidence.

Denoising Machine Reading Comprehension +3

Global Bootstrapping Neural Network for Entity Set Expansion

1 code implementation Findings of the Association for Computational Linguistics 2020 Lingyong Yan, Xianpei Han, Ben He, Le Sun

Bootstrapping for entity set expansion (ESE) has been studied for a long period, which expands new entities using only a few seed entities as supervision.

Decoder

End-to-End Neural Event Coreference Resolution

1 code implementation17 Sep 2020 Yaojie Lu, Hongyu Lin, Jialong Tang, Xianpei Han, Le Sun

Traditional event coreference systems usually rely on pipeline framework and hand-crafted features, which often face error propagation problem and have poor generalization ability.

coreference-resolution Event Coreference Resolution +1

ISCAS at SemEval-2020 Task 5: Pre-trained Transformers for Counterfactual Statement Modeling

1 code implementation SEMEVAL 2020 Yaojie Lu, Annan Li, Hongyu Lin, Xianpei Han, Le Sun

ISCAS participated in two subtasks of SemEval 2020 Task 5: detecting counterfactual statements and detecting antecedent and consequence.

counterfactual Question Answering

A Rigorous Study on Named Entity Recognition: Can Fine-tuning Pretrained Model Lead to the Promised Land?

no code implementations EMNLP 2020 Hongyu Lin, Yaojie Lu, Jialong Tang, Xianpei Han, Le Sun, Zhicheng Wei, Nicholas Jing Yuan

Specifically, we erase name regularity, mention coverage and context diversity respectively from the benchmarks, in order to explore their impact on the generalization ability of models.

Diversity named-entity-recognition +2

EUSP: An Easy-to-Use Semantic Parsing PlatForm

no code implementations IJCNLP 2019 Bo An, Chen Bo, Xianpei Han, Le Sun

Semantic parsing aims to map natural language utterances into structured meaning representations.

Semantic Parsing

Learning to Bootstrap for Entity Set Expansion

no code implementations IJCNLP 2019 Lingyong Yan, Xianpei Han, Le Sun, Ben He

Bootstrapping for Entity Set Expansion (ESE) aims at iteratively acquiring new instances of a specific target category.

Gazetteer-Enhanced Attentive Neural Networks for Named Entity Recognition

no code implementations IJCNLP 2019 Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun, Bin Dong, Shanshan Jiang

Current region-based NER models only rely on fully-annotated training data to learn effective region encoder, which often face the training data bottleneck.

named-entity-recognition Named Entity Recognition +1

Distilling Discrimination and Generalization Knowledge for Event Detection via Delta-Representation Learning

1 code implementation ACL 2019 Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun

Event detection systems rely on discrimination knowledge to distinguish ambiguous trigger words and generalization knowledge to detect unseen/sparse trigger words.

Event Detection Representation Learning

Cost-sensitive Regularization for Label Confusion-aware Event Detection

1 code implementation ACL 2019 Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun

In supervised event detection, most of the mislabeling occurs between a small number of confusing type pairs, including trigger-NIL pairs and sibling sub-types of the same coarse type.

Event Detection Vocal Bursts Type Prediction

Sequence-to-Nuggets: Nested Entity Mention Detection via Anchor-Region Networks

1 code implementation ACL 2019 Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun

In this paper, we propose to resolve this problem by modeling and leveraging the head-driven phrase structures of entity mentions, i. e., although a mention can nest other mentions, they will not share the same head word.

NER Nested Mention Recognition +1

Sentence Rewriting for Semantic Parsing

no code implementations ACL 2016 Bo Chen, Le Sun, Xianpei Han, Bo An

A major challenge of semantic parsing is the vocabulary mismatch problem between natural language and target ontology.

Form Semantic Parsing +2

Overview of CAIL2018: Legal Judgment Prediction Competition

2 code implementations13 Oct 2018 Haoxi Zhong, Chaojun Xiao, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Yansong Feng, Xianpei Han, Zhen Hu, Heng Wang, Jianfeng Xu

In this paper, we give an overview of the Legal Judgment Prediction (LJP) competition at Chinese AI and Law challenge (CAIL2018).

Prediction

Sequence-to-Action: End-to-End Semantic Graph Generation for Semantic Parsing

1 code implementation ACL 2018 Bo Chen, Le Sun, Xianpei Han

This paper proposes a neural semantic parsing approach -- Sequence-to-Action, which models semantic parsing as an end-to-end semantic graph generation process.

Graph Generation Representation Learning +2

Model-Free Context-Aware Word Composition

no code implementations COLING 2018 Bo An, Xianpei Han, Le Sun

Word composition is a promising technique for representation learning of large linguistic units (e. g., phrases, sentences and documents).

Dimensionality Reduction Learning Word Embeddings +5

CAIL2018: A Large-Scale Legal Dataset for Judgment Prediction

3 code implementations4 Jul 2018 Chaojun Xiao, Haoxi Zhong, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Yansong Feng, Xianpei Han, Zhen Hu, Heng Wang, Jianfeng Xu

In this paper, we introduce the \textbf{C}hinese \textbf{AI} and \textbf{L}aw challenge dataset (CAIL2018), the first large-scale Chinese legal dataset for judgment prediction.

Prediction Text Classification

Accurate Text-Enhanced Knowledge Graph Representation Learning

no code implementations NAACL 2018 Bo An, Bo Chen, Xianpei Han, Le Sun

Previous representation learning techniques for knowledge graph representation usually represent the same entity or relation in different triples with the same representation, without considering the ambiguity of relations and entities.

General Classification Graph Representation Learning +4

Nugget Proposal Networks for Chinese Event Detection

1 code implementation ACL 2018 Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun

Neural network based models commonly regard event detection as a word-wise classification task, which suffer from the mismatch problem between words and event triggers, especially in languages without natural word delimiters such as Chinese.

Event Detection General Classification

Adaptive Scaling for Sparse Detection in Information Extraction

1 code implementation ACL 2018 Hongyu Lin, Yaojie Lu, Xianpei Han, Le Sun

This paper focuses on detection tasks in information extraction, where positive instances are sparsely distributed and models are usually evaluated using F-measure on positive classes.

Variational Recurrent Neural Machine Translation

no code implementations16 Jan 2018 Jinsong Su, Shan Wu, Deyi Xiong, Yaojie Lu, Xianpei Han, Biao Zhang

Partially inspired by successful applications of variational recurrent neural networks, we propose a novel variational recurrent neural machine translation (VRNMT) model in this paper.

Decoder Machine Translation +3

Reasoning with Heterogeneous Knowledge for Commonsense Machine Comprehension

no code implementations EMNLP 2017 Hongyu Lin, Le Sun, Xianpei Han

Then we propose a multi-knowledge reasoning model, which selects inference rules for a specific reasoning context using attention mechanism, and reasons by summarizing all valid inference rules.

Natural Language Understanding Reading Comprehension +1

Context-Sensitive Inference Rule Discovery: A Graph-Based Method

no code implementations COLING 2016 Xianpei Han, Le Sun

Firstly, these methods are mostly context-insensitive, cannot accurately measure the similarity between two predicates in a specific context.

Natural Language Inference Question Answering

Cannot find the paper you are looking for? You can Submit a new open access paper.