Search Results for author: Yanghua Xiao

Found 140 papers, 70 papers with code

Parsing Natural Language into Propositional and First-Order Logic with Dual Reinforcement Learning

no code implementations COLING 2022 Xuantao Lu, Jingping Liu, Zhouhong Gu, Hanwen Tong, Chenhao Xie, Junyang Huang, Yanghua Xiao, Wenguang Wang

In this paper, we propose a scoring model to automatically learn a model-based reward, and an effective training strategy based on curriculum learning is further proposed to stabilize the training process.

Natural Language Inference reinforcement-learning +3

PowerAttention: Exponentially Scaling of Receptive Fields for Effective Sparse Attention

no code implementations5 Mar 2025 Lida Chen, Dong Xu, Chenxin An, Xintao Wang, Yikai Zhang, Jiangjie Chen, Zujie Liang, Feng Wei, Jiaqing Liang, Yanghua Xiao, Wei Wang

Large Language Models (LLMs) face efficiency bottlenecks due to the quadratic complexity of the attention mechanism when processing long contexts.

MciteBench: A Benchmark for Multimodal Citation Text Generation in MLLMs

1 code implementation4 Mar 2025 Caiyu Hu, Yikai Zhang, Tinghui Zhu, Yiwei Ye, Yanghua Xiao

To address this gap, we introduce MCiteBench, the first benchmark designed to evaluate and analyze the multimodal citation text generation ability of MLLMs.

Hallucination Text Generation

Reward Shaping to Mitigate Reward Hacking in RLHF

1 code implementation26 Feb 2025 Jiayi Fu, Xuandong Zhao, Chengyuan Yao, Heng Wang, Qi Han, Yanghua Xiao

Reinforcement Learning from Human Feedback (RLHF) is essential for aligning large language models (LLMs) with human values.

Order Matters: Investigate the Position Bias in Multi-constraint Instruction Following

1 code implementation24 Feb 2025 Jie Zeng, Qianyu He, Qingyu Ren, Jiaqing Liang, Yanghua Xiao, Weikang Zhou, Zeye Sun, Fei Yu

Real-world instructions with multiple constraints pose a significant challenge to existing large language models (LLMs).

Instruction Following Position

DEEPER Insight into Your User: Directed Persona Refinement for Dynamic Persona Modeling

1 code implementation16 Feb 2025 Aili Chen, Chengyu Du, Jiangjie Chen, Jinghan Xu, Yikai Zhang, Siyu Yuan, Zulong Chen, Liangyue Li, Yanghua Xiao

To advance personalized applications such as recommendation systems and user behavior prediction, recent research increasingly adopts large language models (LLMs) for human -readable persona modeling.

Prediction Recommendation Systems

CoSER: Coordinating LLM-Based Persona Simulation of Established Roles

1 code implementation13 Feb 2025 Xintao Wang, Heng Wang, Yifei Zhang, Xinfeng Yuan, Rui Xu, Jen-tse Huang, Siyu Yuan, Haoran Guo, Jiangjie Chen, Wei Wang, Yanghua Xiao, Shuchang Zhou

It provides authentic dialogues with real-world intricacies, as well as diverse data types such as conversation setups, character experiences and internal thoughts.

AdaptiveLog: An Adaptive Log Analysis Framework with the Collaboration of Large and Small Language Model

1 code implementation19 Jan 2025 Lipeng Ma, Weidong Yang, Yixuan Li, Ben Fei, Mingjie Zhou, Shuhao Li, Sihang Jiang, Bo Xu, Yanghua Xiao

Specifically, to efficiently query the LLM, we propose an adaptive selection strategy based on the uncertainty estimation of the SLM, where the LLM is invoked only when the SLM is uncertain.

In-Context Learning Language Modeling +1

CDS: Data Synthesis Method Guided by Cognitive Diagnosis Theory

no code implementations13 Jan 2025 Haokun Zhao, Jinyi Han, Jiaqing Liang, Yanghua Xiao

Large Language Models (LLMs) have demonstrated outstanding capabilities across various domains, but the increasing complexity of new challenges demands enhanced performance and adaptability.

cognitive diagnosis Data Augmentation +1

Step-by-Step Mastery: Enhancing Soft Constraint Following Ability of Large Language Models

1 code implementation9 Jan 2025 Qingyu Ren, Jie Zeng, Qianyu He, Jiaqing Liang, Yanghua Xiao, Weikang Zhou, Zeye Sun, Fei Yu

It is crucial for large language models (LLMs) to follow instructions that involve multiple constraints.

MultiLingPoT: Enhancing Mathematical Reasoning with Multilingual Program Fine-tuning

1 code implementation17 Dec 2024 Nianqi Li, Zujie Liang, Siyu Yuan, Jiaqing Liang, Feng Wei, Yanghua Xiao

Since different programming languages excel in different areas, it is natural to use the most suitable language for solving specific problems.

Mathematical Reasoning

Fast and Robust Contextual Node Representation Learning over Dynamic Graphs

no code implementations11 Nov 2024 Xingzhi Guo, Silong Wang, Baojian Zhou, Yanghua Xiao, Steven Skiena

However, most PPR-based GNNs are designed for static graphs, and efficient PPR maintenance remains as an open problem.

Graph Learning Representation Learning

QUILL: Quotation Generation Enhancement of Large Language Models

1 code implementation6 Nov 2024 Jin Xiao, Bowei Zhang, Qianyu He, Jiaqing Liang, Feng Wei, Jinglei Chen, Zujie Liang, Deqing Yang, Yanghua Xiao

To improve the LLMs' quotation generation abilities, we construct a bilingual knowledge base that is broad in scope and rich in dimensions, containing up to 32, 022 quotes.

Faster Local Solvers for Graph Diffusion Equations

1 code implementation29 Oct 2024 Jiahe Bai, Baojian Zhou, Deqing Yang, Yanghua Xiao

Standard iterative methods require accessing the whole graph per iteration, making them time-consuming for large-scale graphs.

Iterative Methods via Locally Evolving Set Process

1 code implementation19 Oct 2024 Baojian Zhou, Yifan Sun, Reza Babanezhad Harikandeh, Xingzhi Guo, Deqing Yang, Yanghua Xiao

We propose to use the \textit{locally evolving set process}, a novel framework to characterize the algorithm locality, and demonstrate that many standard solvers can be effectively localized.

Think Thrice Before You Act: Progressive Thought Refinement in Large Language Models

no code implementations17 Oct 2024 Chengyu Du, Jinyi Han, Yizhou Ying, Aili Chen, Qianyu He, Haokun Zhao, Sirui Xia, Haoran Guo, Jiaqing Liang, Zulong Chen, Liangyue Li, Yanghua Xiao

To address these limitations, we propose Progressive Thought Refinement (PTR), a framework that enables LLMs to refine their responses progressively.

Avg

Revealing the Barriers of Language Agents in Planning

1 code implementation16 Oct 2024 Jian Xie, Kexun Zhang, Jiangjie Chen, Siyu Yuan, Kai Zhang, Yikai Zhang, Lei LI, Yanghua Xiao

Although existing studies have highlighted weak performance in agent planning, the deeper underlying issues and the mechanisms and limitations of the strategies proposed to address them remain insufficiently understood.

Evaluating Semantic Variation in Text-to-Image Synthesis: A Causal Perspective

1 code implementation14 Oct 2024 Xiangru Zhu, Penglei Sun, Yaoxian Song, Yanghua Xiao, Zhixu Li, Chengyu Wang, Jun Huang, Bei Yang, Xiaoxiao Xu

To address these deficiencies, we propose a novel metric called SemVarEffect and a benchmark named SemVarBench, designed to evaluate the causality between semantic variations in inputs and outputs in T2I synthesis.

cross-modal alignment Image Generation

Do Large Language Models have Problem-Solving Capability under Incomplete Information Scenarios?

no code implementations23 Sep 2024 Yuyan Chen, Tianhao Yu, Yueze Li, Songzhou Yan, Sijia Liu, Jiaqing Liang, Yanghua Xiao

Therefore, in this paper, we introduce a novel game named BrainKing based on the ``Who is undercover'' and ``Twenty Questions'' for evaluating LLM capabilities under incomplete information scenarios.

Past Meets Present: Creating Historical Analogy with Large Language Models

1 code implementation23 Sep 2024 Nianqi Li, Siyu Yuan, Jiangjie Chen, Jiaqing Liang, Feng Wei, Zujie Liang, Deqing Yang, Yanghua Xiao

Historical analogies, which compare known past events with contemporary but unfamiliar events, are important abilities that help people make decisions and understand the world.

Retrieval

HOTVCOM: Generating Buzzworthy Comments for Videos

no code implementations23 Sep 2024 Yuyan Chen, Yiwen Qian, Songzhou Yan, Jiyuan Jia, Zhixu Li, Yanghua Xiao, Xiaobo Li, Ming Yang, Qingpei Guo

In the era of social media video platforms, popular ``hot-comments'' play a crucial role in attracting user impressions of short-form videos, making them vital for marketing and branding purpose.

Descriptive Marketing

Recent Advancement of Emotion Cognition in Large Language Models

no code implementations20 Sep 2024 Yuyan Chen, Yanghua Xiao

Emotion cognition in large language models (LLMs) is crucial for enhancing performance across various applications, such as social media, human-computer interaction, and mental health assessment.

Contrastive Learning Emotion Classification +1

EmotionQueen: A Benchmark for Evaluating Empathy of Large Language Models

no code implementations20 Sep 2024 Yuyan Chen, Hao Wang, Songzhou Yan, Sijia Liu, Yueze Li, Yi Zhao, Yanghua Xiao

The framework includes four distinctive tasks: Key Event Recognition, Mixed Event Recognition, Implicit Emotional Recognition, and Intention Recognition.

Emotional Intelligence Emotion Recognition +2

TravelAgent: An AI Assistant for Personalized Travel Planning

no code implementations12 Sep 2024 Aili Chen, Xuyang Ge, Ziquan Fu, Yanghua Xiao, Jiangjie Chen

As global tourism expands and artificial intelligence technology advances, intelligent travel planning services have emerged as a significant research focus.

LUK: Empowering Log Understanding with Expert Knowledge from Large Language Models

1 code implementation3 Sep 2024 Lipeng Ma, Weidong Yang, Sihang Jiang, Ben Fei, Mingjie Zhou, Shuhao Li, Mingyu Zhao, Bo Xu, Yanghua Xiao

To address the lack of expert knowledge and enhance log understanding for smaller PLMs, this paper introduces a novel and practical knowledge enhancement framework, called LUK, which acquires expert knowledge from LLMs automatically and then enhances the smaller PLM for log analysis with these expert knowledge.

Dr.Academy: A Benchmark for Evaluating Questioning Capability in Education for Large Language Models

no code implementations20 Aug 2024 Yuyan Chen, Chenwei Wu, Songzhou Yan, Panjun Liu, Haoyu Zhou, Yanghua Xiao

Therefore, our research introduces a benchmark to evaluate the questioning capability in education as a teacher of LLMs through evaluating their generated educational questions, utilizing Anderson and Krathwohl's taxonomy across general, monodisciplinary, and interdisciplinary domains.

XMeCap: Meme Caption Generation with Sub-Image Adaptability

no code implementations24 Jul 2024 Yuyan Chen, Songzhou Yan, Zhihong Zhu, Zhixu Li, Yanghua Xiao

Humor, deeply rooted in societal meanings and cultural details, poses a unique challenge for machines.

Caption Generation Meme Captioning

Can Pre-trained Language Models Understand Chinese Humor?

no code implementations4 Jul 2024 Yuyan Chen, Zhixu Li, Jiaqing Liang, Yanghua Xiao, Bang Liu, Yunwen Chen

Humor understanding is an important and challenging research in natural language processing.

MAPO: Boosting Large Language Model Performance with Model-Adaptive Prompt Optimization

no code implementations4 Jul 2024 Yuyan Chen, Zhihao Wen, Ge Fan, Zhengyu Chen, Wei Wu, Dayiheng Liu, Zhixu Li, Bang Liu, Yanghua Xiao

Prompt engineering, as an efficient and effective way to leverage Large Language Models (LLM), has drawn a lot of attention from the research community.

Language Modeling Language Modelling +3

Hallucination Detection: Robustly Discerning Reliable Answers in Large Language Models

no code implementations4 Jul 2024 Yuyan Chen, Qiang Fu, Yichen Yuan, Zhihao Wen, Ge Fan, Dayiheng Liu, Dongmei Zhang, Zhixu Li, Yanghua Xiao

Large Language Models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems.

Hallucination Question Answering

Capturing Minds, Not Just Words: Enhancing Role-Playing Language Models with Personality-Indicative Data

1 code implementation27 Jun 2024 Yiting Ran, Xintao Wang, Rui Xu, Xinfeng Yuan, Jiaqing Liang, Deqing Yang, Yanghua Xiao

Role-playing agents (RPA) have been a popular application area for large language models (LLMs), attracting significant interest from both industry and academia. While existing RPAs well portray the characters' knowledge and tones, they face challenges in capturing their minds, especially for small role-playing language models (RPLMs).

ESC-Eval: Evaluating Emotion Support Conversations in Large Language Models

3 code implementations21 Jun 2024 Haiquan Zhao, Lingyu Li, Shisong Chen, Shuqi Kong, Jiaan Wang, Kexin Huang, Tianle Gu, Yixu Wang, Wang Jian, Dandan Liang, Zhixu Li, Yan Teng, Yanghua Xiao, Yingchun Wang

Inspired by the awesome development of role-playing agents, we propose an ESC Evaluation framework (ESC-Eval), which uses a role-playing agent to interact with ESC models, followed by a manual evaluation of the interactive dialogues.

DetectBench: Can Large Language Model Detect and Piece Together Implicit Evidence?

1 code implementation18 Jun 2024 Zhouhong Gu, Lin Zhang, Xiaoxuan Zhu, Jiangjie Chen, Wenhao Huang, Yikai Zhang, Shusen Wang, Zheyu Ye, Yan Gao, Hongwei Feng, Yanghua Xiao

This paper proposes a benchmark called DetectBench for verifying the ability to detect and piece together implicit evidence within a long context.

Language Modeling Language Modelling +2

Teaching Large Language Models to Express Knowledge Boundary from Their Own Signals

no code implementations16 Jun 2024 Lida Chen, Zujie Liang, Xintao Wang, Jiaqing Liang, Yanghua Xiao, Feng Wei, Jinglei Chen, Zhenghong Hao, Bing Han, Wei Wang

Large language models (LLMs) have achieved great success, but their occasional content fabrication, or hallucination, limits their practical application.

Hallucination

VCEval: Rethinking What is a Good Educational Video and How to Automatically Evaluate It

no code implementations15 Jun 2024 Xiaoxuan Zhu, Zhouhong Gu, Sihang Jiang, Zhixu Li, Hongwei Feng, Yanghua Xiao

Online courses have significantly lowered the barrier to accessing education, yet the varying content quality of these videos poses challenges.

Language Modeling Language Modelling +2

StrucText-Eval: Evaluating Large Language Model's Reasoning Ability in Structure-Rich Text

1 code implementation15 Jun 2024 Zhouhong Gu, Haoning Ye, Xingzhou Chen, Zeyang Zhou, Hongwei Feng, Yanghua Xiao

The effective utilization of structured data, integral to corporate data strategies, has been challenged by the rise of large language models (LLMs) capable of processing unstructured information.

SelfGoal: Your Language Agents Already Know How to Achieve High-level Goals

no code implementations7 Jun 2024 Ruihan Yang, Jiangjie Chen, Yikai Zhang, Siyu Yuan, Aili Chen, Kyle Richardson, Yanghua Xiao, Deqing Yang

Language agents powered by large language models (LLMs) are increasingly valuable as decision-making tools in domains such as gaming and programming.

Decision Making

SED: Self-Evaluation Decoding Enhances Large Language Models for Better Generation

no code implementations26 May 2024 Ziqin Luo, Haixia Han, Haokun Zhao, Guochao Jiang, Chengyu Du, Tingyun Li, Jiaqing Liang, Deqing Yang, Yanghua Xiao

Existing Large Language Models (LLMs) generate text through unidirectional autoregressive decoding methods to respond to various user queries.

Decision Making

From Persona to Personalization: A Survey on Role-Playing Language Agents

no code implementations28 Apr 2024 Jiangjie Chen, Xintao Wang, Rui Xu, Siyu Yuan, Yikai Zhang, Wei Shi, Jian Xie, Shuang Li, Ruihan Yang, Tinghui Zhu, Aili Chen, Nianqi Li, Lida Chen, Caiyu Hu, Siye Wu, Scott Ren, Ziquan Fu, Yanghua Xiao

Through this work, we aim to establish a clear taxonomy of RPLA research and applications, and facilitate future research in this critical and ever-evolving field, and pave the way for a future where humans and RPLAs coexist in harmony.

In-Context Learning Instruction Following

From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large Language Models

1 code implementation24 Apr 2024 Qianyu He, Jie Zeng, Qianxi He, Jiaqing Liang, Yanghua Xiao

It is imperative for Large language models (LLMs) to follow instructions with elaborate requirements (i. e. Complex Instructions Following).

Instruction Following

AutoScraper: A Progressive Understanding Web Agent for Web Scraper Generation

2 code implementations19 Apr 2024 Wenhao Huang, Zhouhong Gu, Chenghao Peng, Zhixu Li, Jiaqing Liang, Yanghua Xiao, Liqian Wen, Zulong Chen

In this work, we introduce the paradigm of generating web scrapers with LLMs and propose AutoScraper, a two-stage framework that can handle diverse and changing web environments more efficiently.

Action Generation

Enhancing Confidence Expression in Large Language Models Through Learning from Past Experience

no code implementations16 Apr 2024 Haixia Han, Tingyun Li, Shisong Chen, Jie Shi, Chengyu Du, Yanghua Xiao, Jiaqing Liang, Xin Lin

Specifically, we first identify three key problems: (1) How to capture the inherent confidence of the LLM?

Negation Triplet Extraction with Syntactic Dependency and Semantic Consistency

1 code implementation15 Apr 2024 Yuchen Shi, Deqing Yang, Jingping Liu, Yanghua Xiao, ZongYu Wang, Huimin Xu

To achieve NTE, we devise a novel Syntax&Semantic-Enhanced Negation Extraction model, namely SSENE, which is built based on a generative pretrained language model (PLM) {of Encoder-Decoder architecture} with a multi-task learning framework.

Decoder Language Modelling +4

CEM: A Data-Efficient Method for Large Language Models to Continue Evolving From Mistakes

no code implementations11 Apr 2024 Haokun Zhao, Haixia Han, Jie Shi, Chengyu Du, Jiaqing Liang, Yanghua Xiao

As world knowledge advances and new task schemas emerge, Continual Learning (CL) becomes essential for keeping Large Language Models (LLMs) current and addressing their shortcomings.

Continual Learning Continual Pretraining +4

SurveyAgent: A Conversational System for Personalized and Efficient Research Survey

no code implementations9 Apr 2024 Xintao Wang, Jiangjie Chen, Nianqi Li, Lida Chen, Xinfeng Yuan, Wei Shi, Xuyang Ge, Rui Xu, Yanghua Xiao

In the rapidly advancing research fields such as AI, managing and staying abreast of the latest scientific literature has become a significant challenge for researchers.

Management Question Answering +1

How Easily do Irrelevant Inputs Skew the Responses of Large Language Models?

1 code implementation4 Apr 2024 Siye Wu, Jian Xie, Jiangjie Chen, Tinghui Zhu, Kai Zhang, Yanghua Xiao

By leveraging the retrieval of information from external knowledge databases, Large Language Models (LLMs) exhibit enhanced capabilities for accomplishing many knowledge-intensive tasks.

Retrieval

Reason from Fallacy: Enhancing Large Language Models' Logical Reasoning through Logical Fallacy Understanding

no code implementations4 Apr 2024 Yanda Li, Dixuan Wang, Jiaqing Liang, Guochao Jiang, Qianyu He, Yanghua Xiao, Deqing Yang

Large Language Models (LLMs) have demonstrated good performance in many reasoning tasks, but they still struggle with some complicated reasoning tasks including logical reasoning.

Logical Fallacies Logical Reasoning

AgentGroupChat: An Interactive Group Chat Simulacra For Better Eliciting Emergent Behavior

1 code implementation20 Mar 2024 Zhouhong Gu, Xiaoxuan Zhu, Haoran Guo, Lin Zhang, Yin Cai, Hao Shen, Jiangjie Chen, Zheyu Ye, Yifei Dai, Yan Gao, Yao Hu, Hongwei Feng, Yanghua Xiao

Language significantly influences the formation and evolution of Human emergent behavior, which is crucial in understanding collective intelligence within human societies.

OVEL: Large Language Model as Memory Manager for Online Video Entity Linking

no code implementations3 Mar 2024 Haiquan Zhao, Xuwu Wang, Shisong Chen, Zhixu Li, Xin Zheng, Yanghua Xiao

In this paper, we propose a task called Online Video Entity Linking OVEL, aiming to establish connections between mentions in online videos and a knowledge base with high accuracy and timeliness.

Entity Linking Language Modeling +3

GumbelSoft: Diversified Language Model Watermarking via the GumbelMax-trick

1 code implementation20 Feb 2024 Jiayi Fu, Xuandong Zhao, Ruihan Yang, Yuansen Zhang, Jiangjie Chen, Yanghua Xiao

Large language models (LLMs) excellently generate human-like text, but also raise concerns about misuse in fake news and academic dishonesty.

Diversity Language Modeling +1

TimeArena: Shaping Efficient Multitasking Language Agents in a Time-Aware Simulation

no code implementations8 Feb 2024 Yikai Zhang, Siyu Yuan, Caiyu Hu, Kyle Richardson, Yanghua Xiao, Jiangjie Chen

Despite remarkable advancements in emulating human-like behavior through Large Language Models (LLMs), current textual simulations do not adequately address the notion of time.

TravelPlanner: A Benchmark for Real-World Planning with Language Agents

2 code implementations2 Feb 2024 Jian Xie, Kai Zhang, Jiangjie Chen, Tinghui Zhu, Renze Lou, Yuandong Tian, Yanghua Xiao, Yu Su

Are these language agents capable of planning in more complex settings that are out of the reach of prior AI agents?

Exploiting Duality in Open Information Extraction with Predicate Prompt

1 code implementation20 Jan 2024 Zhen Chen, Jingping Liu, Deqing Yang, Yanghua Xiao, Huimin Xu, ZongYu Wang, Rui Xie, Yunsen Xian

Open information extraction (OpenIE) aims to extract the schema-free triplets in the form of (\emph{subject}, \emph{predicate}, \emph{object}) from a given sentence.

Open Information Extraction Sentence

Small Language Model Can Self-correct

no code implementations14 Jan 2024 Haixia Han, Jiaqing Liang, Jie Shi, Qianyu He, Yanghua Xiao

In this paper, we introduce the \underline{I}ntrinsic \underline{S}elf-\underline{C}orrection (ISC) in generative language models, aiming to correct the initial output of LMs in a self-triggered manner, even for those small LMs with 6 billion parameters.

Language Modeling Language Modelling +1

ConcEPT: Concept-Enhanced Pre-Training for Language Models

no code implementations11 Jan 2024 Xintao Wang, Zhouhong Gu, Jiaqing Liang, Dakuan Lu, Yanghua Xiao, Wei Wang

In this paper, we propose ConcEPT, which stands for Concept-Enhanced Pre-Training for language models, to infuse conceptual knowledge into PLMs.

Entity Linking Entity Typing

Enhancing Quantitative Reasoning Skills of Large Language Models through Dimension Perception

no code implementations29 Dec 2023 Yuncheng Huang, Qianyu He, Jiaqing Liang, Sihang Jiang, Yanghua Xiao, Yunwen Chen

Hence, we present a framework to enhance the quantitative reasoning ability of language models based on dimension perception.

M^2ConceptBase: A Fine-Grained Aligned Concept-Centric Multimodal Knowledge Base

1 code implementation16 Dec 2023 Zhiwei Zha, Jiaan Wang, Zhixu Li, Xiangru Zhu, Wei Song, Yanghua Xiao

Comprising 951K images and 152K concepts, M^2ConceptBase links each concept to an average of 6. 27 images and a single description, ensuring comprehensive visual and textual semantics.

cross-modal alignment Knowledge Graphs +2

InCharacter: Evaluating Personality Fidelity in Role-Playing Agents through Psychological Interviews

2 code implementations27 Oct 2023 Xintao Wang, Yunze Xiao, Jen-tse Huang, Siyu Yuan, Rui Xu, Haoran Guo, Quan Tu, Yaying Fei, Ziang Leng, Wei Wang, Jiangjie Chen, Cheng Li, Yanghua Xiao

Then, with InCharacter, we show that state-of-the-art RPAs exhibit personalities highly aligned with the human-perceived personalities of the characters, achieving an accuracy up to 80. 7%.

Towards Visual Taxonomy Expansion

1 code implementation12 Sep 2023 Tinghui Zhu, Jingping Liu, Jiaqing Liang, Haiyun Jiang, Yanghua Xiao, ZongYu Wang, Rui Xie, Yunsen Xian

Specifically, on the Chinese taxonomy dataset, our method significantly improves accuracy by 8. 75 %.

Taxonomy Expansion

Translate Meanings, Not Just Words: IdiomKB's Role in Optimizing Idiomatic Translation with Language Models

1 code implementation26 Aug 2023 Shuang Li, Jiangjie Chen, Siyu Yuan, Xinyi Wu, Hao Yang, Shimin Tao, Yanghua Xiao

To translate well, machine translation (MT) systems and general-purposed language models (LMs) need a deep understanding of both source and target languages and cultures.

Machine Translation Translation

KnowledGPT: Enhancing Large Language Models with Retrieval and Storage Access on Knowledge Bases

no code implementations17 Aug 2023 Xintao Wang, Qianwen Yang, Yongting Qiu, Jiaqing Liang, Qianyu He, Zhouhong Gu, Yanghua Xiao, Wei Wang

Large language models (LLMs) have demonstrated impressive impact in the field of natural language processing, but they still struggle with several issues regarding, such as completeness, timeliness, faithfulness and adaptability.

Retrieval World Knowledge

AspectMMKG: A Multi-modal Knowledge Graph with Aspect-aware Entities

1 code implementation9 Aug 2023 Jingdan Zhang, Jiaan Wang, Xiaodan Wang, Zhixu Li, Yanghua Xiao

Multi-modal knowledge graphs (MMKGs) combine different modal data (e. g., text and image) for a comprehensive understanding of entities.

Extract Aspect Image Retrieval +2

Piecing Together Clues: A Benchmark for Evaluating the Detective Skills of Large Language Models

no code implementations11 Jul 2023 Zhouhong Gu, Lin Zhang, Jiangjie Chen, Haoning Ye, Xiaoxuan Zhu, Zihan Li, Zheyu Ye, Yan Gao, Yao Hu, Yanghua Xiao, Hongwei Feng

We introduces the DetectBench, a reading comprehension dataset designed to assess a model's ability to jointly ability in key information detection and multi-hop reasoning when facing complex and implicit information.

Common Sense Reasoning Decision Making +2

M3PT: A Multi-Modal Model for POI Tagging

no code implementations16 Jun 2023 Jingsong Yang, Guanzhou Han, Deqing Yang, Jingping Liu, Yanghua Xiao, Xiang Xu, Baohua Wu, Shenghua Ni

In this paper, we propose a novel Multi-Modal Model for POI Tagging, namely M3PT, which achieves enhanced POI tagging through fusing the target POI's textual and visual features, and the precise matching between the multi-modal representations.

Contrastive Learning model

Beneath Surface Similarity: Large Language Models Make Reasonable Scientific Analogies after Structure Abduction

1 code implementation22 May 2023 Siyu Yuan, Jiangjie Chen, Xuyang Ge, Yanghua Xiao, Deqing Yang

The vital role of analogical reasoning in human cognition allows us to grasp novel concepts by linking them with familiar ones through shared relational structures.

Novel Concepts Question Answering

Distilling Script Knowledge from Large Language Models for Constrained Language Planning

1 code implementation9 May 2023 Siyu Yuan, Jiangjie Chen, Ziquan Fu, Xuyang Ge, Soham Shah, Charles Robert Jankowski, Yanghua Xiao, Deqing Yang

In everyday life, humans often plan their actions by following step-by-step instructions in the form of goal-oriented scripts.

Knowledge Distillation

Causality-aware Concept Extraction based on Knowledge-guided Prompting

1 code implementation3 May 2023 Siyu Yuan, Deqing Yang, Jinxi Liu, Shuyu Tian, Jiaqing Liang, Yanghua Xiao, Rui Xie

The prompt adopts the topic of the given entity from the existing knowledge in KGs to mitigate the spurious co-occurrence correlations between entities and biased concepts.

Knowledge Graphs Natural Language Understanding

GANTEE: Generative Adversatial Network for Taxonomy Entering Evaluation

no code implementations25 Mar 2023 Zhouhong Gu, Sihang Jiang, Jingping Liu, Yanghua Xiao, Hongwei Feng, Zhixu Li, Jiaqing Liang, Jian Zhong

The previous methods suffer from low-efficiency since they waste much time when most of the new coming concepts are indeed noisy concepts.

Generative Adversarial Network Taxonomy Expansion

MAPS-KB: A Million-scale Probabilistic Simile Knowledge Base

2 code implementations10 Dec 2022 Qianyu He, Xintao Wang, Jiaqing Liang, Yanghua Xiao

The ability to understand and generate similes is an imperative step to realize human-level AI.

Knowledge Base Construction

Harnessing Knowledge and Reasoning for Human-Like Natural Language Generation: A Brief Review

no code implementations7 Dec 2022 Jiangjie Chen, Yanghua Xiao

The rapid development and application of natural language generation (NLG) techniques has revolutionized the field of automatic text production.

Text Generation

Group Buying Recommendation Model Based on Multi-task Learning

1 code implementation25 Nov 2022 Shuoyao Zhai, Baichuan Liu, Deqing Yang, Yanghua Xiao

Furthermore, we propose two auxiliary losses corresponding to the two sub-tasks, to refine the representation learning in our model.

Multi-Task Learning Representation Learning

Improving Continual Relation Extraction through Prototypical Contrastive Learning

no code implementations COLING 2022 Chengwei Hu, Deqing Yang, Haoliang Jin, Zhen Chen, Yanghua Xiao

Continual relation extraction (CRE) aims to extract relations towards the continuous and iterative arrival of new data, of which the major challenge is the catastrophic forgetting of old tasks.

Continual Relation Extraction Contrastive Learning +1

Generative Entity Typing with Curriculum Learning

1 code implementation6 Oct 2022 Siyu Yuan, Deqing Yang, Jiaqing Liang, Zhixu Li, Jinxi Liu, Jingyue Huang, Yanghua Xiao

To overcome these drawbacks, we propose a novel generative entity typing (GET) paradigm: given a text with an entity mention, the multiple types for the role that the entity plays in the text are generated with a pre-trained language model (PLM).

Entity Typing Language Modelling

Large-scale Multi-granular Concept Extraction Based on Machine Reading Comprehension

1 code implementation30 Aug 2022 Siyu Yuan, Deqing Yang, Jiaqing Liang, Jilun Sun, Jingyue Huang, Kaiyan Cao, Yanghua Xiao, Rui Xie

In order to supply existing KGs with more fine-grained and new concepts, we propose a novel concept extraction framework, namely MRC-CE, to extract large-scale multi-granular concepts from the descriptive texts of entities.

Descriptive Knowledge Graphs +1

Factorial User Modeling with Hierarchical Graph Neural Network for Enhanced Sequential Recommendation

1 code implementation27 Jul 2022 Lyuxin Xue, Deqing Yang, Yanghua Xiao

Most sequential recommendation (SR) systems employing graph neural networks (GNNs) only model a user's interaction sequence as a flat graph without hierarchy, overlooking diverse factors in the user's preference.

Graph Neural Network Sequential Recommendation

Contextual Information and Commonsense Based Prompt for Emotion Recognition in Conversation

1 code implementation27 Jul 2022 Jingjie Yi, Deqing Yang, Siyu Yuan, Caiyan Cao, Zhiyao Zhang, Yanghua Xiao

The newly proposed ERC models have leveraged pre-trained language models (PLMs) with the paradigm of pre-training and fine-tuning to obtain good performance.

Emotion Recognition in Conversation Language Modeling +1

Language Models as Knowledge Embeddings

1 code implementation25 Jun 2022 Xintao Wang, Qianyu He, Jiaqing Liang, Yanghua Xiao

In this paper, we propose LMKE, which adopts Language Models to derive Knowledge Embeddings, aiming at both enriching representations of long-tail entities and solving problems of prior description-based methods.

Contrastive Learning Link Prediction +1

Tackling Math Word Problems with Fine-to-Coarse Abstracting and Reasoning

no code implementations17 May 2022 Ailisi Li, Xueyao Jiang, Bang Liu, Jiaqing Liang, Yanghua Xiao

Math Word Problems (MWP) is an important task that requires the ability of understanding and reasoning over mathematical text.

Math

Neighbors Are Not Strangers: Improving Non-Autoregressive Translation under Low-Frequency Lexical Constraints

1 code implementation NAACL 2022 Chun Zeng, Jiangjie Chen, Tianyi Zhuang, Rui Xu, Hao Yang, Ying Qin, Shimin Tao, Yanghua Xiao

To this end, we propose a plug-in algorithm for this line of work, i. e., Aligned Constrained Training (ACT), which alleviates this problem by familiarizing the model with the source-side context of the constraints.

Translation

WikiDiverse: A Multimodal Entity Linking Dataset with Diversified Contextual Topics and Entity Types

3 code implementations ACL 2022 Xuwu Wang, Junfeng Tian, Min Gui, Zhixu Li, Rui Wang, Ming Yan, Lihan Chen, Yanghua Xiao

In this paper, we present WikiDiverse, a high-quality human-annotated MEL dataset with diversified contextual topics and entity types from Wikinews, which uses Wikipedia as the corresponding knowledge base.

Entity Linking

Learning What You Need from What You Did: Product Taxonomy Expansion with User Behaviors Supervision

1 code implementation28 Mar 2022 Sijie Cheng, Zhouhong Gu, Bang Liu, Rui Xie, Wei Wu, Yanghua Xiao

Specifically, i) to fully exploit user behavioral information, we extract candidate hyponymy relations that match user interests from query-click concepts; ii) to enhance the semantic information of new concepts and better detect hyponymy relations, we model concepts and relations through both user-generated content and structural information in existing taxonomies and user click logs, by leveraging Pre-trained Language Models and Graph Neural Network combined with Contrastive Learning; iii) to reduce the cost of dataset construction and overcome data skews, we construct a high-quality and balanced training dataset from existing taxonomy with no supervision.

Contrastive Learning Graph Neural Network +1

E-KAR: A Benchmark for Rationalizing Natural Language Analogical Reasoning

no code implementations Findings (ACL) 2022 Jiangjie Chen, Rui Xu, Ziquan Fu, Wei Shi, Zhongqiao Li, Xinbo Zhang, Changzhi Sun, Lei LI, Yanghua Xiao, Hao Zhou

Holding the belief that models capable of reasoning should be right for the right reasons, we propose a first-of-its-kind Explainable Knowledge-intensive Analogical Reasoning benchmark (E-KAR).

Explanation Generation Question Answering

Can Pre-trained Language Models Interpret Similes as Smart as Human?

1 code implementation ACL 2022 Qianyu He, Sijie Cheng, Zhixu Li, Rui Xie, Yanghua Xiao

In this paper, we investigate the ability of PLMs in simile interpretation by designing a novel task named Simile Property Probing, i. e., to let the PLMs infer the shared properties of similes.

Sentiment Analysis Sentiment Classification

Rule Mining over Knowledge Graphs via Reinforcement Learning

no code implementations21 Feb 2022 Lihan Chen, Sihang Jiang, Jingping Liu, Chao Wang, Sheng Zhang, Chenhao Xie, Jiaqing Liang, Yanghua Xiao, Rui Song

Knowledge graphs (KGs) are an important source repository for a wide range of applications and rule mining from KGs recently attracts wide research interest in the KG-related research community.

Knowledge Graphs reinforcement-learning +2

Multi-Modal Knowledge Graph Construction and Application: A Survey

no code implementations11 Feb 2022 Xiangru Zhu, Zhixu Li, Xiaodan Wang, Xueyao Jiang, Penglei Sun, Xuwu Wang, Yanghua Xiao, Nicholas Jing Yuan

In this survey on MMKGs constructed by texts and images, we first give definitions of MMKGs, followed with the preliminaries on multi-modal tasks and techniques.

graph construction Knowledge Graphs +2

Grow-and-Clip: Informative-yet-Concise Evidence Distillation for Answer Explanation

no code implementations13 Jan 2022 Yuyan Chen, Yanghua Xiao, Bang Liu

In this research, we argue that the evidences of an answer is critical to enhancing the interpretability of QA models.

Informativeness Question Answering +2

Semantic-based Data Augmentation for Math Word Problems

no code implementations7 Jan 2022 Ailisi Li, Jiaqing Liang, Yanghua Xiao

In this paper, we propose a set of novel data augmentation approaches to supplement existing datasets with such data that are augmented with different kinds of local variances, and help to improve the generalization ability of current neural models.

Data Augmentation Math

Unsupervised Editing for Counterfactual Stories

1 code implementation10 Dec 2021 Jiangjie Chen, Chun Gan, Sijie Cheng, Hao Zhou, Yanghua Xiao, Lei LI

We also propose a new metric to alleviate the shortcomings of current automatic metrics and better evaluate the trade-off.

counterfactual

A Review on Graph Neural Network Methods in Financial Applications

no code implementations27 Nov 2021 Jianian Wang, Sheng Zhang, Yanghua Xiao, Rui Song

With multiple components and relations, financial data are often presented as graph data, since it could represent both the individual features and the complicated relations.

Graph Neural Network

A Probit Tensor Factorization Model For Relational Learning

no code implementations6 Nov 2021 Ye Liu, Rui Song, Wenbin Lu, Yanghua Xiao

A large number of models and algorithms have been proposed to perform link prediction, among which tensor factorization method has proven to achieve state-of-the-art performance in terms of computation efficiency and prediction accuracy.

Knowledge Graphs Link Prediction +3

FedGEMS: Federated Learning of Larger Server Models via Selective Knowledge Fusion

no code implementations21 Oct 2021 Sijie Cheng, Jingwen Wu, Yanghua Xiao, Yang Liu

Today data is often scattered among billions of resource-constrained edge devices with security and privacy constraints.

Federated Learning Image Classification

Relation Aware Semi-autoregressive Semantic Parsing for NL2SQL

no code implementations2 Aug 2021 Junyang Huang, Yongbo Wang, Yongliang Wang, Yang Dong, Yanghua Xiao

It first learns relation embedding over the schema entities and question words with predefined schema relations with ELECTRA and relation aware transformer layer as backbone.

Relation Semantic Parsing

Refining Sample Embeddings with Relation Prototypes to Enhance Continual Relation Extraction

1 code implementation ACL 2021 Li Cui, Deqing Yang, Jiaxin Yu, Chengwei Hu, Jiayang Cheng, Jingjie Yi, Yanghua Xiao

As a typical task of continual learning, continual relation extraction (CRE) aims to extract relations between entities from texts, where the samples of different relations are delivered into the model continuously.

Continual Learning Continual Relation Extraction +1

A Question-answering Based Framework for Relation Extraction Validation

no code implementations7 Apr 2021 Jiayang Cheng, Haiyun Jiang, Deqing Yang, Yanghua Xiao

However, few works have focused on how to validate and correct the results generated by the existing relation extraction models.

Question Answering Relation +1

Complex Relation Extraction: Challenges and Opportunities

no code implementations9 Dec 2020 Haiyun Jiang, Qiaoben Bao, Qiao Cheng, Deqing Yang, Li Wang, Yanghua Xiao

In recent years, many complex relation extraction tasks, i. e., the variants of simple binary relation extraction, are proposed to meet the complex applications in practice.

Binary Relation Extraction Knowledge Base Construction +1

Convolutional Gaussian Embeddings for Personalized Recommendation with Uncertainty

1 code implementation19 Jun 2020 Junyang Jiang, Deqing Yang, Yanghua Xiao, Chenlu Shen

Most of existing embedding based recommendation models use embeddings (vectors) corresponding to a single fixed point in low-dimensional space, to represent users and items.

Recommendation Systems

A Knowledge-Enhanced Recommendation Model with Attribute-Level Co-Attention

no code implementations18 Jun 2020 Deqing Yang, Zengcun Song, Lvxin Xue, Yanghua Xiao

Deep neural networks (DNNs) have been widely employed in recommender systems including incorporating attention mechanism for performance improvement.

Attribute Knowledge Graphs +1

Incorporating User Micro-behaviors and Item Knowledge into Multi-task Learning for Session-based Recommendation

1 code implementation12 Jun 2020 Wenjing Meng, Deqing Yang, Yanghua Xiao

These insights motivate us to propose a novel SR model MKM-SR in this paper, which incorporates user Micro-behaviors and item Knowledge into Multi-task learning for Session-based Recommendation.

Multi-Task Learning Session-Based Recommendations

Collective Loss Function for Positive and Unlabeled Learning

no code implementations6 May 2020 Chenhao Xie, Qiao Cheng, Jiaqing Liang, Lihan Chen, Yanghua Xiao

On the contrary, traditional machine learning algorithms often rely on negative examples, otherwise the model would be prone to collapse and always-true predictions.

Bayes EMbedding (BEM): Refining Representation by Integrating Knowledge Graphs and Behavior-specific Networks

1 code implementation28 Aug 2019 Yuting Ye, Xuwu Wang, Jiangchao Yao, Kunyang Jia, Jingren Zhou, Yanghua Xiao, Hongxia Yang

Low-dimensional embeddings of knowledge graphs and behavior graphs have proved remarkably powerful in varieties of tasks, from predicting unobserved edges between entities to content recommendation.

General Classification Knowledge Graph Embedding +4

Ensuring Readability and Data-fidelity using Head-modifier Templates in Deep Type Description Generation

no code implementations ACL 2019 Jiangjie Chen, Ao Wang, Haiyun Jiang, Suo Feng, Chenguang Li, Yanghua Xiao

A type description is a succinct noun compound which helps human and machines to quickly grasp the informative and distinctive information of an entity.

Knowledge Graphs

KBQA: Learning Question Answering over QA Corpora and Knowledge Bases

no code implementations6 Mar 2019 Wanyun Cui, Yanghua Xiao, Haixun Wang, Yangqiu Song, Seung-won Hwang, Wei Wang

Based on these templates, our QA system KBQA effectively supports binary factoid questions, as well as complex questions which are composed of a series of binary factoid questions.

Question Answering

Deep Short Text Classification with Knowledge Powered Attention

1 code implementation21 Feb 2019 Jindong Chen, Yizhou Hu, Jingping Liu, Yanghua Xiao, Haiyun Jiang

For the purpose of measuring the importance of knowledge, we introduce attention mechanisms and propose deep Short Text Classification with Knowledge powered Attention (STCKA).

General Classification text-classification +1

Verb Pattern: A Probabilistic Semantic Representation on Verbs

no code implementations20 Oct 2017 Wanyun Cui, Xiyou Zhou, Hangyu Lin, Yanghua Xiao, Haixun Wang, Seung-won Hwang, Wei Wang

In this paper, we introduce verb patterns to represent verbs' semantics, such that each pattern corresponds to a single semantic of the verb.

Specificity

Entity Suggestion by Example using a Conceptual Taxonomy

no code implementations29 Nov 2015 Yi Zhang, Yanghua Xiao, Seung-won Hwang, Haixun Wang, X. Sean Wang, Wei Wang

This paper provides a query processing method based on the relevance models between entity sets and concepts.

Cannot find the paper you are looking for? You can Submit a new open access paper.