Search Results for author: Yongliang Shen

Found 35 papers, 23 papers with code

De-Bias for Generative Extraction in Unified NER Task

no code implementations ACL 2022 Shuai Zhang, Yongliang Shen, Zeqi Tan, Yiquan Wu, Weiming Lu

Named entity recognition (NER) is a fundamental task to recognize specific types of entities from a given sentence.

Attribute Data Augmentation +4

2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining

1 code implementation1 Jan 2025 Wenqi Zhang, Hang Zhang, Xin Li, Jiashuo Sun, Yongliang Shen, Weiming Lu, Deli Zhao, Yueting Zhuang, Lidong Bing

Compared to its counterparts, our video-centric textbook offers more coherent context, richer knowledge, and better image-text alignment.

Optical Character Recognition (OCR)

GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable Recommendation

1 code implementation15 Oct 2024 Fei Tang, Yongliang Shen, Hang Zhang, Zeqi Tan, Wenqi Zhang, Guiyang Hou, Kaitao Song, Weiming Lu, Yueting Zhuang

GaVaMoE introduces two key components: (1) a rating reconstruction module that employs Variational Autoencoder (VAE) with a Gaussian Mixture Model (GMM) to capture complex user-item collaborative preferences, serving as a pre-trained multi-gating mechanism; and (2) a set of fine-grained expert models coupled with the multi-gating mechanism for generating highly personalized explanations.

Explainable Recommendation Language Modelling +1

Entering Real Social World! Benchmarking the Social Intelligence of Large Language Models from a First-person Perspective

1 code implementation8 Oct 2024 Guiyang Hou, Wenqi Zhang, Yongliang Shen, Zeqi Tan, Sihao Shen, Weiming Lu

(3) a lack of comprehensive evaluation of behavioral intelligence, with specific emphasis on incorporating critical human-machine interaction scenarios.

Attribute Benchmarking +2

Multimodal Self-Instruct: Synthetic Abstract Image and Visual Reasoning Instruction Using Language Model

1 code implementation9 Jul 2024 Wenqi Zhang, Zhenglin Cheng, Yuanyu He, Mengna Wang, Yongliang Shen, Zeqi Tan, Guiyang Hou, Mingqian He, Yanna Ma, Weiming Lu, Yueting Zhuang

In light of this, we design a multi-modal self-instruct, utilizing large language models and their code capabilities to synthesize massive abstract images and visual reasoning instructions across daily scenarios.

Chart Understanding Language Modeling +2

TimeToM: Temporal Space is the Key to Unlocking the Door of Large Language Models' Theory-of-Mind

no code implementations1 Jul 2024 Guiyang Hou, Wenqi Zhang, Yongliang Shen, Linjuan Wu, Weiming Lu

Theory of Mind (ToM)-the cognitive ability to reason about mental states of ourselves and others, is the foundation of social interaction.

Advancing Process Verification for Large Language Models via Tree-Based Preference Learning

no code implementations29 Jun 2024 Mingqian He, Yongliang Shen, Wenqi Zhang, Zeqi Tan, Weiming Lu

For instance, Tree-PLV achieved substantial performance gains over the Mistral-7B self-consistency baseline on GSM8K (67. 55% to 82. 79%), MATH (17. 00% to 26. 80%), CSQA (68. 14% to 72. 97%), and StrategyQA (82. 86% to 83. 25%). Additionally, our study explores the appropriate granularity for applying preference learning, revealing that step-level guidance provides feedback that better aligns with the evaluation of the reasoning process.

Binary Classification GSM8K +2

EASYTOOL: Enhancing LLM-based Agents with Concise Tool Instruction

1 code implementation11 Jan 2024 Siyu Yuan, Kaitao Song, Jiangjie Chen, Xu Tan, Yongliang Shen, Ren Kan, Dongsheng Li, Deqing Yang

EasyTool purifies essential information from extensive tool documentation of different sources, and elaborates a unified interface (i. e., tool instruction) to offer standardized tool descriptions and functionalities for LLM-based agents.

Self-Contrast: Better Reflection Through Inconsistent Solving Perspectives

no code implementations4 Jan 2024 Wenqi Zhang, Yongliang Shen, Linjuan Wu, Qiuying Peng, Jun Wang, Yueting Zhuang, Weiming Lu

Experiments conducted on a series of reasoning and translation tasks with different LLMs serve to underscore the effectiveness and generality of our strategy.

Language Modeling Language Modelling +1

TaskBench: Benchmarking Large Language Models for Task Automation

1 code implementation30 Nov 2023 Yongliang Shen, Kaitao Song, Xu Tan, Wenqi Zhang, Kan Ren, Siyu Yuan, Weiming Lu, Dongsheng Li, Yueting Zhuang

To address this, we introduce TaskBench, a comprehensive framework to evaluate the capability of LLMs in task automation.

Benchmarking Parameter Prediction

An Expression Tree Decoding Strategy for Mathematical Equation Generation

1 code implementation14 Oct 2023 Wenqi Zhang, Yongliang Shen, Qingpeng Nong, Zeqi Tan, Yanna Ma, Weiming Lu

To generate a tree with expression as its node, we employ a layer-wise parallel decoding strategy: we decode multiple independent expressions (leaf nodes) in parallel at each layer and repeat parallel decoding layer by layer to sequentially generate these parent node expressions that depend on others.

Math Math Word Problem Solving +1

MProto: Multi-Prototype Network with Denoised Optimal Transport for Distantly Supervised Named Entity Recognition

1 code implementation12 Oct 2023 Shuhui Wu, Yongliang Shen, Zeqi Tan, Wenqi Ren, Jietian Guo, ShiLiang Pu, Weiming Lu

Distantly supervised named entity recognition (DS-NER) aims to locate entity mentions and classify their types with only knowledge bases or gazetteers and unlabeled corpus.

named-entity-recognition Named Entity Recognition +1

Data-Copilot: Bridging Billions of Data and Humans with Autonomous Workflow

1 code implementation12 Jun 2023 Wenqi Zhang, Yongliang Shen, Weiming Lu, Yueting Zhuang

The advancements are twofold: First, it is a code-centric agent that receives human requests and generates code as an intermediary to handle massive data, which is quite flexible for large-scale data processing tasks.

DiffusionNER: Boundary Diffusion for Named Entity Recognition

2 code implementations22 May 2023 Yongliang Shen, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, Yueting Zhuang

In this paper, we propose DiffusionNER, which formulates the named entity recognition task as a boundary-denoising diffusion process and thus generates named entities from noisy spans.

Chinese Named Entity Recognition Denoising +4

Insert or Attach: Taxonomy Completion via Box Embedding

1 code implementation18 May 2023 Wei Xue, Yongliang Shen, Wenqi Ren, Jietian Guo, ShiLiang Pu, Weiming Lu

Taxonomy completion, enriching existing taxonomies by inserting new concepts as parents or attaching them as children, has gained significant interest.

Multi-View Reasoning: Consistent Contrastive Learning for Math Word Problem

1 code implementation21 Oct 2022 Wenqi Zhang, Yongliang Shen, Yanna Ma, Xiaoxia Cheng, Zeqi Tan, Qingpeng Nong, Weiming Lu

Math word problem solver requires both precise relation reasoning about quantities in the text and reliable generation for the diverse equation.

Ranked #2 on Math Word Problem Solving on Math23K (using extra training data)

Contrastive Learning Math +3

Molecular Substructure-Aware Network for Drug-Drug Interaction Prediction

1 code implementation24 Aug 2022 Xinyu Zhu, Yongliang Shen, Weiming Lu

Concomitant administration of drugs can cause drug-drug interactions (DDIs).

Prompting to Distill: Boosting Data-Free Knowledge Distillation via Reinforced Prompt

no code implementations16 May 2022 Xinyin Ma, Xinchao Wang, Gongfan Fang, Yongliang Shen, Weiming Lu

Data-free knowledge distillation (DFKD) conducts knowledge distillation via eliminating the dependence of original training data, and has recently achieved impressive results in accelerating pre-trained language models.

Data-free Knowledge Distillation

Propose-and-Refine: A Two-Stage Set Prediction Network for Nested Named Entity Recognition

1 code implementation27 Apr 2022 Shuhui Wu, Yongliang Shen, Zeqi Tan, Weiming Lu

In the refine stage, proposals interact with each other, and richer contextual information is incorporated into the proposal representations.

named-entity-recognition Named Entity Recognition +3

Towards Data-Efficient Detection Transformers

2 code implementations17 Mar 2022 Wen Wang, Jing Zhang, Yang Cao, Yongliang Shen, DaCheng Tao

Besides, we introduce a simple yet effective label augmentation method to provide richer supervision and improve data efficiency.

Heterogeneous Graph Neural Networks for Concept Prerequisite Relation Learning in Educational Data

no code implementations NAACL 2021 Chenghao Jia, Yongliang Shen, Yechun Tang, Lu Sun, Weiming Lu

Prerequisite relations among concepts are crucial for educational applications, such as curriculum planning and intelligent tutoring.

Relation

A Sequence-to-Set Network for Nested Named Entity Recognition

1 code implementation19 May 2021 Zeqi Tan, Yongliang Shen, Shuai Zhang, Weiming Lu, Yueting Zhuang

We utilize a non-autoregressive decoder to predict the final set of entities in one pass, in which we are able to capture dependencies between entities.

Decoder named-entity-recognition +3

Locate and Label: A Two-stage Identifier for Nested Named Entity Recognition

1 code implementation ACL 2021 Yongliang Shen, Xinyin Ma, Zeqi Tan, Shuai Zhang, Wen Wang, Weiming Lu

Although these methods have the innate ability to handle nested NER, they suffer from high computational cost, ignorance of boundary information, under-utilization of the spans that partially match with entities, and difficulties in long entity recognition.

Chinese Named Entity Recognition named-entity-recognition +3

A Trigger-Sense Memory Flow Framework for Joint Entity and Relation Extraction

1 code implementation25 Jan 2021 Yongliang Shen, Xinyin Ma, Yechun Tang, Weiming Lu

Joint entity and relation extraction framework constructs a unified model to perform entity recognition and relation extraction simultaneously, which can exploit the dependency between the two tasks to mitigate the error propagation problem suffered by the pipeline model.

 Ranked #1 on Relation Extraction on CoNLL04 (NER Micro F1 metric)

Joint Entity and Relation Extraction Reading Comprehension +2

Multi-hop Reading Comprehension across Documents with Path-based Graph Convolutional Network

no code implementations11 Jun 2020 Zeyun Tang, Yongliang Shen, Xinyin Ma, Wei Xu, Jiale Yu, Weiming Lu

Meanwhile, we propose Gated-RGCN to accumulate evidence on the path-based reasoning graph, which contains a new question-aware gating mechanism to regulate the usefulness of information propagating across documents and add question information during reasoning.

Multi-Hop Reading Comprehension

Cannot find the paper you are looking for? You can Submit a new open access paper.