Search Results for author: Longtao Huang

Found 25 papers, 18 papers with code

Correlation-Aware Graph Convolutional Networks for Multi-Label Node Classification

1 code implementation26 Nov 2024 Yuanchen Bei, Weizhi Chen, Hao Chen, Sheng Zhou, Carl Yang, Jiapei Fan, Longtao Huang, Jiajun Bu

Multi-label node classification is an important yet under-explored domain in graph mining as many real-world nodes belong to multiple categories rather than just a single one.

Classification Graph Mining +1

Towards Rehearsal-Free Multilingual ASR: A LoRA-based Case Study on Whisper

no code implementations20 Aug 2024 Tianyi Xu, Kaixun Huang, Pengcheng Guo, Yu Zhou, Longtao Huang, Hui Xue, Lei Xie

Pre-trained multilingual speech foundation models, like Whisper, have shown impressive performance across different languages.

On the Role of Long-tail Knowledge in Retrieval Augmented Large Language Models

no code implementations24 Jun 2024 Dongyang Li, Junbing Yan, Taolin Zhang, Chengyu Wang, Xiaofeng He, Longtao Huang, Hui Xue, Jun Huang

Retrieval augmented generation (RAG) exhibits outstanding performance in promoting the knowledge capabilities of large language models (LLMs) with retrieved documents related to user queries.

RAG Retrieval +1

UniPSDA: Unsupervised Pseudo Semantic Data Augmentation for Zero-Shot Cross-Lingual Natural Language Understanding

1 code implementation24 Jun 2024 Dongyang Li, Taolin Zhang, Jiali Deng, Longtao Huang, Chengyu Wang, Xiaofeng He, Hui Xue

Specifically, to retrieve the tokens with similar meanings for the semantic data augmentation across different languages, we propose a sequential clustering process in 3 stages: within a single language, across multiple languages of a language family, and across languages from multiple language families.

Data Augmentation Natural Language Understanding +2

KEHRL: Learning Knowledge-Enhanced Language Representations with Hierarchical Reinforcement Learning

1 code implementation24 Jun 2024 Dongyang Li, Taolin Zhang, Longtao Huang, Chengyu Wang, Xiaofeng He, Hui Xue

Knowledge-enhanced pre-trained language models (KEPLMs) leverage relation triples from knowledge graphs (KGs) and integrate these external data sources into language models via self-supervised learning.

Hierarchical Reinforcement Learning Knowledge Graphs +5

NoisyGL: A Comprehensive Benchmark for Graph Neural Networks under Label Noise

1 code implementation6 Jun 2024 Zhonghao Wang, Danyu Sun, Sheng Zhou, Haobo Wang, Jiapei Fan, Longtao Huang, Jiajun Bu

However, due to variations in dataset selection, data splitting, and preprocessing techniques, the community currently lacks a comprehensive benchmark, which impedes deeper understanding and further development of GLN.

Node Classification

DAFNet: Dynamic Auxiliary Fusion for Sequential Model Editing in Large Language Models

1 code implementation31 May 2024 Taolin Zhang, Qizhou Chen, Dongyang Li, Chengyu Wang, Xiaofeng He, Longtao Huang, Hui Xue, Jun Huang

(2) Considering that auxiliary parameters are required to store the knowledge for sequential editing, we construct a new dataset named \textbf{DAFSet}, fulfilling recent, popular, long-tail and robust properties to enhance the generality of sequential editing.

Hallucination Model Editing

Lifelong Knowledge Editing for LLMs with Retrieval-Augmented Continuous Prompt Learning

1 code implementation6 May 2024 Qizhou Chen, Taolin Zhang, Xiaofeng He, Dongyang Li, Chengyu Wang, Longtao Huang, Hui Xue

Model editing aims to correct outdated or erroneous knowledge in large language models (LLMs) without the need for costly retraining.

knowledge editing Retrieval

R4: Reinforced Retriever-Reorder-Responder for Retrieval-Augmented Large Language Models

no code implementations4 May 2024 Taolin Zhang, Dongyang Li, Qizhou Chen, Chengyu Wang, Longtao Huang, Hui Xue, Xiaofeng He, Jun Huang

The reordering learning process is divided into two steps according to the quality of the generated responses: document order adjustment and document representation enhancement.

Graph Attention Hallucination +5

General Phrase Debiaser: Debiasing Masked Language Models at a Multi-Token Level

1 code implementation23 Nov 2023 Bingkang Shi, Xiaodan Zhang, Dehan Kong, Yulei Wu, Zongzhen Liu, Honglei Lyu, Longtao Huang

The social biases and unwelcome stereotypes revealed by pretrained language models are becoming obstacles to their application.

From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework

1 code implementation29 May 2023 Yangyi Chen, Hongcheng Gao, Ganqu Cui, Lifan Yuan, Dehan Kong, Hanlu Wu, Ning Shi, Bo Yuan, Longtao Huang, Hui Xue, Zhiyuan Liu, Maosong Sun, Heng Ji

In our experiments, we conduct a robustness evaluation of RoBERTa models to demonstrate the effectiveness of our evaluation framework, and further show the rationality of each component in the framework.

Adversarial Attack

Large Language Models Can be Lazy Learners: Analyze Shortcuts in In-Context Learning

no code implementations26 May 2023 Ruixiang Tang, Dehan Kong, Longtao Huang, Hui Xue

Large language models (LLMs) have recently shown great potential for in-context learning, where LLMs learn a new task simply by conditioning on a few input-label pairs (prompts).

In-Context Learning

Decoder Tuning: Efficient Language Understanding as Decoding

3 code implementations16 Dec 2022 Ganqu Cui, Wentao Li, Ning Ding, Longtao Huang, Zhiyuan Liu, Maosong Sun

With the evergrowing sizes of pre-trained models (PTMs), it has been an emerging practice to only provide the inference APIs for users, namely model-as-a-service (MaaS) setting.

Decoder Natural Language Understanding

Text Editing as Imitation Game

1 code implementation21 Oct 2022 Ning Shi, Bin Tang, Bo Yuan, Longtao Huang, Yewen Pu, Jie Fu, Zhouhan Lin

Text editing, such as grammatical error correction, arises naturally from imperfect textual data.

Action Generation Grammatical Error Correction +1

Syntax-guided Localized Self-attention by Constituency Syntactic Distance

1 code implementation21 Oct 2022 Shengyuan Hou, Jushi Kai, Haotian Xue, Bingyu Zhu, Bo Yuan, Longtao Huang, Xinbing Wang, Zhouhan Lin

Recent works have revealed that Transformers are implicitly learning the syntactic information in its lower layers from data, albeit is highly dependent on the quality and scale of the training data.

Machine Translation Translation

Why Should Adversarial Perturbations be Imperceptible? Rethink the Research Paradigm in Adversarial NLP

2 code implementations19 Oct 2022 Yangyi Chen, Hongcheng Gao, Ganqu Cui, Fanchao Qi, Longtao Huang, Zhiyuan Liu, Maosong Sun

We discuss the deficiencies in previous work and propose our suggestions that the research on the Security-oriented adversarial NLP (SoadNLP) should: (1) evaluate their methods on security tasks to demonstrate the real-world concerns; (2) consider real-world attackers' goals, instead of developing impractical methods.

Data Augmentation

RMGN: A Regional Mask Guided Network for Parser-free Virtual Try-on

2 code implementations24 Apr 2022 Chao Lin, Zhao Li, Sheng Zhou, Shichang Hu, Jialun Zhang, Linhao Luo, Jiarun Zhang, Longtao Huang, Yuan He

Virtual try-on(VTON) aims at fitting target clothes to reference person images, which is widely adopted in e-commerce. Existing VTON approaches can be narrowly categorized into Parser-Based(PB) and Parser-Free(PF) by whether relying on the parser information to mask the persons' clothes and synthesize try-on images.

Virtual Try-on

Prototypical Verbalizer for Prompt-based Few-shot Tuning

1 code implementation ACL 2022 Ganqu Cui, Shengding Hu, Ning Ding, Longtao Huang, Zhiyuan Liu

However, manual verbalizers heavily depend on domain-specific prior knowledge and human efforts, while finding appropriate label words automatically still remains challenging. In this work, we propose the prototypical verbalizer (ProtoVerb) which is built directly from training data.

Contrastive Learning Entity Typing +2

Automated Strabismus Detection for Telemedicine Applications

1 code implementation9 Sep 2018 Jiewei Lu, Zhun Fan, Ce Zheng, Jingan Feng, Longtao Huang, Wenji Li, Erik D. Goodman

Telemedicine, which has great potential to alleviate the growing demand of the diagnosis of ophthalmologic diseases, is an effective method to achieve timely strabismus detection.

Cannot find the paper you are looking for? You can Submit a new open access paper.