Search Results for author: Fei Zhao

Found 16 papers, 12 papers with code

Learning from Adjective-Noun Pairs: A Knowledge-enhanced Framework for Target-Oriented Multimodal Sentiment Classification

1 code implementation COLING 2022 Fei Zhao, Zhen Wu, Siyu Long, Xinyu Dai, ShuJian Huang, Jiajun Chen

Target-oriented multimodal sentiment classification (TMSC) is a new subtask of aspect-based sentiment analysis, which aims to determine the sentiment polarity of the opinion target mentioned in a (sentence, image) pair.

Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +2

AlignGPT: Multi-modal Large Language Models with Adaptive Alignment Capability

no code implementations23 May 2024 Fei Zhao, Taotian Pang, Chunhui Li, Zhen Wu, Junjie Guo, Shangyu Xing, Xinyu Dai

In the pre-training stage, instead of treating all image-text pairs equally, we assign different levels of alignment capabilities to different image-text pairs.

Language Modelling Large Language Model +1

Knowledge-aware Dual-side Attribute-enhanced Recommendation

1 code implementation24 Mar 2024 Taotian Pang, Xingyu Lou, Fei Zhao, Zhen Wu, Kuiyao Dong, Qiuying Peng, Yue Qi, Xinyu Dai

Specifically, we build \textit{user preference representations} and \textit{attribute fusion representations} upon the attribute information in knowledge graphs, which are utilized to enhance \textit{collaborative filtering} (CF) based user and item representations, respectively.

Attribute Collaborative Filtering +3

Cobra Effect in Reference-Free Image Captioning Metrics

no code implementations18 Feb 2024 Zheng Ma, Changxin Wang, Yawen Ouyang, Fei Zhao, Jianbing Zhang, ShuJian Huang, Jiajun Chen

If a certain metric has flaws, it will be exploited by the model and reflected in the generated sentences.

Image Captioning

EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models

no code implementations15 Feb 2024 Shangyu Xing, Fei Zhao, Zhen Wu, Tuo An, WeiHao Chen, Chunhui Li, Jianbing Zhang, Xinyu Dai

Multimodal large language models (MLLMs) have attracted increasing attention in the past few years, but they may still generate descriptions that include objects not present in the corresponding images, a phenomenon known as object hallucination.


DRIN: Dynamic Relation Interactive Network for Multimodal Entity Linking

1 code implementation9 Oct 2023 Shangyu Xing, Fei Zhao, Zhen Wu, Chunhui Li, Jianbing Zhang, Xinyu Dai

Multimodal Entity Linking (MEL) is a task that aims to link ambiguous mentions within multimodal contexts to referential entities in a multimodal knowledge base.

Entity Linking Relation

Dynamic Demonstrations Controller for In-Context Learning

1 code implementation30 Sep 2023 Fei Zhao, Taotian Pang, Zhen Wu, Zheng Ma, ShuJian Huang, Xinyu Dai

Previous studies have revealed that ICL is sensitive to the selection and the ordering of demonstrations.

In-Context Learning Language Modelling +1

Measuring Your ASTE Models in The Wild: A Diversified Multi-domain Dataset For Aspect Sentiment Triplet Extraction

1 code implementation27 May 2023 Ting Xu, Huiyun Yang, Zhen Wu, Jiaze Chen, Fei Zhao, Xinyu Dai

In this paper, we introduce a new dataset, named DMASTE, which is manually annotated to better fit real-world scenarios by providing more diverse and realistic reviews for the task.

Aspect Sentiment Triplet Extraction

Label-Driven Denoising Framework for Multi-Label Few-Shot Aspect Category Detection

1 code implementation9 Oct 2022 Fei Zhao, Yuchen Shen, Zhen Wu, Xinyu Dai

Multi-Label Few-Shot Aspect Category Detection (FS-ACD) is a new sub-task of aspect-based sentiment analysis, which aims to detect aspect categories accurately with limited training instances.

Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +2

Analysing Wideband Absorbance Immittance in Normal and Ears with Otitis Media with Effusion Using Machine Learning

no code implementations4 Mar 2021 Emad M. Grais, Xiaoya Wang, Jie Wang, Fei Zhao, Wen Jiang, Yuexin Cai, Lifang Zhang, Qingwen Lin, Haidi Yang

Wideband Absorbance Immittance (WAI) has been available for more than a decade, however its clinical use still faces the challenges of limited understanding and poor interpretation of WAI results.

BIG-bench Machine Learning

Attention Transfer Network for Aspect-level Sentiment Classification

1 code implementation COLING 2020 Fei Zhao, Zhen Wu, Xinyu Dai

In neural network-based methods for ASC, most works employ the attention mechanism to capture the corresponding sentiment words of the opinion target, then aggregate them as evidence to infer the sentiment of the target.

Classification General Classification +3

Grid Tagging Scheme for Aspect-oriented Fine-grained Opinion Extraction

3 code implementations Findings of the Association for Computational Linguistics 2020 Zhen Wu, Chengcan Ying, Fei Zhao, Zhifang Fan, Xinyu Dai, Rui Xia

To validate the feasibility and compatibility of GTS, we implement three different GTS models respectively based on CNN, BiLSTM, and BERT, and conduct experiments on the aspect-oriented opinion pair extraction and opinion triplet extraction datasets.

Aspect-Sentiment-Opinion Triplet Extraction Aspect Sentiment Triplet Extraction

Latent Opinions Transfer Network for Target-Oriented Opinion Words Extraction

1 code implementation7 Jan 2020 Zhen Wu, Fei Zhao, Xin-yu Dai, Shu-Jian Huang, Jia-Jun Chen

In this paper, we propose a novel model to transfer these opinions knowledge from resource-rich review sentiment classification datasets to low-resource task TOWE.

Aspect-oriented Opinion Extraction General Classification +3

Multi-Band Multi-Resolution Fully Convolutional Neural Networks for Singing Voice Separation

1 code implementation21 Oct 2019 Emad M. Grais, Fei Zhao, Mark D. Plumbley

In the spectrogram of a mixture of singing voices and music signals, there is usually more information about the voice in the low frequency bands than the high frequency bands.

Dimensionality Reduction

Cannot find the paper you are looking for? You can Submit a new open access paper.