Search Results for author: Yanyan Zhao

Found 27 papers, 7 papers with code

Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors

1 code implementation Findings (ACL) 2022 Yang Wu, Yanyan Zhao, Hao Yang, Song Chen, Bing Qin, Xiaohuan Cao, Wenting Zhao

Through further analysis of the ASR outputs, we find that in some cases the sentiment words, the key sentiment elements in the textual modality, are recognized as other words, which makes the sentiment of the text change and hurts the performance of multimodal sentiment models directly.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

An Early Evaluation of GPT-4V(ision)

1 code implementation25 Oct 2023 Yang Wu, Shilong Wang, Hao Yang, Tian Zheng, Hongbo Zhang, Yanyan Zhao, Bing Qin

In this paper, we evaluate different abilities of GPT-4V including visual understanding, language understanding, visual puzzle solving, and understanding of other modalities such as depth, thermal, video, and audio.

Math

Knowledge-Bridged Causal Interaction Network for Causal Emotion Entailment

1 code implementation6 Dec 2022 Weixiang Zhao, Yanyan Zhao, Zhuojun Li, Bing Qin

Moreover, social-interaction CSK serves as emotion-level bridge (E-bridge) and action-level bridge (A-bridge) to connect candidate utterances with the target one, which provides explicit causal clues for the Emotional Interaction module and Actional Interaction module to reason the target emotion.

Causal Emotion Entailment Graph Attention

TransESC: Smoothing Emotional Support Conversation via Turn-Level State Transition

1 code implementation5 May 2023 Weixiang Zhao, Yanyan Zhao, Shilong Wang, Bing Qin

Specifically, we construct the state transition graph with a two-step way, named transit-then-interact, to grasp such three types of turn-level transition information.

MuCDN: Mutual Conversational Detachment Network for Emotion Recognition in Multi-Party Conversations

1 code implementation COLING 2022 Weixiang Zhao, Yanyan Zhao, Bing Qin

Specifically, two detachment ways are devised to perform context and speaker-specific modeling within detached threads and they are bridged through a mutual module.

Emotion Recognition

Exploring Periodicity and Interactivity in Multi-Interest Framework for Sequential Recommendation

1 code implementation7 Jun 2021 Gaode Chen, Xinghua Zhang, Yanyan Zhao, Cong Xue, Ji Xiang

Meanwhile, an ingenious graph is proposed to enhance the interactivity between items in user's behavior sequence, which can capture both global and local item features.

Sequential Recommendation

Don't Lose Yourself! Empathetic Response Generation via Explicit Self-Other Awareness

1 code implementation8 Oct 2022 Weixiang Zhao, Yanyan Zhao, Xin Lu, Bing Qin

As a critical step to achieve human-like chatbots, empathetic response generation has attained increasing interests.

Empathetic Response Generation Response Generation

An Iterative Emotion Interaction Network for Emotion Recognition in Conversations

no code implementations COLING 2020 Xin Lu, Yanyan Zhao, Yang Wu, Yijian Tian, Huipeng Chen, Bing Qin

We noticed that the gold emotion labels of the context utterances can provide explicit and accurate emotion interaction, but it is impossible to input gold labels at inference time.

Emotion Recognition in Conversation

Learning to Share by Masking the Non-shared for Multi-domain Sentiment Classification

no code implementations17 Apr 2021 Jianhua Yuan, Yanyan Zhao, Bing Qin, Ting Liu

To this end, we propose the BertMasker network which explicitly masks domain-related words from texts, learns domain-invariant sentiment features from these domain-agnostic texts, and uses those masked words to form domain-aware sentence representations.

General Classification Multi-Domain Sentiment Classification +3

MACSA: A Multimodal Aspect-Category Sentiment Analysis Dataset with Multimodal Fine-grained Aligned Annotations

no code implementations28 Jun 2022 Hao Yang, Yanyan Zhao, Jianwei Liu, Yang Wu, Bing Qin

In this paper, we propose a new dataset, the Multimodal Aspect-Category Sentiment Analysis (MACSA) dataset, which contains more than 21K text-image pairs.

Aspect Category Sentiment Analysis Sentiment Analysis

Zero-shot Aspect-level Sentiment Classification via Explicit Utilization of Aspect-to-Document Sentiment Composition

no code implementations6 Sep 2022 Pengfei Deng, Jianhua Yuan, Yanyan Zhao, Bing Qin

Our key intuition is that the sentiment representation of a document is composed of the sentiment representations of all the aspects of that document.

Classification Sentiment Analysis +1

An Efficient End-to-End Transformer with Progressive Tri-modal Attention for Multi-modal Emotion Recognition

no code implementations20 Sep 2022 Yang Wu, Pai Peng, Zhenyu Zhang, Yanyan Zhao, Bing Qin

At the low-level, we propose the progressive tri-modal attention, which can model the tri-modal feature interactions by adopting a two-pass strategy and can further leverage such interactions to significantly reduce the computation and memory complexity through reducing the input token length.

Emotion Recognition

SSR: Utilizing Simplified Stance Reasoning Process for Robust Stance Detection

no code implementations COLING 2022 Jianhua Yuan, Yanyan Zhao, Yanyue Lu, Bing Qin

Motivated by how humans tackle stance detection tasks, we propose to incorporate the stance reasoning process as task knowledge to assist in learning genuine features and reducing reliance on bias features.

Sentence Stance Detection

Debiasing Stance Detection Models with Counterfactual Reasoning and Adversarial Bias Learning

no code implementations20 Dec 2022 Jianhua Yuan, Yanyan Zhao, Bing Qin

Stance detection models may tend to rely on dataset bias in the text part as a shortcut and thus fail to sufficiently learn the interaction between the targets and texts.

counterfactual Counterfactual Inference +2

Is ChatGPT Equipped with Emotional Dialogue Capabilities?

no code implementations19 Apr 2023 Weixiang Zhao, Yanyan Zhao, Xin Lu, Shilong Wang, Yanpeng Tong, Bing Qin

This report presents a study on the emotional dialogue capability of ChatGPT, an advanced language model developed by OpenAI.

Dialogue Understanding Language Modelling

Improving Cross-Task Generalization with Step-by-Step Instructions

no code implementations8 May 2023 Yang Wu, Yanyan Zhao, Zhongyang Li, Bing Qin, Kai Xiong

Instruction tuning has been shown to be able to improve cross-task generalization of language models.

UNIMO-3: Multi-granularity Interaction for Vision-Language Representation Learning

no code implementations23 May 2023 Hao Yang, Can Gao, Hao Líu, Xinyan Xiao, Yanyan Zhao, Bing Qin

The experimental results show that our model achieves state-of-the-art performance in various downstream tasks, and through ablation study can prove that effective cross-layer learning improves the model's ability of multimodal representation.

Representation Learning

SAPT: A Shared Attention Framework for Parameter-Efficient Continual Learning of Large Language Models

no code implementations16 Jan 2024 Weixiang Zhao, Shilong Wang, Yulin Hu, Yanyan Zhao, Bing Qin, Xuanyu Zhang, Qing Yang, Dongliang Xu, Wanxiang Che

Existing methods devise the learning module to acquire task-specific knowledge with parameter-efficient tuning (PET) block and the selection module to pick out the corresponding one for the testing input, aiming at handling the challenges of catastrophic forgetting and knowledge transfer in CL.

Continual Learning Transfer Learning

Both Matter: Enhancing the Emotional Intelligence of Large Language Models without Compromising the General Intelligence

no code implementations15 Feb 2024 Weixiang Zhao, Zhuojun Li, Shilong Wang, Yang Wang, Yulin Hu, Yanyan Zhao, Chen Wei, Bing Qin

Emotional Intelligence (EI), consisting of emotion perception, emotion cognition and emotion expression, plays the critical roles in improving user interaction experience for the current large language model (LLM) based conversational general AI assistants.

Emotional Intelligence Language Modelling +1

How does Architecture Influence the Base Capabilities of Pre-trained Language Models? A Case Study Based on FFN-Wider Transformer Models

no code implementations4 Mar 2024 Xin Lu, Yanyan Zhao, Bing Qin

In this work, we attempt to explain and reverse the decline in base capabilities caused by the architecture of FFN-Wider Transformers, seeking to provide some insights.

Few-Shot Learning Language Modelling +1

Vanilla Transformers are Transfer Capability Teachers

no code implementations4 Mar 2024 Xin Lu, Yanyan Zhao, Bing Qin

However, studies have indicated that MoE Transformers underperform vanilla Transformers in many downstream tasks, significantly diminishing the practical value of MoE models.

Computational Efficiency

Cannot find the paper you are looking for? You can Submit a new open access paper.