Search Results for author: Jinming Zhao

Found 21 papers, 15 papers with code

M3ED: Multi-modal Multi-scene Multi-label Emotional Dialogue Database

1 code implementation ACL 2022 Jinming Zhao, Tenggan Zhang, Jingwen Hu, Yuchen Liu, Qin Jin, Xinchao Wang, Haizhou Li

In this work, we propose a Multi-modal Multi-scene Multi-label Emotional Dialogue dataset, M3ED, which contains 990 dyadic emotional dialogues from 56 different TV series, a total of 9, 082 turns and 24, 449 utterances.

Cultural Vocal Bursts Intensity Prediction Emotion Recognition

MMGCN: Multimodal Fusion via Deep Graph Convolution Network for Emotion Recognition in Conversation

1 code implementation ACL 2021 Jingwen Hu, Yuchen Liu, Jinming Zhao, Qin Jin

Emotion recognition in conversation (ERC) is a crucial component in affective dialogue systems, which helps the system understand users' emotions and generate empathetic responses.

Emotion Recognition in Conversation

Missing Modality Imagination Network for Emotion Recognition with Uncertain Missing Modalities

1 code implementation ACL 2021 Jinming Zhao, Ruichen Li, Qin Jin

However, in real-world applications, we often encounter the problem of missing modality, and which modalities will be missing is uncertain.

Emotion Recognition

SummPip: Unsupervised Multi-Document Summarization with Sentence Graph Compression

1 code implementation17 Jul 2020 Jinming Zhao, Ming Liu, Longxiang Gao, Yuan Jin, Lan Du, He Zhao, He Zhang, Gholamreza Haffari

Obtaining training data for multi-document summarization (MDS) is time consuming and resource-intensive, so recent neural models can only be trained for limited domains.

Clustering Document Summarization +2

Towards Relation Extraction From Speech

1 code implementation17 Oct 2022 Tongtong Wu, Guitao Wang, Jinming Zhao, Zhaoran Liu, Guilin Qi, Yuan-Fang Li, Gholamreza Haffari

We explore speech relation extraction via two approaches: the pipeline approach conducting text-based extraction with a pretrained ASR module, and the end2end approach via a new proposed encoder-decoder model, or what we called SpeechRE.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

Multi-Task Learning Framework for Emotion Recognition in-the-wild

1 code implementation19 Jul 2022 Tenggan Zhang, Chuanhe Liu, Xiaolong Liu, Yuchen Liu, Liyu Meng, Lei Sun, Wenqiang Jiang, Fengyuan Zhang, Jinming Zhao, Qin Jin

This paper presents our system for the Multi-Task Learning (MTL) Challenge in the 4th Affective Behavior Analysis in-the-wild (ABAW) competition.

Emotion Recognition Multi-Task Learning +1

Generating Synthetic Speech from SpokenVocab for Speech Translation

1 code implementation15 Oct 2022 Jinming Zhao, Gholamreza Haffar, Ehsan Shareghi

Training end-to-end speech translation (ST) systems requires sufficiently large-scale data, which is unavailable for most language pairs and domains.

Data Augmentation Machine Translation +1

M-Adapter: Modality Adaptation for End-to-End Speech-to-Text Translation

1 code implementation3 Jul 2022 Jinming Zhao, Hao Yang, Ehsan Shareghi, Gholamreza Haffari

End-to-end speech-to-text translation models are often initialized with pre-trained speech encoder and pre-trained text decoder.

Speech-to-Text Translation Translation

Investigating Pre-trained Audio Encoders in the Low-Resource Condition

1 code implementation28 May 2023 Hao Yang, Jinming Zhao, Gholamreza Haffari, Ehsan Shareghi

Pre-trained speech encoders have been central to pushing state-of-the-art results across various speech understanding and generation tasks.

MEmoBERT: Pre-training Model with Prompt-based Learning for Multimodal Emotion Recognition

no code implementations27 Oct 2021 Jinming Zhao, Ruichen Li, Qin Jin, Xinchao Wang, Haizhou Li

Multimodal emotion recognition study is hindered by the lack of labelled corpora in terms of scale and diversity, due to the high annotation cost and label ambiguity.

Emotion Classification Multimodal Emotion Recognition +1

RedApt: An Adaptor for wav2vec 2 Encoding \\ Faster and Smaller Speech Translation without Quality Compromise

no code implementations16 Oct 2022 Jinming Zhao, Hao Yang, Gholamreza Haffari, Ehsan Shareghi

Pre-trained speech Transformers in speech translation (ST) have facilitated state-of-the-art (SotA) results; yet, using such encoders is computationally expensive.

Translation

Towards Event Extraction from Speech with Contextual Clues

1 code implementation27 Jan 2024 Jingqi Kang, Tongtong Wu, Jinming Zhao, Guitao Wang, Guilin Qi, Yuan-Fang Li, Gholamreza Haffari

While text-based event extraction has been an active research area and has seen successful application in many domains, extracting semantic events from speech directly is an under-explored problem.

Event Extraction speech-recognition +1

Double Mixture: Towards Continual Event Detection from Speech

no code implementations20 Apr 2024 Jingqi Kang, Tongtong Wu, Jinming Zhao, Guitao Wang, Yinwei Wei, Hao Yang, Guilin Qi, Yuan-Fang Li, Gholamreza Haffari

To address the challenges of catastrophic forgetting and effective disentanglement, we propose a novel method, 'Double Mixture.'

Cannot find the paper you are looking for? You can Submit a new open access paper.