no code implementations • 27 Jul 2023 • Yu-Ting Lan, Kan Ren, Yansen Wang, Wei-Long Zheng, Dongsheng Li, Bao-liang Lu, Lili Qiu
Seeing is believing, however, the underlying mechanism of how human visual perceptions are intertwined with our cognitions is still a mystery.
no code implementations • 4 Apr 2020 • Xun Wu, Wei-Long Zheng, Bao-liang Lu
The discrimination ability of the EEG connectivity features in emotion recognition is evaluated on three public emotion EEG datasets: SEED, SEED-V, and DEAP.
1 code implementation • 13 Aug 2019 • Wei Liu, Jie-Lin Qiu, Wei-Long Zheng, Bao-liang Lu
We evaluate the performance of DCCA on five multimodal datasets: the SEED, SEED-IV, SEED-V, DEAP, and DREAMER datasets.
no code implementations • 25 Apr 2017 • Changde Du, Changying Du, Jinpeng Li, Wei-Long Zheng, Bao-liang Lu, Huiguang He
In this paper, we first build a multi-view deep generative model to simulate the generative process of multi-modality emotional data.
no code implementations • 26 Feb 2016 • Wei Liu, Wei-Long Zheng, Bao-liang Lu
To enhance the performance of affective models and reduce the cost of acquiring physiological signals for real-world applications, we adopt multimodal deep learning approach to construct affective models from multiple physiological signals.
no code implementations • 10 Jan 2016 • Wei-Long Zheng, Jia-Yi Zhu, Bao-liang Lu
In this paper, we investigate stable patterns of electroencephalogram (EEG) over time for emotion recognition using a machine learning approach.
no code implementations • IEEE Transactions on Autonomous Mental Development ( Volume: 7 , Issue: 3 , Sept. 2015 ) 2015 • Wei-Long Zheng, Bao-liang Lu
To investigate critical frequency bands and channels, this paper introduces deep belief networks (DBNs) to constructing EEG-based emotion recognition models for three emotions: positive, neutral and negative.
Ranked #1 on Electroencephalogram (EEG) on SEED