Research Article

Semisupervised Deep Features of Time-Frequency Maps for Multimodal Emotion Recognition

Table 1

Summary of multimodal emotion recognition from biological signals.

ReferencesModalityDatasetDomain analysisFusion/classification

[7]EEGDEAPTime domainAfter feature extraction from each modal/decision tree
PPSMAHNOB-HCI

[8]EEGDEAPTime-frequency domain (MSST) for EEGAfter feature extraction from each modal/CNN
PPSMAHNOB-HCITime domain for PPS
VideoTime and frequency domains for video

[9]EEGDEAPTime domainBefore feature extraction/ensemble CNN
PPS

[10]EEGDEAPTime domainAfter feature extraction from each modal/MLP
PPS

[11]EEGDEAPTime and frequency domains for different modalitiesAfter feature extraction from each modal/CNN
PPS

[12]PPSDEAPTime-frequency domainAfter computing the CWT of each modal/CNN

[17]EEGDEAPTime domainAfter feature extraction from each modal by CNN/MLP
VideoMAHNOB-HCI

[19]EEGDEAPTime domainAfter feature extraction from each modal by CNN/Decision tree
GSRLUMED-2
Video

[20]EEGDEAPTime domainAfter feature extraction from each modal by 3D-CNN/MLP
Video

[22]EEGPrivateFrequency domainAfter feature extraction from each modal/CNN
Audio