Exploring temporal representations by leveraging attention-based bidirectional LSTM-RNNs for multi-modal emotion recognition
作者:
Highlights:
• Compared with hand-craft features, the proposed method can automatically extract temporal features from raw physiological signals by Attention-based BLSTM-RNNs, which is capable of learning feature representations and modeling the temporal dependencies between their activation.
• We also investigate the usage of attention-based architectures to improve physiological-based emotion recognition. The attention mechanism allows the network to focus on the emotionally salient parts of a sequence.
• Decision level fusion is implemented to capture complementary information from different modalities for enhancing emotion recognition system.
摘要
•Compared with hand-craft features, the proposed method can automatically extract temporal features from raw physiological signals by Attention-based BLSTM-RNNs, which is capable of learning feature representations and modeling the temporal dependencies between their activation.•We also investigate the usage of attention-based architectures to improve physiological-based emotion recognition. The attention mechanism allows the network to focus on the emotionally salient parts of a sequence.•Decision level fusion is implemented to capture complementary information from different modalities for enhancing emotion recognition system.
论文关键词:Emotion recognition,EEG signals,Physiological signals,Deep learning,Multimedia content,Multi-modal fusion,00-01,99-00
论文评审过程:Received 29 July 2019, Revised 15 November 2019, Accepted 16 December 2019, Available online 3 January 2020, Version of Record 3 January 2020.
论文官网地址:https://doi.org/10.1016/j.ipm.2019.102185