TY - JOUR
T1 - Emotion recognition with short-period physiological signals using bimodal sparse autoencoders
AU - Lee, Yun Kyu
AU - Pae, Dong Sung
AU - Hong, Dae Ki
AU - Lim, Myo Taeg
AU - Kang, Tae Koo
N1 - Funding Information:
Funding Statement: This research was supported by the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (NRF-2019R1A2C1089742).
Publisher Copyright:
© 2022, Tech Science Press. All rights reserved.
PY - 2022
Y1 - 2022
N2 - With the advancement of human-computer interaction and artificial intelligence, emotion recognition has received significant research attention. The most commonly used technique for emotion recognition is EEG, which is directly associated with the central nervous system and contains strong emotional features. However, there are some disadvantages to using EEG signals. They require high dimensionality, diverse and complex processing procedures which make real-time computation difficult. In addition, there are problems in data acquisition and interpretation due to body movement or reduced concentration of the experimenter. In this paper, we used photoplethysmography (PPG) and electromyography (EMG) to record signals. Firstly, we segmented the emotion data into 10-pulses during preprocessing to identify emotions with short period signals. These segmented data were input to the proposed bimodal stacked sparse auto-encoder model. To enhance recognition performance, we adopted a bimodal structure to extract shared PPG and EMG representations. This approach provided more detailed arousal-valence mapping compared with the current high/low binary classification. We created a dataset of PPG and EMG signals, called the emotion dataset dividing into four classes to help understand emotion levels. We achieved high performance of 80.18% and 75.86% for arousal and valence, respectively, despite more class classification. Experimental results validated that the proposed method significantly enhanced emotion recognition performance.
AB - With the advancement of human-computer interaction and artificial intelligence, emotion recognition has received significant research attention. The most commonly used technique for emotion recognition is EEG, which is directly associated with the central nervous system and contains strong emotional features. However, there are some disadvantages to using EEG signals. They require high dimensionality, diverse and complex processing procedures which make real-time computation difficult. In addition, there are problems in data acquisition and interpretation due to body movement or reduced concentration of the experimenter. In this paper, we used photoplethysmography (PPG) and electromyography (EMG) to record signals. Firstly, we segmented the emotion data into 10-pulses during preprocessing to identify emotions with short period signals. These segmented data were input to the proposed bimodal stacked sparse auto-encoder model. To enhance recognition performance, we adopted a bimodal structure to extract shared PPG and EMG representations. This approach provided more detailed arousal-valence mapping compared with the current high/low binary classification. We created a dataset of PPG and EMG signals, called the emotion dataset dividing into four classes to help understand emotion levels. We achieved high performance of 80.18% and 75.86% for arousal and valence, respectively, despite more class classification. Experimental results validated that the proposed method significantly enhanced emotion recognition performance.
KW - Bimodal structure network
KW - EDPE dataset
KW - Emotion recognition
KW - Physiological signal
KW - Stacked sparse autoencoder
UR - http://www.scopus.com/inward/record.url?scp=85119824009&partnerID=8YFLogxK
U2 - 10.32604/iasc.2022.020849
DO - 10.32604/iasc.2022.020849
M3 - Article
AN - SCOPUS:85119824009
SN - 1079-8587
VL - 32
SP - 657
EP - 673
JO - Intelligent Automation and Soft Computing
JF - Intelligent Automation and Soft Computing
IS - 2
ER -