Advances in Manufacturing ›› 2025, Vol. 13 ›› Issue (3): 525-538.doi: 10.1007/s40436-024-00519-8

Previous Articles    

AI-enabled intelligent cockpit proactive affective interaction: middle-level feature fusion dual-branch deep learning network for driver emotion recognition

Ying-Zhang Wu1, Wen-Bo Li1, Yu-Jing Liu1, Guan-Zhong Zeng2, Cheng-Mou Li1, Hua-Min Jin3, Shen Li4, Gang Guo1   

  1. 1. College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, 400044, People's Republic of China;
    2. Hikvision Research Institute, Hangzhou, 311599, People's Republic of China;
    3. China Society of Automotive Engineers, Beijing, 100021, People's Republic of China;
    4. Department of Civil Engineering, Tsinghua University, Beijing, 100084, People's Republic of China
  • Received:2023-11-01 Revised:2023-11-23 Published:2025-09-19
  • Supported by:
    This work is supported by the National Natural Science Foundation of China (Grant No. 52302497).

Abstract: Advances in artificial intelligence (AI) technology are propelling the rapid development of automotive intelligent cockpits. The active perception of driver emotions significantly impacts road traffic safety. Consequently, the development of driver emotion recognition technology is crucial for ensuring driving safety in the advanced driver assistance system (ADAS) of the automotive intelligent cockpit. The ongoing advancements in AI technology offer a compelling avenue for implementing proactive affective interaction technology. This study introduced the multimodal driver emotion recognition network (MDERNet), a dual-branch deep learning network that temporally fused driver facial expression features and driving behavior features for non-contact driver emotion recognition. The proposed model was validated on publicly available datasets such as CK+, RAVDESS, DEAP, and PPB-Emo, recognizing discrete and dimensional emotions. The results indicated that the proposed model demonstrated advanced recognition performance, and ablation experiments confirmed the significance of various model components. The proposed method serves as a fundamental reference for multimodal feature fusion in driver emotion recognition and contributes to the advancement of ADAS within automotive intelligent cockpits.

The full text can be downloaded at https://link.springer.com/article/10.1007/s40436-024-00519-8

Key words: Driver emotion, Artificial intelligence (AI), Facial expression, Driving behavior, Intelligent cockpit