Advances in Manufacturing ›› 2025, Vol. 13 ›› Issue (3): 525-538.doi: 10.1007/s40436-024-00519-8

• • 上一篇    

AI-enabled intelligent cockpit proactive affective interaction: middle-level feature fusion dual-branch deep learning network for driver emotion recognition

Ying-Zhang Wu1, Wen-Bo Li1, Yu-Jing Liu1, Guan-Zhong Zeng2, Cheng-Mou Li1, Hua-Min Jin3, Shen Li4, Gang Guo1   

  1. 1. College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, 400044, People's Republic of China;
    2. Hikvision Research Institute, Hangzhou, 311599, People's Republic of China;
    3. China Society of Automotive Engineers, Beijing, 100021, People's Republic of China;
    4. Department of Civil Engineering, Tsinghua University, Beijing, 100084, People's Republic of China
  • 收稿日期:2023-11-01 修回日期:2023-11-23 发布日期:2025-09-19
  • 通讯作者: Wen-Bo Li,E-mail:wenbo_li@cqu.edu.cn E-mail:wenbo_li@cqu.edu.cn
  • 作者简介:Ying-Zhang Wu received a B.S. degree in mechanical engineering from Chongqing University, Chongqing, China, in 2017. He is working toward a Ph.D. with the Advanced Manufacturing and Information Technology Laboratory, College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China. His research interests include intelligent vehicles, intelligent cockpits, driver emotion detection, driving fatigue, human-machine interaction, and brain-computer interface.
    Wen-Bo Li received a B.S., M.Sc., and Ph.D. in automotive engineering from Chongqing University, Chongqing, China, in 2014, 2017, and 2021, respectively. From 2018 to 2020, he was a Visiting Ph.D. Student with the Department of Mechanical and Mechatronics Engineering, University of Waterloo, Waterloo, ON, Canada. From 2021 to 2023, he was a Postdoctoral Research Fellow at the School of Vehicle and Mobility at Tsinghua University, Beijing, China. He is an associate professor at the College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China. His research interests include intelligent vehicles, intelligent cockpits, human emotion and cognition, driver emotion detection and regulation, human-machine interaction, affective computing, and brain–computer interface.
    Yu-Jing Liu received her B.S. degree in industrial design from the College of Mechanical Engineering, Chongqing University, China, in 2020. She is working toward a Ph.D. with the Advanced Manufacturing and Information Technology Laboratory, College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China. Her research interests include human-vehicle interaction design for intelligent cockpit, user experience, and car seat comfort.
    Guan-Zhong Zeng received a B.S. and M.Sc. degree in automotive engineering from Chongqing University, Chongqing, China, in 2018 and 2021. He is an artificial intelligence engineer at Hikvision Research Institute, Hangzhou, China. His research interests include gaze estimation, domain generalization, and domain adaptation.[Inline Image Removed]Cheng-Mou Li received a B.S. degree in mechanical engineering from the College of Mechanical Engineering, Chongqing University, China, in 2020. He is working toward a Ph.D. with the Advanced Manufacturing and Information Technology Laboratory, College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China. His research interests include intelligent transportation systems, human-computer interaction, driver distraction detection, and brain-computer interface.
    Cheng-Mou Li received a B.S. degree in mechanical engineering from the College of Mechanical Engineering, Chongqing University, China, in 2020. He is working toward a Ph.D. with the Advanced Manufacturing and Information Technology Laboratory, College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China. His research interests include intelligent transportation systems, human-computer interaction, driver distraction detection, and brain-computer interface.
    Hua-Min Jin received an M.Sc. degree in Industrial Engineering from Seoul National University, Seoul, Korea, in 2019. She is an intelligent cockpit researcher at the China Society of Automotive Engineers in Beijing, China. Her research interests include intelligent vehicles, cockpits, human factors, user experience, and humancomputer interaction.
    Shen Li received a Ph.D. from the University of Wisconsin– Madison, USA, in 2018. He is a Research Associate at Tsinghua University. His research interests include intelligent transportation systems (ITS), architecture design of CAVH system, vehicle infrastructure cooperative planning and decision method, traffic data mining based on cellular data, and traffic operations and management.
    Gang Guo received a Ph.D. degree in mechanical engineering from Chongqing University, Chongqing, China, in 1994. He is a professor at the College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China. He has authored and co-authored over 100 refereed journal and conference publications. His research interests include human-machine interaction, user experience, intelligent cockpits, intelligent vehicles, brain-computer interfaces, and intelligent manufacturing.
  • 基金资助:
    This work is supported by the National Natural Science Foundation of China (Grant No. 52302497).

AI-enabled intelligent cockpit proactive affective interaction: middle-level feature fusion dual-branch deep learning network for driver emotion recognition

Ying-Zhang Wu1, Wen-Bo Li1, Yu-Jing Liu1, Guan-Zhong Zeng2, Cheng-Mou Li1, Hua-Min Jin3, Shen Li4, Gang Guo1   

  1. 1. College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, 400044, People's Republic of China;
    2. Hikvision Research Institute, Hangzhou, 311599, People's Republic of China;
    3. China Society of Automotive Engineers, Beijing, 100021, People's Republic of China;
    4. Department of Civil Engineering, Tsinghua University, Beijing, 100084, People's Republic of China
  • Received:2023-11-01 Revised:2023-11-23 Published:2025-09-19
  • Supported by:
    This work is supported by the National Natural Science Foundation of China (Grant No. 52302497).

摘要: Advances in artificial intelligence (AI) technology are propelling the rapid development of automotive intelligent cockpits. The active perception of driver emotions significantly impacts road traffic safety. Consequently, the development of driver emotion recognition technology is crucial for ensuring driving safety in the advanced driver assistance system (ADAS) of the automotive intelligent cockpit. The ongoing advancements in AI technology offer a compelling avenue for implementing proactive affective interaction technology. This study introduced the multimodal driver emotion recognition network (MDERNet), a dual-branch deep learning network that temporally fused driver facial expression features and driving behavior features for non-contact driver emotion recognition. The proposed model was validated on publicly available datasets such as CK+, RAVDESS, DEAP, and PPB-Emo, recognizing discrete and dimensional emotions. The results indicated that the proposed model demonstrated advanced recognition performance, and ablation experiments confirmed the significance of various model components. The proposed method serves as a fundamental reference for multimodal feature fusion in driver emotion recognition and contributes to the advancement of ADAS within automotive intelligent cockpits.

The full text can be downloaded at https://link.springer.com/article/10.1007/s40436-024-00519-8

关键词: Driver emotion, Artificial intelligence (AI), Facial expression, Driving behavior, Intelligent cockpit

Abstract: Advances in artificial intelligence (AI) technology are propelling the rapid development of automotive intelligent cockpits. The active perception of driver emotions significantly impacts road traffic safety. Consequently, the development of driver emotion recognition technology is crucial for ensuring driving safety in the advanced driver assistance system (ADAS) of the automotive intelligent cockpit. The ongoing advancements in AI technology offer a compelling avenue for implementing proactive affective interaction technology. This study introduced the multimodal driver emotion recognition network (MDERNet), a dual-branch deep learning network that temporally fused driver facial expression features and driving behavior features for non-contact driver emotion recognition. The proposed model was validated on publicly available datasets such as CK+, RAVDESS, DEAP, and PPB-Emo, recognizing discrete and dimensional emotions. The results indicated that the proposed model demonstrated advanced recognition performance, and ablation experiments confirmed the significance of various model components. The proposed method serves as a fundamental reference for multimodal feature fusion in driver emotion recognition and contributes to the advancement of ADAS within automotive intelligent cockpits.

The full text can be downloaded at https://link.springer.com/article/10.1007/s40436-024-00519-8

Key words: Driver emotion, Artificial intelligence (AI), Facial expression, Driving behavior, Intelligent cockpit