Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification

The evolution of Intelligent Vehicles (IV) has enabled various degrees of autonomous driving, aiming to enhance road safety through Advanced Driver Assistance Systems (ADAS). Apart from road obstacle detection, the research on IV extends to driver-state monitoring, specifically on driver distract...

全面介紹

Saved in:
書目詳細資料
主要作者: Low, Daniel Teck Fatt
其他作者: Lyu Chen
格式: Final Year Project
語言:English
出版: Nanyang Technological University 2024
主題:
在線閱讀:https://hdl.handle.net/10356/177419
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
實物特徵
總結:The evolution of Intelligent Vehicles (IV) has enabled various degrees of autonomous driving, aiming to enhance road safety through Advanced Driver Assistance Systems (ADAS). Apart from road obstacle detection, the research on IV extends to driver-state monitoring, specifically on driver distraction to promote safe driving and minimise the likelihood of road accidents due to human error. Past studies focused on attaining high accuracy in driver activity recognition through deeper convolutional neural networks (CNN) with more parameters, which require more computational power, making them less viable for real-time classification. This report presents efficient CNN model architectures: MobileNetV3 and MobileVGG, designed for edge and mobile-like system, predominantly for driver activity recognition. Employing transfer learning approach, the models utilised parameters pretrained on large dataset for model training, enhancing data generalisation and model performance. The findings indicate that MobileNetV3 Large is the most effective for driver activity recognition. A dual-stream model, using MobileNetV3 Large as its backbone, has been developed to address occlusion and variations in camera angles by processing images from the driver’s front and side views. This model achieved 81% classification accuracy on real-world data with 10.9M parameters, about 50% less than the state-of-the-art models, and delivered 27 FPS in real-time.