Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification
The growing interest in Intelligent Vehicles (IV) worldwide have made it possible to operate vehicles with varying degrees of autonomy. Driver Activity Recognition (DAR) is an important area of research which aims to reduce road accidents caused by distracted driving. Past research conducted have a...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/167301 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-167301 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1673012023-07-06T08:45:18Z Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification Lim, Cai Yin Lyu Chen School of Mechanical and Aerospace Engineering lyuchen@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence The growing interest in Intelligent Vehicles (IV) worldwide have made it possible to operate vehicles with varying degrees of autonomy. Driver Activity Recognition (DAR) is an important area of research which aims to reduce road accidents caused by distracted driving. Past research conducted have a strong emphasis on using datasets in optimal condition to achieve high accuracy. The lack of variations in data for the models to train may not perform as well in real time as it has been optimised. Therefore, this report aims to develop a deep learning model which has the lowest computational cost while maintaining a high accuracy. In this research, Convolutional Neural Network (CNN), transfer learning and two-stream neural network will be investigated. The dataset used to train models are Kaggle for single-frame and Ben Khalifa’s for multi-frame models. 2D CNN and 2D CNN with transfer learning was applied on single frame dataset while 3D CNN with transfer learning, Two-Stream transfer learning and Two-Stream CNN was used on multi-frame dataset. 2D CNN with transfer learning, ResNet18, performs the best for single-frame dataset with a high accuracy of 99.44%. Two-Stream transfer learning with RGB and optical flow was able to attain an accuracy of 97.16% for multi-frame dataset. The best performing model for each type was then compared before selecting the final model for testing with real-world data. Bachelor of Engineering (Mechanical Engineering) 2023-05-25T08:54:35Z 2023-05-25T08:54:35Z 2023 Final Year Project (FYP) Lim, C. Y. (2023). Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/167301 https://hdl.handle.net/10356/167301 en C073 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Lim, Cai Yin Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification |
description |
The growing interest in Intelligent Vehicles (IV) worldwide have made it possible to operate vehicles with varying degrees of autonomy. Driver Activity Recognition (DAR) is an important area of research which aims to reduce road accidents caused by distracted driving. Past research conducted have a strong emphasis on using datasets in optimal condition to achieve high accuracy. The lack of variations in data for the models to train may not perform as well in real time as it has been optimised.
Therefore, this report aims to develop a deep learning model which has the lowest computational cost while maintaining a high accuracy. In this research, Convolutional Neural Network (CNN), transfer learning and two-stream neural network will be investigated. The dataset used to train models are Kaggle for single-frame and Ben Khalifa’s for multi-frame models. 2D CNN and 2D CNN with transfer learning was applied on single frame dataset while 3D CNN with transfer learning, Two-Stream transfer learning and Two-Stream CNN was used on multi-frame dataset.
2D CNN with transfer learning, ResNet18, performs the best for single-frame dataset with a high accuracy of 99.44%. Two-Stream transfer learning with RGB and optical flow was able to attain an accuracy of 97.16% for multi-frame dataset. The best performing model for each type was then compared before selecting the final model for testing with real-world data. |
author2 |
Lyu Chen |
author_facet |
Lyu Chen Lim, Cai Yin |
format |
Final Year Project |
author |
Lim, Cai Yin |
author_sort |
Lim, Cai Yin |
title |
Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification |
title_short |
Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification |
title_full |
Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification |
title_fullStr |
Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification |
title_full_unstemmed |
Driver state monitoring for intelligent vehicles - part I: in-cabin activity identification |
title_sort |
driver state monitoring for intelligent vehicles - part i: in-cabin activity identification |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/167301 |
_version_ |
1772826360019943424 |