Visual analytics using artificial intelligence (multi-modality driver action recognition)

A report published by the National Highway Traffic Safety Administration (NHTSA) in the United States showed that up to 3522 people were killed due to distracted driving. Various driver monitoring system were developed to tackle this issue and potentially saving lives and increasing road safety,...

Full description

Saved in:
Bibliographic Details
Main Author: Lee, Jaron Jin-An
Other Authors: Yap Kim Hui
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/176634
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-176634
record_format dspace
spelling sg-ntu-dr.10356-1766342024-05-24T15:50:25Z Visual analytics using artificial intelligence (multi-modality driver action recognition) Lee, Jaron Jin-An Yap Kim Hui School of Electrical and Electronic Engineering EKHYap@ntu.edu.sg Engineering A report published by the National Highway Traffic Safety Administration (NHTSA) in the United States showed that up to 3522 people were killed due to distracted driving. Various driver monitoring system were developed to tackle this issue and potentially saving lives and increasing road safety, one such system includes a driver video action recognition system. The project aims to develop a robust and stable driver action recognition model utilizing multimodality data streams, including RGB, IR and depth. A literature review was carried out to determine suitable model and dataset for this project. Following model and dataset selection, hyperparameters tuning is conducted to optimize VideoMAE V2 for improved accuracy and efficiency on the Drive&Act (DAA) dataset. Various fusion learning technique were explored and implemented into the system for evaluation. Early fusion achieves an average Top-1 accuracy of 82.40%, while late fusion obtains an average Top-1 accuracy of 84.30% on the test set. Overall, the project demonstrated the capability of incorporating early and late fusion methods with VideoMAE V2 model to achieve satisfactory results. This suggests the potential applicability of this model to different multi-modality action recognition tasks. Future work explores alternative fusion techniques and expanding the model to other driver datasets. Bachelor's degree 2024-05-19T23:15:40Z 2024-05-19T23:15:40Z 2024 Final Year Project (FYP) Lee, J. J. (2024). Visual analytics using artificial intelligence (multi-modality driver action recognition). Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/176634 https://hdl.handle.net/10356/176634 en A3256-231 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
spellingShingle Engineering
Lee, Jaron Jin-An
Visual analytics using artificial intelligence (multi-modality driver action recognition)
description A report published by the National Highway Traffic Safety Administration (NHTSA) in the United States showed that up to 3522 people were killed due to distracted driving. Various driver monitoring system were developed to tackle this issue and potentially saving lives and increasing road safety, one such system includes a driver video action recognition system. The project aims to develop a robust and stable driver action recognition model utilizing multimodality data streams, including RGB, IR and depth. A literature review was carried out to determine suitable model and dataset for this project. Following model and dataset selection, hyperparameters tuning is conducted to optimize VideoMAE V2 for improved accuracy and efficiency on the Drive&Act (DAA) dataset. Various fusion learning technique were explored and implemented into the system for evaluation. Early fusion achieves an average Top-1 accuracy of 82.40%, while late fusion obtains an average Top-1 accuracy of 84.30% on the test set. Overall, the project demonstrated the capability of incorporating early and late fusion methods with VideoMAE V2 model to achieve satisfactory results. This suggests the potential applicability of this model to different multi-modality action recognition tasks. Future work explores alternative fusion techniques and expanding the model to other driver datasets.
author2 Yap Kim Hui
author_facet Yap Kim Hui
Lee, Jaron Jin-An
format Final Year Project
author Lee, Jaron Jin-An
author_sort Lee, Jaron Jin-An
title Visual analytics using artificial intelligence (multi-modality driver action recognition)
title_short Visual analytics using artificial intelligence (multi-modality driver action recognition)
title_full Visual analytics using artificial intelligence (multi-modality driver action recognition)
title_fullStr Visual analytics using artificial intelligence (multi-modality driver action recognition)
title_full_unstemmed Visual analytics using artificial intelligence (multi-modality driver action recognition)
title_sort visual analytics using artificial intelligence (multi-modality driver action recognition)
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/176634
_version_ 1800916365259833344