Challenges in action recognition in videos

With the discovery of action recognition and object detection in computer vision, there has been an increase in research interest among this field of study as many different methodologies and techniques are being explored, experimented, analyzed and proposed to improve the efficiency and robustness...

Full description

Saved in:
Bibliographic Details
Main Author: Lee, Liang Cheng
Other Authors: Teoh Eam Khwang
Format: Final Year Project
Language:English
Published: 2016
Subjects:
Online Access:http://hdl.handle.net/10356/67734
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-67734
record_format dspace
spelling sg-ntu-dr.10356-677342023-07-07T17:07:33Z Challenges in action recognition in videos Lee, Liang Cheng Teoh Eam Khwang School of Electrical and Electronic Engineering DRNTU::Engineering DRNTU::Engineering With the discovery of action recognition and object detection in computer vision, there has been an increase in research interest among this field of study as many different methodologies and techniques are being explored, experimented, analyzed and proposed to improve the efficiency and robustness of the existing ability of the program to recognize or detect an object. However, there are certain challenges that will always pose as an obstacle which will affect the accuracy of the program. Several well-known challenges such as background cluttering, view point, camera motion and occlusions often posed as an obstacle in action recognition. Due to the complexity in nature of these challenges especially occlusion, which makes detection in video difficult, much attention and research focus is required to address all this challenges. To depict the challenges faced in action recognition in videos, a relatively new yet efficient technique known as dense trajectory will be adopted while utilizing several different types of dataset to determine the output accuracy, so as to determine and study the effects that these challenges could cause. In this report, the dense trajectory method will compute the features using dense sampling and optical flow field to extract dense trajectory followed by concatenation of computed features by 4 descriptors to determine its robustness in handling challenges of action recognition. On top of basic action recognition testing, different types of test known as view base testing and occlusion effect testing was being carried out as well. These challenges of action recognition was then identified and analyze based on the eventual accuracy attain. The next part is to adopt the second method known as log-covariance matrix method with motion context to serve as a comparative study with dense trajectory method in terms of robustness in handling these challenges. Based on the eventual accuracy attained from these two methods, it can be shown that dense trajectory method outperforms log-covariance matrix with motion context method. Bachelor of Engineering 2016-05-19T08:36:37Z 2016-05-19T08:36:37Z 2016 Final Year Project (FYP) http://hdl.handle.net/10356/67734 en Nanyang Technological University 170 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering
DRNTU::Engineering
spellingShingle DRNTU::Engineering
DRNTU::Engineering
Lee, Liang Cheng
Challenges in action recognition in videos
description With the discovery of action recognition and object detection in computer vision, there has been an increase in research interest among this field of study as many different methodologies and techniques are being explored, experimented, analyzed and proposed to improve the efficiency and robustness of the existing ability of the program to recognize or detect an object. However, there are certain challenges that will always pose as an obstacle which will affect the accuracy of the program. Several well-known challenges such as background cluttering, view point, camera motion and occlusions often posed as an obstacle in action recognition. Due to the complexity in nature of these challenges especially occlusion, which makes detection in video difficult, much attention and research focus is required to address all this challenges. To depict the challenges faced in action recognition in videos, a relatively new yet efficient technique known as dense trajectory will be adopted while utilizing several different types of dataset to determine the output accuracy, so as to determine and study the effects that these challenges could cause. In this report, the dense trajectory method will compute the features using dense sampling and optical flow field to extract dense trajectory followed by concatenation of computed features by 4 descriptors to determine its robustness in handling challenges of action recognition. On top of basic action recognition testing, different types of test known as view base testing and occlusion effect testing was being carried out as well. These challenges of action recognition was then identified and analyze based on the eventual accuracy attain. The next part is to adopt the second method known as log-covariance matrix method with motion context to serve as a comparative study with dense trajectory method in terms of robustness in handling these challenges. Based on the eventual accuracy attained from these two methods, it can be shown that dense trajectory method outperforms log-covariance matrix with motion context method.
author2 Teoh Eam Khwang
author_facet Teoh Eam Khwang
Lee, Liang Cheng
format Final Year Project
author Lee, Liang Cheng
author_sort Lee, Liang Cheng
title Challenges in action recognition in videos
title_short Challenges in action recognition in videos
title_full Challenges in action recognition in videos
title_fullStr Challenges in action recognition in videos
title_full_unstemmed Challenges in action recognition in videos
title_sort challenges in action recognition in videos
publishDate 2016
url http://hdl.handle.net/10356/67734
_version_ 1772826696887566336