Pedestrian attribute recognition

Pedestrian Attribute Recognition aims to develop a machine learning model to recognize person attributes including age, gender, clothing and accessories in a given person image. It is a challenging task due to high quality variance and inconsistent number of attributes in input image samples. Many m...

Full description

Saved in:
Bibliographic Details
Main Author: Liu, Yuan
Other Authors: Lin Zhiping
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/139139
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-139139
record_format dspace
spelling sg-ntu-dr.10356-1391392023-07-07T18:52:03Z Pedestrian attribute recognition Liu, Yuan Lin Zhiping School of Electrical and Electronic Engineering ezplin@ntu.edu.sg Engineering::Electrical and electronic engineering::Computer hardware, software and systems Engineering::Computer science and engineering::Computer applications::Computer-aided engineering Pedestrian Attribute Recognition aims to develop a machine learning model to recognize person attributes including age, gender, clothing and accessories in a given person image. It is a challenging task due to high quality variance and inconsistent number of attributes in input image samples. Many models have been proposed to tackle this problem. However, there are two open problems that remain to be addressed. On the one hand, visually similar pedestrian images tend to share similar attributes, and the upper and lower attributes normally match in style. However, the existing methods normally either ignore these relations or capture them ineffectively. On the other hand, how to maximise the benefit from each input pedestrian sample based on their importance to the machine learning model has not been explored, which can be critical given the huge variety in training images. In this report, we propose two deep learning based neural network models to tackle the two problems separately. The “Pentadent-Net” (PD-Net) is introduced with two parallel branches, where the two branches capture the relations among different samples and different attributes with a graphical method. The “Reinforced Sample Re-weighting” (RSR) is also proposed to re-weight samples in a batch during back-propagation through reinforcement learning, which can facilitate model training based on each input sample’s importance. The proposed PD-Net and RSR achieve state-of-the-art performance against existing methods on two large scale pedestrian attribute benchmarks PETA and PA-100K, which validates their effectives and advantages. The research paper generated from RSR has been accepted by WCCI 2020. The research paper on PD-Net is pending review for ICIP 2020. Notably, this project has been shortlisted as one of the 8 finalists for the EEE FYP Challenge 2020. Bachelor of Engineering (Electrical and Electronic Engineering) 2020-05-16T05:17:43Z 2020-05-16T05:17:43Z 2020 Final Year Project (FYP) https://hdl.handle.net/10356/139139 en B3132-191 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering::Computer hardware, software and systems
Engineering::Computer science and engineering::Computer applications::Computer-aided engineering
spellingShingle Engineering::Electrical and electronic engineering::Computer hardware, software and systems
Engineering::Computer science and engineering::Computer applications::Computer-aided engineering
Liu, Yuan
Pedestrian attribute recognition
description Pedestrian Attribute Recognition aims to develop a machine learning model to recognize person attributes including age, gender, clothing and accessories in a given person image. It is a challenging task due to high quality variance and inconsistent number of attributes in input image samples. Many models have been proposed to tackle this problem. However, there are two open problems that remain to be addressed. On the one hand, visually similar pedestrian images tend to share similar attributes, and the upper and lower attributes normally match in style. However, the existing methods normally either ignore these relations or capture them ineffectively. On the other hand, how to maximise the benefit from each input pedestrian sample based on their importance to the machine learning model has not been explored, which can be critical given the huge variety in training images. In this report, we propose two deep learning based neural network models to tackle the two problems separately. The “Pentadent-Net” (PD-Net) is introduced with two parallel branches, where the two branches capture the relations among different samples and different attributes with a graphical method. The “Reinforced Sample Re-weighting” (RSR) is also proposed to re-weight samples in a batch during back-propagation through reinforcement learning, which can facilitate model training based on each input sample’s importance. The proposed PD-Net and RSR achieve state-of-the-art performance against existing methods on two large scale pedestrian attribute benchmarks PETA and PA-100K, which validates their effectives and advantages. The research paper generated from RSR has been accepted by WCCI 2020. The research paper on PD-Net is pending review for ICIP 2020. Notably, this project has been shortlisted as one of the 8 finalists for the EEE FYP Challenge 2020.
author2 Lin Zhiping
author_facet Lin Zhiping
Liu, Yuan
format Final Year Project
author Liu, Yuan
author_sort Liu, Yuan
title Pedestrian attribute recognition
title_short Pedestrian attribute recognition
title_full Pedestrian attribute recognition
title_fullStr Pedestrian attribute recognition
title_full_unstemmed Pedestrian attribute recognition
title_sort pedestrian attribute recognition
publisher Nanyang Technological University
publishDate 2020
url https://hdl.handle.net/10356/139139
_version_ 1772825649116872704