Sound-event classification for robot hearing

Throughout the years, there have been several methods of executing the process of sound-event classification. The use of spectrograms and a time-frequency spectral analysis that illustrates the magnitude spectrum of the signal within a 2D time-frequency plane are some examples of the well known meth...

Full description

Saved in:
Bibliographic Details
Main Author: Yong, Shu Ching
Other Authors: Jiang Xudong
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2022
Subjects:
Online Access:https://hdl.handle.net/10356/157932
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-157932
record_format dspace
spelling sg-ntu-dr.10356-1579322023-07-07T19:20:23Z Sound-event classification for robot hearing Yong, Shu Ching Jiang Xudong School of Electrical and Electronic Engineering EXDJiang@ntu.edu.sg Engineering::Electrical and electronic engineering Throughout the years, there have been several methods of executing the process of sound-event classification. The use of spectrograms and a time-frequency spectral analysis that illustrates the magnitude spectrum of the signal within a 2D time-frequency plane are some examples of the well known methods. Even though intensive research was done, there are still greater developments that can be achieved. For instance, for sound-based recognition, there still exists a research gap to enhance its accuracy and reliability. By using a spectrogram, audio signals can be visualised and evaluated into a time-frequency spectral analysis of a magnitude spectrum on a 2D plane. However, the magnitude spectrum is not enough to classify the audio sources. To address this issue, a method, first proposed by Jiang Xudong and Ren Jianfeng, called “Regularised 2D complex-log-Fourier transform” is introduced. The addition to this process is a phase spectrum which will also be used to do sound-event classification. On top of this, the Principal Component Analysis (PCA) is used to extract out significant information and remove unnecessary data in the audio samples. Last but not least, the calculated values using the Mahalanobis Distance will be used to identify the belonging classes of the sound events. Bachelor of Engineering (Electrical and Electronic Engineering) 2022-05-25T03:23:40Z 2022-05-25T03:23:40Z 2022 Final Year Project (FYP) Yong, S. C. (2022). Sound-event classification for robot hearing. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/157932 https://hdl.handle.net/10356/157932 en A3090-211 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
spellingShingle Engineering::Electrical and electronic engineering
Yong, Shu Ching
Sound-event classification for robot hearing
description Throughout the years, there have been several methods of executing the process of sound-event classification. The use of spectrograms and a time-frequency spectral analysis that illustrates the magnitude spectrum of the signal within a 2D time-frequency plane are some examples of the well known methods. Even though intensive research was done, there are still greater developments that can be achieved. For instance, for sound-based recognition, there still exists a research gap to enhance its accuracy and reliability. By using a spectrogram, audio signals can be visualised and evaluated into a time-frequency spectral analysis of a magnitude spectrum on a 2D plane. However, the magnitude spectrum is not enough to classify the audio sources. To address this issue, a method, first proposed by Jiang Xudong and Ren Jianfeng, called “Regularised 2D complex-log-Fourier transform” is introduced. The addition to this process is a phase spectrum which will also be used to do sound-event classification. On top of this, the Principal Component Analysis (PCA) is used to extract out significant information and remove unnecessary data in the audio samples. Last but not least, the calculated values using the Mahalanobis Distance will be used to identify the belonging classes of the sound events.
author2 Jiang Xudong
author_facet Jiang Xudong
Yong, Shu Ching
format Final Year Project
author Yong, Shu Ching
author_sort Yong, Shu Ching
title Sound-event classification for robot hearing
title_short Sound-event classification for robot hearing
title_full Sound-event classification for robot hearing
title_fullStr Sound-event classification for robot hearing
title_full_unstemmed Sound-event classification for robot hearing
title_sort sound-event classification for robot hearing
publisher Nanyang Technological University
publishDate 2022
url https://hdl.handle.net/10356/157932
_version_ 1772827922891014144