Automatic video segmentation tool for laughter detection based on audio features
Non-linguistic signals, specifically, laughter offers a lot of information such as cues on the emotional state of a person a topic changes in meetings. The numerous benefits of laughter, ranging from identifying activities to improving speech-to-text accuracy, have gained the interests of many resea...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Animo Repository
2010
|
Subjects: | |
Online Access: | https://animorepository.dlsu.edu.ph/etd_bachelors/11194 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | De La Salle University |
Language: | English |
id |
oai:animorepository.dlsu.edu.ph:etd_bachelors-11839 |
---|---|
record_format |
eprints |
spelling |
oai:animorepository.dlsu.edu.ph:etd_bachelors-118392022-03-03T07:49:59Z Automatic video segmentation tool for laughter detection based on audio features Bantiling, Hans Paulo Gadi, Stefan Ron Lee, Jasper Charles Yang, John Vincent Non-linguistic signals, specifically, laughter offers a lot of information such as cues on the emotional state of a person a topic changes in meetings. The numerous benefits of laughter, ranging from identifying activities to improving speech-to-text accuracy, have gained the interests of many researchers. Most of the recent studies regarding laughter have successfully and effectively detected laughter using different modalities. However, they have encountered several problems in manually segmenting laughter as there is no automated segmentation tool for video containing laughter. This paper presents a prototype that would automatically segment laughter from videos of meetings. The prototype uses a segmentation algorithm pattered over the model-based segmentation approach. A SVM trained on audio features (such as MFCC, Pitch, and Formants) to classify instances. In addition, this research involves the creation of a Filipino Meeting corpus which consists of videos of spontaneous meetings. The classifier achieved an accuracy of over 90%. The segmentation algorithm achieved AUC-ROC of 0.79. The prototype is able to accurately segment laughter in videos however, errors are encountered (which is usually composed of onset and offsets of laughter and other sound accompanied with breathing). In addition, the research was also able to characterize the laughter models built and define laughter in the corpus. 2010-01-01T08:00:00Z text https://animorepository.dlsu.edu.ph/etd_bachelors/11194 Bachelor's Theses English Animo Repository Laughter Nonverbal communication Image processing--Digital techniques Computer Sciences |
institution |
De La Salle University |
building |
De La Salle University Library |
continent |
Asia |
country |
Philippines Philippines |
content_provider |
De La Salle University Library |
collection |
DLSU Institutional Repository |
language |
English |
topic |
Laughter Nonverbal communication Image processing--Digital techniques Computer Sciences |
spellingShingle |
Laughter Nonverbal communication Image processing--Digital techniques Computer Sciences Bantiling, Hans Paulo Gadi, Stefan Ron Lee, Jasper Charles Yang, John Vincent Automatic video segmentation tool for laughter detection based on audio features |
description |
Non-linguistic signals, specifically, laughter offers a lot of information such as cues on the emotional state of a person a topic changes in meetings. The numerous benefits of laughter, ranging from identifying activities to improving speech-to-text accuracy, have gained the interests of many researchers. Most of the recent studies regarding laughter have successfully and effectively detected laughter using different modalities. However, they have encountered several problems in manually segmenting laughter as there is no automated segmentation tool for video containing laughter. This paper presents a prototype that would automatically segment laughter from videos of meetings. The prototype uses a segmentation algorithm pattered over the model-based segmentation approach. A SVM trained on audio features (such as MFCC, Pitch, and Formants) to classify instances. In addition, this research involves the creation of a Filipino Meeting corpus which consists of videos of spontaneous meetings. The classifier achieved an accuracy of over 90%. The segmentation algorithm achieved AUC-ROC of 0.79. The prototype is able to accurately segment laughter in videos however, errors are encountered (which is usually composed of onset and offsets of laughter and other sound accompanied with breathing). In addition, the research was also able to characterize the laughter models built and define laughter in the corpus. |
format |
text |
author |
Bantiling, Hans Paulo Gadi, Stefan Ron Lee, Jasper Charles Yang, John Vincent |
author_facet |
Bantiling, Hans Paulo Gadi, Stefan Ron Lee, Jasper Charles Yang, John Vincent |
author_sort |
Bantiling, Hans Paulo |
title |
Automatic video segmentation tool for laughter detection based on audio features |
title_short |
Automatic video segmentation tool for laughter detection based on audio features |
title_full |
Automatic video segmentation tool for laughter detection based on audio features |
title_fullStr |
Automatic video segmentation tool for laughter detection based on audio features |
title_full_unstemmed |
Automatic video segmentation tool for laughter detection based on audio features |
title_sort |
automatic video segmentation tool for laughter detection based on audio features |
publisher |
Animo Repository |
publishDate |
2010 |
url |
https://animorepository.dlsu.edu.ph/etd_bachelors/11194 |
_version_ |
1728621048938627072 |