The augmented human — seeing sounds
“Dinner table syndrome” describes the difficulties faced by those with hearing impairments in engaging in conversations with other speaking participants, which may lead to fatigue and social exclusion. This project designed and implemented an application which provides accessibility to hearing-impai...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/175150 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-175150 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1751502024-04-26T15:41:08Z The augmented human — seeing sounds Lim, Nicole Sze Ting Cham Tat Jen School of Computer Science and Engineering ASTJCham@ntu.edu.sg Computer and Information Science Assistive technology Speech separation Speech recognition Hearing impariment Software engineering Web application “Dinner table syndrome” describes the difficulties faced by those with hearing impairments in engaging in conversations with other speaking participants, which may lead to fatigue and social exclusion. This project designed and implemented an application which provides accessibility to hearing-impaired users in multi-speaker conversations, through automated captioning and speaker identification. Through a user-friendly web interface, a user uploads a video and adjusts settings for its processing. Speech separation and speech recognition models are then applied to produce an annotated video with captions differentiated by speaker. Four video annotation interfaces were proposed, of which floating captions tagged to each speaker’s face were preferred. The user is then able to monitor the processing of the video, and download the annotated video and a transcript with captions divided by speaker identity when processing is complete. Design choices made in the implementation of the final application are evaluated. Bachelor's degree 2024-04-22T05:35:24Z 2024-04-22T05:35:24Z 2024 Final Year Project (FYP) Lim, N. S. T. (2024). The augmented human — seeing sounds. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/175150 https://hdl.handle.net/10356/175150 en SCSE23-0038 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Assistive technology Speech separation Speech recognition Hearing impariment Software engineering Web application |
spellingShingle |
Computer and Information Science Assistive technology Speech separation Speech recognition Hearing impariment Software engineering Web application Lim, Nicole Sze Ting The augmented human — seeing sounds |
description |
“Dinner table syndrome” describes the difficulties faced by those with hearing impairments in engaging in conversations with other speaking participants, which may lead to fatigue and social exclusion. This project designed and implemented an application which provides accessibility to hearing-impaired users in multi-speaker conversations, through automated captioning and speaker identification. Through a user-friendly web interface, a user uploads a video and adjusts settings for its processing. Speech separation and speech recognition models are then applied to produce an annotated video with captions differentiated by speaker. Four video annotation interfaces were proposed, of which floating captions tagged to each speaker’s face were preferred. The user is then able to monitor the processing of the video, and download the annotated video and a transcript with captions divided by speaker identity when processing is complete. Design choices made in the implementation of the final application are evaluated. |
author2 |
Cham Tat Jen |
author_facet |
Cham Tat Jen Lim, Nicole Sze Ting |
format |
Final Year Project |
author |
Lim, Nicole Sze Ting |
author_sort |
Lim, Nicole Sze Ting |
title |
The augmented human — seeing sounds |
title_short |
The augmented human — seeing sounds |
title_full |
The augmented human — seeing sounds |
title_fullStr |
The augmented human — seeing sounds |
title_full_unstemmed |
The augmented human — seeing sounds |
title_sort |
augmented human — seeing sounds |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/175150 |
_version_ |
1814047432157167616 |