Optimizing model training for speech recognition

Modern speech recognition systems are generally based on statistical models which output a sequence of symbols or quantities. These models can be trained automatically and are simple and computationally feasible to use. To reduce long computational time, the model training can be distributed to many...

Full description

Saved in:
Bibliographic Details
Main Author: Chak, Hui Ping
Other Authors: Lee Bu Sung
Format: Final Year Project
Language:English
Published: 2010
Subjects:
Online Access:http://hdl.handle.net/10356/40059
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-40059
record_format dspace
spelling sg-ntu-dr.10356-400592023-03-03T20:43:56Z Optimizing model training for speech recognition Chak, Hui Ping Lee Bu Sung School of Computer Engineering Parallel and Distributed Computing Centre DRNTU::Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Modern speech recognition systems are generally based on statistical models which output a sequence of symbols or quantities. These models can be trained automatically and are simple and computationally feasible to use. To reduce long computational time, the model training can be distributed to many machines for parallel processing. Apache Hadoop is a Java software framework that uses the Map-Reduce architecture to support data-intensive parallel and distributed processing. The objective of this project is to tune the performance of model training for speech recognition by distributing and parallelizing the model training process using the Hadoop framework. Performance of the optimization is measured for comparison and analysis. The report also shows how the legacy scripts are ported into the Map-Reduce architecture and discusses the issues and challenges involved. With the aid of the Swimlanes visualization tools [1] in understanding and tuning the performance of the job, various methods of processing data for the training are explored and discussed in the report. Performance is measured for 100 iterations of the model training process for 4 nodes using the various methods discussed. From the results of the experiment, it is found that model training can be optimized by taking data locality into consideration in the software design. Bachelor of Engineering (Computer Engineering) 2010-06-10T01:16:54Z 2010-06-10T01:16:54Z 2010 2010 Final Year Project (FYP) http://hdl.handle.net/10356/40059 en Nanyang Technological University 48 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering::Computing methodologies::Pattern recognition
spellingShingle DRNTU::Engineering::Computer science and engineering::Computing methodologies::Pattern recognition
Chak, Hui Ping
Optimizing model training for speech recognition
description Modern speech recognition systems are generally based on statistical models which output a sequence of symbols or quantities. These models can be trained automatically and are simple and computationally feasible to use. To reduce long computational time, the model training can be distributed to many machines for parallel processing. Apache Hadoop is a Java software framework that uses the Map-Reduce architecture to support data-intensive parallel and distributed processing. The objective of this project is to tune the performance of model training for speech recognition by distributing and parallelizing the model training process using the Hadoop framework. Performance of the optimization is measured for comparison and analysis. The report also shows how the legacy scripts are ported into the Map-Reduce architecture and discusses the issues and challenges involved. With the aid of the Swimlanes visualization tools [1] in understanding and tuning the performance of the job, various methods of processing data for the training are explored and discussed in the report. Performance is measured for 100 iterations of the model training process for 4 nodes using the various methods discussed. From the results of the experiment, it is found that model training can be optimized by taking data locality into consideration in the software design.
author2 Lee Bu Sung
author_facet Lee Bu Sung
Chak, Hui Ping
format Final Year Project
author Chak, Hui Ping
author_sort Chak, Hui Ping
title Optimizing model training for speech recognition
title_short Optimizing model training for speech recognition
title_full Optimizing model training for speech recognition
title_fullStr Optimizing model training for speech recognition
title_full_unstemmed Optimizing model training for speech recognition
title_sort optimizing model training for speech recognition
publishDate 2010
url http://hdl.handle.net/10356/40059
_version_ 1759856122180665344