Neural network implementation on a graphics processing unit using CUDA

The Graphics Processing Unit (GPU) parallel architecture is now being used not just for graphics but also for general purpose computations. This emerging field of general-purpose computation on graphics hardware is referred to as General Purpose-computing on a Graphics Processing Unit (GPGPU). Neura...

Full description

Saved in:
Bibliographic Details
Main Authors: Barcelona, Ma. Francesca, Bautista, Byron Joshua, Candano, Shaun Raphael, Tadios, Marie Katherine
Format: text
Language:English
Published: Animo Repository 2010
Subjects:
Online Access:https://animorepository.dlsu.edu.ph/etd_bachelors/11504
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: De La Salle University
Language: English
id oai:animorepository.dlsu.edu.ph:etd_bachelors-12149
record_format eprints
spelling oai:animorepository.dlsu.edu.ph:etd_bachelors-121492022-03-11T07:40:43Z Neural network implementation on a graphics processing unit using CUDA Barcelona, Ma. Francesca Bautista, Byron Joshua Candano, Shaun Raphael Tadios, Marie Katherine The Graphics Processing Unit (GPU) parallel architecture is now being used not just for graphics but also for general purpose computations. This emerging field of general-purpose computation on graphics hardware is referred to as General Purpose-computing on a Graphics Processing Unit (GPGPU). Neural networks are highly parallel algorithms which may be implemented in the GPU. Previous research works used shading languages that are used mainly for graphics computations. NVIDIA, a leading GPU vendor developed a technology called Compute Unified Device Architecture (CUDA) which is appropriate for GPGPU programming. In this research, the proponents have developed a GPU-based implementation of Kohonen's Self-Organizing Map using CUDA. The proponents used Animal SOM data and Music Classification data to test the network and used the GPUs, NVIDIA GeForce 9400M and NVIDIA GeFore 9800GT for the testing. The minimum speedup using the Animal SOM data on the NVIDIA GeForce 9400M was 1.35 using 64 x 64 network and maximum speedup of 3.11 using 256x256 network. Using the Music Classification data, the minimum speedup on the NVIDIA GeForce 9400m was 1.17 using 64x64 network and maximum speed of 1.64 using 512x512 network. The minimum speedup on the NVIDIA GeForce 9800GT was 1.32 using 32x32 network and maximum speedup of 2.72 using 256x256 network. 2010-01-01T08:00:00Z text https://animorepository.dlsu.edu.ph/etd_bachelors/11504 Bachelor's Theses English Animo Repository Neural networks (Computer science) Computer Sciences
institution De La Salle University
building De La Salle University Library
continent Asia
country Philippines
Philippines
content_provider De La Salle University Library
collection DLSU Institutional Repository
language English
topic Neural networks (Computer science)
Computer Sciences
spellingShingle Neural networks (Computer science)
Computer Sciences
Barcelona, Ma. Francesca
Bautista, Byron Joshua
Candano, Shaun Raphael
Tadios, Marie Katherine
Neural network implementation on a graphics processing unit using CUDA
description The Graphics Processing Unit (GPU) parallel architecture is now being used not just for graphics but also for general purpose computations. This emerging field of general-purpose computation on graphics hardware is referred to as General Purpose-computing on a Graphics Processing Unit (GPGPU). Neural networks are highly parallel algorithms which may be implemented in the GPU. Previous research works used shading languages that are used mainly for graphics computations. NVIDIA, a leading GPU vendor developed a technology called Compute Unified Device Architecture (CUDA) which is appropriate for GPGPU programming. In this research, the proponents have developed a GPU-based implementation of Kohonen's Self-Organizing Map using CUDA. The proponents used Animal SOM data and Music Classification data to test the network and used the GPUs, NVIDIA GeForce 9400M and NVIDIA GeFore 9800GT for the testing. The minimum speedup using the Animal SOM data on the NVIDIA GeForce 9400M was 1.35 using 64 x 64 network and maximum speedup of 3.11 using 256x256 network. Using the Music Classification data, the minimum speedup on the NVIDIA GeForce 9400m was 1.17 using 64x64 network and maximum speed of 1.64 using 512x512 network. The minimum speedup on the NVIDIA GeForce 9800GT was 1.32 using 32x32 network and maximum speedup of 2.72 using 256x256 network.
format text
author Barcelona, Ma. Francesca
Bautista, Byron Joshua
Candano, Shaun Raphael
Tadios, Marie Katherine
author_facet Barcelona, Ma. Francesca
Bautista, Byron Joshua
Candano, Shaun Raphael
Tadios, Marie Katherine
author_sort Barcelona, Ma. Francesca
title Neural network implementation on a graphics processing unit using CUDA
title_short Neural network implementation on a graphics processing unit using CUDA
title_full Neural network implementation on a graphics processing unit using CUDA
title_fullStr Neural network implementation on a graphics processing unit using CUDA
title_full_unstemmed Neural network implementation on a graphics processing unit using CUDA
title_sort neural network implementation on a graphics processing unit using cuda
publisher Animo Repository
publishDate 2010
url https://animorepository.dlsu.edu.ph/etd_bachelors/11504
_version_ 1728621121678344192