Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU

This paper explores the utilization of hardware acceleration techniques for the non-linear layers in Transformer networks, specifically within the context of RISC-V CPU archi- tectures. The growing complexity of Transformer-based models, highlighted by their significant computational demands, unders...

全面介紹

Saved in:
書目詳細資料
主要作者: Seenivasagan Haresh
其他作者: Chang Chip Hong
格式: Final Year Project
語言:English
出版: Nanyang Technological University 2024
主題:
在線閱讀:https://hdl.handle.net/10356/177093
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
id sg-ntu-dr.10356-177093
record_format dspace
spelling sg-ntu-dr.10356-1770932024-05-31T15:44:43Z Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU Seenivasagan Haresh Chang Chip Hong School of Electrical and Electronic Engineering ECHChang@ntu.edu.sg Engineering Hardware acceleration RISC-V Transformers This paper explores the utilization of hardware acceleration techniques for the non-linear layers in Transformer networks, specifically within the context of RISC-V CPU archi- tectures. The growing complexity of Transformer-based models, highlighted by their significant computational demands, underscores the need for optimized computing solu- tions. Despite the widespread application of these models in generating human-like text and other multi-modal AI tasks, their deployment is often hampered by the high volume of Floating Point Operations (FLOPs) required, particularly for activation functions like GELU, Softmax, and SiLU. RISC-V, an open Instruction Set Architecture (ISA), offers a promising avenue for addressing these challenges due to its customizable and royalty-free nature. This paper investigates the potential of RISC-V CPUs to provide efficient hard- ware acceleration for the computationally intensive layers of Transformer networks. By focusing on non-linear layers, we aim to enhance the overall execution speed and energy efficiency of these models . Bachelor's degree 2024-05-27T02:39:17Z 2024-05-27T02:39:17Z 2024 Final Year Project (FYP) Seenivasagan Haresh (2024). Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177093 https://hdl.handle.net/10356/177093 en A2043-231 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Hardware acceleration
RISC-V
Transformers
spellingShingle Engineering
Hardware acceleration
RISC-V
Transformers
Seenivasagan Haresh
Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU
description This paper explores the utilization of hardware acceleration techniques for the non-linear layers in Transformer networks, specifically within the context of RISC-V CPU archi- tectures. The growing complexity of Transformer-based models, highlighted by their significant computational demands, underscores the need for optimized computing solu- tions. Despite the widespread application of these models in generating human-like text and other multi-modal AI tasks, their deployment is often hampered by the high volume of Floating Point Operations (FLOPs) required, particularly for activation functions like GELU, Softmax, and SiLU. RISC-V, an open Instruction Set Architecture (ISA), offers a promising avenue for addressing these challenges due to its customizable and royalty-free nature. This paper investigates the potential of RISC-V CPUs to provide efficient hard- ware acceleration for the computationally intensive layers of Transformer networks. By focusing on non-linear layers, we aim to enhance the overall execution speed and energy efficiency of these models .
author2 Chang Chip Hong
author_facet Chang Chip Hong
Seenivasagan Haresh
format Final Year Project
author Seenivasagan Haresh
author_sort Seenivasagan Haresh
title Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU
title_short Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU
title_full Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU
title_fullStr Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU
title_full_unstemmed Hardware acceleration for non-linear layers of transformer networks on RISC-V CPU
title_sort hardware acceleration for non-linear layers of transformer networks on risc-v cpu
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/177093
_version_ 1814047374256898048