Comparison between AI local execution and cloud offloading for AIoT

The rapid evolution of Artificial Intelligence (AI) and the Internet of Things (IoT) has led to the development of AIoT (Artificial Intelligence of Things), where AI empowers IoT devices with intelligent processing and autonomous decision-making capabilities. While AIoT systems drive innovatio...

Full description

Saved in:
Bibliographic Details
Main Author: Quek, Wei Quan
Other Authors: Tan Rui
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/181127
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-181127
record_format dspace
spelling sg-ntu-dr.10356-1811272024-11-15T12:03:38Z Comparison between AI local execution and cloud offloading for AIoT Quek, Wei Quan Tan Rui College of Computing and Data Science tanrui@ntu.edu.sg Computer and Information Science The rapid evolution of Artificial Intelligence (AI) and the Internet of Things (IoT) has led to the development of AIoT (Artificial Intelligence of Things), where AI empowers IoT devices with intelligent processing and autonomous decision-making capabilities. While AIoT systems drive innovation across industries like healthcare, smart homes, and autonomous vehicles, they face significant challenges related to computational resources, energy efficiency, and latency. To address these limitations, AI offloading to the cloud has been a common solution, although it introduces concerns like network latency, data privacy risks, and energy consumption. This research investigates whether running a reduced AI model locally on resource constrained IoT devices can achieve performance levels comparable to cloud-based AI offloading. Specifically, the study aims to determine the minimum memory threshold required for local execution to match the speed and accuracy of cloud inference. Machine learning models such as MobileNetV2, EfficientNetV2B2, and DenseNet121 were employed for image classification across different datasets, using various quantization techniques, including dynamic range, full integer, and INT16 quantization. The results show that local inference with smaller models, like MobileNet, generally outperforms cloud inference in terms of latency with minor differences in accuracy. For larger models, like DenseNet, compressed versions for local and cloud inference achieved similar speeds and accuracy, though the uncompressed base model performed better in the cloud. The study also highlights the varying effects of compression techniques, with Dynamic Range Quantization offering the smallest model size but impacting latency significantly on MobileNet. In conclusion, as model size and complexity grow, cloud inference may become more advantageous, and thorough validation is essential before deploying compressed models in real-world applications. Bachelor's degree 2024-11-15T12:03:38Z 2024-11-15T12:03:38Z 2024 Final Year Project (FYP) Quek, W. Q. (2024). Comparison between AI local execution and cloud offloading for AIoT. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181127 https://hdl.handle.net/10356/181127 en SCSE23-0791 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
spellingShingle Computer and Information Science
Quek, Wei Quan
Comparison between AI local execution and cloud offloading for AIoT
description The rapid evolution of Artificial Intelligence (AI) and the Internet of Things (IoT) has led to the development of AIoT (Artificial Intelligence of Things), where AI empowers IoT devices with intelligent processing and autonomous decision-making capabilities. While AIoT systems drive innovation across industries like healthcare, smart homes, and autonomous vehicles, they face significant challenges related to computational resources, energy efficiency, and latency. To address these limitations, AI offloading to the cloud has been a common solution, although it introduces concerns like network latency, data privacy risks, and energy consumption. This research investigates whether running a reduced AI model locally on resource constrained IoT devices can achieve performance levels comparable to cloud-based AI offloading. Specifically, the study aims to determine the minimum memory threshold required for local execution to match the speed and accuracy of cloud inference. Machine learning models such as MobileNetV2, EfficientNetV2B2, and DenseNet121 were employed for image classification across different datasets, using various quantization techniques, including dynamic range, full integer, and INT16 quantization. The results show that local inference with smaller models, like MobileNet, generally outperforms cloud inference in terms of latency with minor differences in accuracy. For larger models, like DenseNet, compressed versions for local and cloud inference achieved similar speeds and accuracy, though the uncompressed base model performed better in the cloud. The study also highlights the varying effects of compression techniques, with Dynamic Range Quantization offering the smallest model size but impacting latency significantly on MobileNet. In conclusion, as model size and complexity grow, cloud inference may become more advantageous, and thorough validation is essential before deploying compressed models in real-world applications.
author2 Tan Rui
author_facet Tan Rui
Quek, Wei Quan
format Final Year Project
author Quek, Wei Quan
author_sort Quek, Wei Quan
title Comparison between AI local execution and cloud offloading for AIoT
title_short Comparison between AI local execution and cloud offloading for AIoT
title_full Comparison between AI local execution and cloud offloading for AIoT
title_fullStr Comparison between AI local execution and cloud offloading for AIoT
title_full_unstemmed Comparison between AI local execution and cloud offloading for AIoT
title_sort comparison between ai local execution and cloud offloading for aiot
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/181127
_version_ 1816858991739797504