DiTMoS: Delving into diverse tiny-model selection on microcontrollers

Enabling efficient and accurate deep neural network (DNN) inference on microcontrollers is non-trivial due to the constrained on-chip resources. Current methodologies primarily focus on compressing larger models yet at the expense of model accuracy. In this paper, we rethink the problem from the inv...

Full description

Saved in:
Bibliographic Details
Main Authors: MA, Xiao, HE, Shengfeng, QIAO, Hezhe, MA, Dong
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8670
https://ink.library.smu.edu.sg/context/sis_research/article/9673/viewcontent/2403.09035.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9673
record_format dspace
spelling sg-smu-ink.sis_research-96732024-03-27T02:40:58Z DiTMoS: Delving into diverse tiny-model selection on microcontrollers MA, Xiao HE, Shengfeng QIAO, Hezhe MA, Dong Enabling efficient and accurate deep neural network (DNN) inference on microcontrollers is non-trivial due to the constrained on-chip resources. Current methodologies primarily focus on compressing larger models yet at the expense of model accuracy. In this paper, we rethink the problem from the inverse perspective by constructing small/weak models directly and improving their accuracy. Thus, we introduce DiTMoS, a novel DNN training and inference framework with a selectorclassifiers architecture, where the selector routes each input sample to the appropriate classifier for classification. DiTMoS is grounded on a key insight: a composition of weak models can exhibit high diversity and the union of them can significantly boost the accuracy upper bound. To approach the upper bound, DiTMoS introduces three strategies including diverse training data splitting to increase the classifiers’ diversity, adversarial selectorclassifiers training to ensure synergistic interactions thereby maximizing their complementarity, and heterogeneous feature aggregation to improve the capacity of classifiers. We further propose a network slicing technique to alleviate the extra memory overhead incurred by feature aggregation. We deploy DiTMoS on the Neucleo STM32F767ZI board and evaluate it based on three time-series datasets for human activity recognition, keywords spotting, and emotion recognition, respectively. The experiment results manifest that: (a) DiTMoS achieves up to 13.4% accuracy improvement compared to the best baseline; (b) network slicing almost completely eliminates the memory overhead incurred by feature aggregation with a marginal increase of latency. Code is released at https://github.com/TheMaXiao/DiTMoS 2024-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8670 https://ink.library.smu.edu.sg/context/sis_research/article/9673/viewcontent/2403.09035.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University embedded machine learning model diversity model selection adversarial training Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic embedded machine learning
model diversity
model selection
adversarial training
Software Engineering
spellingShingle embedded machine learning
model diversity
model selection
adversarial training
Software Engineering
MA, Xiao
HE, Shengfeng
QIAO, Hezhe
MA, Dong
DiTMoS: Delving into diverse tiny-model selection on microcontrollers
description Enabling efficient and accurate deep neural network (DNN) inference on microcontrollers is non-trivial due to the constrained on-chip resources. Current methodologies primarily focus on compressing larger models yet at the expense of model accuracy. In this paper, we rethink the problem from the inverse perspective by constructing small/weak models directly and improving their accuracy. Thus, we introduce DiTMoS, a novel DNN training and inference framework with a selectorclassifiers architecture, where the selector routes each input sample to the appropriate classifier for classification. DiTMoS is grounded on a key insight: a composition of weak models can exhibit high diversity and the union of them can significantly boost the accuracy upper bound. To approach the upper bound, DiTMoS introduces three strategies including diverse training data splitting to increase the classifiers’ diversity, adversarial selectorclassifiers training to ensure synergistic interactions thereby maximizing their complementarity, and heterogeneous feature aggregation to improve the capacity of classifiers. We further propose a network slicing technique to alleviate the extra memory overhead incurred by feature aggregation. We deploy DiTMoS on the Neucleo STM32F767ZI board and evaluate it based on three time-series datasets for human activity recognition, keywords spotting, and emotion recognition, respectively. The experiment results manifest that: (a) DiTMoS achieves up to 13.4% accuracy improvement compared to the best baseline; (b) network slicing almost completely eliminates the memory overhead incurred by feature aggregation with a marginal increase of latency. Code is released at https://github.com/TheMaXiao/DiTMoS
format text
author MA, Xiao
HE, Shengfeng
QIAO, Hezhe
MA, Dong
author_facet MA, Xiao
HE, Shengfeng
QIAO, Hezhe
MA, Dong
author_sort MA, Xiao
title DiTMoS: Delving into diverse tiny-model selection on microcontrollers
title_short DiTMoS: Delving into diverse tiny-model selection on microcontrollers
title_full DiTMoS: Delving into diverse tiny-model selection on microcontrollers
title_fullStr DiTMoS: Delving into diverse tiny-model selection on microcontrollers
title_full_unstemmed DiTMoS: Delving into diverse tiny-model selection on microcontrollers
title_sort ditmos: delving into diverse tiny-model selection on microcontrollers
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/8670
https://ink.library.smu.edu.sg/context/sis_research/article/9673/viewcontent/2403.09035.pdf
_version_ 1795302168321851392