MLModelCI : an automatic cloud platform for efficient MLaaS
MLModelCI provides multimedia researchers and developers with a one-stop platform for efficient machine learning (ML) services. The system leverages DevOps techniques to optimize, test, and manage models. It also containerizes and deploys these optimized and validated models as cloud services (ML...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/152994 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-152994 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1529942021-10-27T08:47:47Z MLModelCI : an automatic cloud platform for efficient MLaaS Zhang, Huaizheng Li, Yuanming Huang, Yizheng Wen, Yonggang Yin, Jianxiong Guan, Kyle School of Computer Science and Engineering 28th ACM International Conference on Multimedia Engineering::Computer science and engineering Engineering::Computer science and engineering::Computing methodologies Cloud Computing Inference Serving MLModelCI provides multimedia researchers and developers with a one-stop platform for efficient machine learning (ML) services. The system leverages DevOps techniques to optimize, test, and manage models. It also containerizes and deploys these optimized and validated models as cloud services (MLaaS). In its essence, MLModelCI serves as a housekeeper to help users publish models. The models are first automatically converted to optimized formats for production purpose and then profiled under different settings (e.g., batch size and hardware). The profiling information can be used as guidelines for balancing the trade-off between performance and cost of MLaaS. Finally, the system dockerizes the models for ease of deployment to cloud environments. A key feature of MLModelCI is the implementation of a controller, which allows elastic evaluation which only utilizes idle workers while maintaining online service quality. Our system bridges the gap between current ML training and serving systems and thus free developers from manual and tedious work often associated with service deployment. We release the platform as an open-source project on GitHub under Apache 2.0 license, with the aim that it will facilitate and streamline more large-scale ML applications and research projects. Energy Market Authority (EMA) Nanyang Technological University National Research Foundation (NRF) This research is supported in part and jointly by the National Research Foundation, Singapore, and the Energy Market Authority, under its Energy Programme (EP Award Ref. NRF2017EWT-EP003- 023) and a project fund from NTU (Ref. NTU–ACE2020-01). 2021-10-27T08:47:47Z 2021-10-27T08:47:47Z 2020 Conference Paper Zhang, H., Li, Y., Huang, Y., Wen, Y., Yin, J. & Guan, K. (2020). MLModelCI : an automatic cloud platform for efficient MLaaS. 28th ACM International Conference on Multimedia, 4453-4456. https://dx.doi.org/10.1145/3394171.3414535 9781450379885 https://hdl.handle.net/10356/152994 10.1145/3394171.3414535 2-s2.0-85106957354 4453 4456 en NRF2017EWT-EP003- 023 NTU–ACE2020-01 © 2020 Association for Computing Machinery. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Engineering::Computer science and engineering::Computing methodologies Cloud Computing Inference Serving |
spellingShingle |
Engineering::Computer science and engineering Engineering::Computer science and engineering::Computing methodologies Cloud Computing Inference Serving Zhang, Huaizheng Li, Yuanming Huang, Yizheng Wen, Yonggang Yin, Jianxiong Guan, Kyle MLModelCI : an automatic cloud platform for efficient MLaaS |
description |
MLModelCI provides multimedia researchers and developers with a one-stop
platform for efficient machine learning (ML) services. The system leverages
DevOps techniques to optimize, test, and manage models. It also containerizes
and deploys these optimized and validated models as cloud services (MLaaS). In
its essence, MLModelCI serves as a housekeeper to help users publish models.
The models are first automatically converted to optimized formats for
production purpose and then profiled under different settings (e.g., batch size
and hardware). The profiling information can be used as guidelines for
balancing the trade-off between performance and cost of MLaaS. Finally, the
system dockerizes the models for ease of deployment to cloud environments. A
key feature of MLModelCI is the implementation of a controller, which allows
elastic evaluation which only utilizes idle workers while maintaining online
service quality. Our system bridges the gap between current ML training and
serving systems and thus free developers from manual and tedious work often
associated with service deployment. We release the platform as an open-source
project on GitHub under Apache 2.0 license, with the aim that it will
facilitate and streamline more large-scale ML applications and research
projects. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Zhang, Huaizheng Li, Yuanming Huang, Yizheng Wen, Yonggang Yin, Jianxiong Guan, Kyle |
format |
Conference or Workshop Item |
author |
Zhang, Huaizheng Li, Yuanming Huang, Yizheng Wen, Yonggang Yin, Jianxiong Guan, Kyle |
author_sort |
Zhang, Huaizheng |
title |
MLModelCI : an automatic cloud platform for efficient MLaaS |
title_short |
MLModelCI : an automatic cloud platform for efficient MLaaS |
title_full |
MLModelCI : an automatic cloud platform for efficient MLaaS |
title_fullStr |
MLModelCI : an automatic cloud platform for efficient MLaaS |
title_full_unstemmed |
MLModelCI : an automatic cloud platform for efficient MLaaS |
title_sort |
mlmodelci : an automatic cloud platform for efficient mlaas |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/152994 |
_version_ |
1715201517708902400 |