DYNAMIC RESOURCE ALLOCATION FOR DEEP LEARNING TRAINING USING TENSORFLOW ON KUBERNETES CLUSTER
Distributed deep learning training nowadays use static resource allocation. Using parameter server architecture, deep learning training is carried out by several parameter server (ps) nodes and worker nodes. Their numbers are constant while the training is running, hence static. Consider a traini...
Saved in:
Main Author: | Yesa Surya, Rahmad |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/39082 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
Similar Items
-
AUTOSCALING FOR MACHINE LEARNING TRAINING USING TENSORFLOW ON KUBERNETES
by: Ardi Mulia, Stefanus -
DYNAMIC RESOURCE SCHEDULER FOR DISTRIBUTED DEEP LEARNING TRAINING IN KUBERNETES
by: Fadhriga Bestari, Muhammad -
Intrusion Detection by Deep Learning with TensorFlow
by: Navaporn Chockwanich, et al.
Published: (2020) -
GitOps in Kubernetes clusters
by: Poh, Kai Kiat
Published: (2022) -
DEVELOPMENT OF A DEEP LEARNING-BASED AUTOSCALER ON KUBERNETES
by: Fadhil Al Hafidz, Rozan