OPTIMIZING INFERENCE PERFORMANCE OF BERT ON CPUS USING APACHE TVM
BERT as a model that composed of Transformer layer is game changing for field of natural language processing (NLP). There has been a lot of study to speedup training the model however only relatively little efforts are made to improve their inference performance. Also, not all machine learning...
Saved in:
Main Author: | Legowo, Setyo |
---|---|
Format: | Theses |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/56144 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
Similar Items
-
TVM telecommunication and TV mobile
by: Cao, C., et al.
Published: (2013) -
ThunderSVM: A fast SVM library on GPUs and CPUs
by: Zeyi Wen, et al.
Published: (2020) -
Augmented virtuality for apache simulated training
by: Lin, Junyuan.
Published: (2011) -
SHORT ANSWER GRADING IN BAHASA USING BERT
by: Putri Dinanti, Anissa -
Using CodeBERT model for vulnerability detection
by: Zhou, ZhiWei
Published: (2022)