DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING
This research presents a deep reinforcement learning model for algorithmic trading of cryptocurrencies. The aim of the model is to help traders earn greater profits than traditional strategies. Traditional strategies can be used to gain profits. However, this strategy tends to require more knowle...
Saved in:
Main Author: | |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/66647 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
id |
id-itb.:66647 |
---|---|
spelling |
id-itb.:666472022-06-29T18:31:36ZDEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING Shan, Elbert Indonesia Final Project cryptocurrency, Bitcoin, trading, deep reinforcement learning INSTITUT TEKNOLOGI BANDUNG https://digilib.itb.ac.id/gdl/view/66647 This research presents a deep reinforcement learning model for algorithmic trading of cryptocurrencies. The aim of the model is to help traders earn greater profits than traditional strategies. Traditional strategies can be used to gain profits. However, this strategy tends to require more knowledge, experience and time compared to machine learning-based solutions in order to generate optimal returns. Models will be trained to trade on the cryptocurrency market. Model inputs are 1 minute interval candlestick data and technical indicators for BTC/USDT cryptocurrency pair. The output of the model is buy, hold, or sell signals. The model development was carried out based on the CRISP-DM methodology. The model will be created with PPO algorithm and a custom environment that follows gym interface. The transaction fee rate of 0.1% will be taken into consideration by the model in determining the strategy. The model will be trained and tested in constant-sized episodes of 720 timestep minutes (12 hours). The trading positions that can be done by the model are neutral (owning 0 amount of assets) and long. Reward for the model will be calculated based on profit earned on each position. The reward system will also include a certain penalty system to prevent the model from always implementing Buy and Hold strategy or never going long. The performance of the model is improved by optimizing the hyperparameter values using Optuna. The performance of the model is compared to the Buy and Hold strategy. The tests results show that the model produced still cannot beat the Buy and Hold strategy. text |
institution |
Institut Teknologi Bandung |
building |
Institut Teknologi Bandung Library |
continent |
Asia |
country |
Indonesia Indonesia |
content_provider |
Institut Teknologi Bandung |
collection |
Digital ITB |
language |
Indonesia |
description |
This research presents a deep reinforcement learning model for algorithmic trading of
cryptocurrencies. The aim of the model is to help traders earn greater profits than
traditional strategies. Traditional strategies can be used to gain profits. However, this
strategy tends to require more knowledge, experience and time compared to machine
learning-based solutions in order to generate optimal returns. Models will be trained
to trade on the cryptocurrency market. Model inputs are 1 minute interval candlestick
data and technical indicators for BTC/USDT cryptocurrency pair. The output of the
model is buy, hold, or sell signals. The model development was carried out based on
the CRISP-DM methodology. The model will be created with PPO algorithm and a
custom environment that follows gym interface. The transaction fee rate of 0.1% will
be taken into consideration by the model in determining the strategy. The model will be
trained and tested in constant-sized episodes of 720 timestep minutes (12 hours). The
trading positions that can be done by the model are neutral (owning 0 amount of assets)
and long. Reward for the model will be calculated based on profit earned on each
position. The reward system will also include a certain penalty system to prevent the
model from always implementing Buy and Hold strategy or never going long. The
performance of the model is improved by optimizing the hyperparameter values using
Optuna. The performance of the model is compared to the Buy and Hold strategy. The
tests results show that the model produced still cannot beat the Buy and Hold strategy. |
format |
Final Project |
author |
Shan, Elbert |
spellingShingle |
Shan, Elbert DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING |
author_facet |
Shan, Elbert |
author_sort |
Shan, Elbert |
title |
DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING |
title_short |
DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING |
title_full |
DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING |
title_fullStr |
DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING |
title_full_unstemmed |
DEEP REINFORCEMENT LEARNING FOR AUTOMATED CRYPTOCURRENCY TRADING |
title_sort |
deep reinforcement learning for automated cryptocurrency trading |
url |
https://digilib.itb.ac.id/gdl/view/66647 |
_version_ |
1822277685203697664 |