Enhancing downstream ML performance with unconditional diffusion models for return predictions
This study addresses the challenge of enhancing model generalization in financial market return predictions crucial due to the dynamic and unpredictable nature of financial markets. Traditional models often fail to generalize across market conditions, largely due to their inability to capture market...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/175255 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-175255 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1752552024-04-26T15:41:14Z Enhancing downstream ML performance with unconditional diffusion models for return predictions Agarwala, Pratham Bo An School of Computer Science and Engineering boan@ntu.edu.sg Computer and Information Science Diffusion Financial market prediction Data augmentation This study addresses the challenge of enhancing model generalization in financial market return predictions crucial due to the dynamic and unpredictable nature of financial markets. Traditional models often fail to generalize across market conditions, largely due to their inability to capture market dynamics effectively. Previous methods, reliant on simple transformations or generative adversarial networks (GANs) with inherent training instability, fall short in addressing these challenges. To bridge this gap, our research introduces an innovative approach that leverages Unconditional Diffusion Models with self-guidance mechanism for conditioning during inference to perform financial data augmentation. We incorporate technical indicators to help the diffusion model comprehend financial time series, and employ ensemble predictions for each context window to construct a stable output. We use refinement process which leverages the implicit probability density learned by the diffusion model serving as a prior, to iteratively improve the output. A filtration layer, based on EMA, RSI and accompanied by some randomness, is added to provide more control over the fidelity and diversity of the augmented dataset. We evaluate augmented dataset quality through metrics assessing affinity (indistinguishability from real data) and diversity (coverage of real data). These include: (1)Discriminative Score: measures classifier’s ability to differentiate between real and synthetic. (2) Mutual Information Coefficient: measures amount of shared information between the synthetic and real. (3) Variation in Information: Evaluates the diversity within the synthetic data. Findings indicate a direct link between these factors and model performance, with high-diversity and optimal fidelity datasets enhancing forecasting accuracy. Our augmentation method outperformed GANs, mixup-based, and transformation-based techniques in predictive accuracy, diversity, and closely aligned with real market dynamics. Finally, we evaluate a real life application, by employing a simple trading strategy using the downstream predictions and its back test on the test data showcased significantly better Return on Investment and Sharpe Ratio when augmentation is applied, indicating better risk-adjusted returns. Bachelor's degree 2024-04-23T05:13:05Z 2024-04-23T05:13:05Z 2024 Final Year Project (FYP) Agarwala, P. (2024). Enhancing downstream ML performance with unconditional diffusion models for return predictions. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/175255 https://hdl.handle.net/10356/175255 en SCSE23-0061 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Diffusion Financial market prediction Data augmentation |
spellingShingle |
Computer and Information Science Diffusion Financial market prediction Data augmentation Agarwala, Pratham Enhancing downstream ML performance with unconditional diffusion models for return predictions |
description |
This study addresses the challenge of enhancing model generalization in financial market return predictions crucial due to the dynamic and unpredictable nature of financial markets. Traditional models often fail to generalize across market conditions, largely due to their inability to capture market dynamics effectively. Previous methods, reliant on simple transformations or generative adversarial networks (GANs) with inherent training instability, fall short in addressing these challenges.
To bridge this gap, our research introduces an innovative approach that leverages Unconditional Diffusion Models with self-guidance mechanism for conditioning during inference to perform financial data augmentation. We incorporate technical indicators to help the diffusion model comprehend financial time series, and employ ensemble predictions for each context window to construct a stable output. We use refinement process which leverages the implicit probability density learned by the diffusion model serving as a prior, to iteratively improve the output. A filtration layer, based on EMA, RSI and accompanied by some randomness, is added to provide more control over the fidelity and diversity of the augmented dataset.
We evaluate augmented dataset quality through metrics assessing affinity (indistinguishability from real data) and diversity (coverage of real data). These include: (1)Discriminative Score: measures classifier’s ability to differentiate between real and synthetic. (2) Mutual Information Coefficient: measures amount of shared information between the synthetic and real. (3) Variation in Information: Evaluates the diversity within the synthetic data. Findings indicate a direct link between these factors and model performance, with high-diversity and optimal fidelity datasets enhancing forecasting accuracy.
Our augmentation method outperformed GANs, mixup-based, and transformation-based techniques in predictive accuracy, diversity, and closely aligned with real market dynamics.
Finally, we evaluate a real life application, by employing a simple trading strategy using the downstream predictions and its back test on the test data showcased significantly better Return on Investment and Sharpe Ratio when augmentation is applied, indicating better risk-adjusted returns. |
author2 |
Bo An |
author_facet |
Bo An Agarwala, Pratham |
format |
Final Year Project |
author |
Agarwala, Pratham |
author_sort |
Agarwala, Pratham |
title |
Enhancing downstream ML performance with unconditional diffusion models for return predictions |
title_short |
Enhancing downstream ML performance with unconditional diffusion models for return predictions |
title_full |
Enhancing downstream ML performance with unconditional diffusion models for return predictions |
title_fullStr |
Enhancing downstream ML performance with unconditional diffusion models for return predictions |
title_full_unstemmed |
Enhancing downstream ML performance with unconditional diffusion models for return predictions |
title_sort |
enhancing downstream ml performance with unconditional diffusion models for return predictions |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/175255 |
_version_ |
1814047181576863744 |