Enhancing downstream ML performance with unconditional diffusion models for return predictions

This study addresses the challenge of enhancing model generalization in financial market return predictions crucial due to the dynamic and unpredictable nature of financial markets. Traditional models often fail to generalize across market conditions, largely due to their inability to capture market...

Full description

Saved in:
Bibliographic Details
Main Author: Agarwala, Pratham
Other Authors: Bo An
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/175255
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-175255
record_format dspace
spelling sg-ntu-dr.10356-1752552024-04-26T15:41:14Z Enhancing downstream ML performance with unconditional diffusion models for return predictions Agarwala, Pratham Bo An School of Computer Science and Engineering boan@ntu.edu.sg Computer and Information Science Diffusion Financial market prediction Data augmentation This study addresses the challenge of enhancing model generalization in financial market return predictions crucial due to the dynamic and unpredictable nature of financial markets. Traditional models often fail to generalize across market conditions, largely due to their inability to capture market dynamics effectively. Previous methods, reliant on simple transformations or generative adversarial networks (GANs) with inherent training instability, fall short in addressing these challenges. To bridge this gap, our research introduces an innovative approach that leverages Unconditional Diffusion Models with self-guidance mechanism for conditioning during inference to perform financial data augmentation. We incorporate technical indicators to help the diffusion model comprehend financial time series, and employ ensemble predictions for each context window to construct a stable output. We use refinement process which leverages the implicit probability density learned by the diffusion model serving as a prior, to iteratively improve the output. A filtration layer, based on EMA, RSI and accompanied by some randomness, is added to provide more control over the fidelity and diversity of the augmented dataset. We evaluate augmented dataset quality through metrics assessing affinity (indistinguishability from real data) and diversity (coverage of real data). These include: (1)Discriminative Score: measures classifier’s ability to differentiate between real and synthetic. (2) Mutual Information Coefficient: measures amount of shared information between the synthetic and real. (3) Variation in Information: Evaluates the diversity within the synthetic data. Findings indicate a direct link between these factors and model performance, with high-diversity and optimal fidelity datasets enhancing forecasting accuracy. Our augmentation method outperformed GANs, mixup-based, and transformation-based techniques in predictive accuracy, diversity, and closely aligned with real market dynamics. Finally, we evaluate a real life application, by employing a simple trading strategy using the downstream predictions and its back test on the test data showcased significantly better Return on Investment and Sharpe Ratio when augmentation is applied, indicating better risk-adjusted returns. Bachelor's degree 2024-04-23T05:13:05Z 2024-04-23T05:13:05Z 2024 Final Year Project (FYP) Agarwala, P. (2024). Enhancing downstream ML performance with unconditional diffusion models for return predictions. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/175255 https://hdl.handle.net/10356/175255 en SCSE23-0061 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Diffusion
Financial market prediction
Data augmentation
spellingShingle Computer and Information Science
Diffusion
Financial market prediction
Data augmentation
Agarwala, Pratham
Enhancing downstream ML performance with unconditional diffusion models for return predictions
description This study addresses the challenge of enhancing model generalization in financial market return predictions crucial due to the dynamic and unpredictable nature of financial markets. Traditional models often fail to generalize across market conditions, largely due to their inability to capture market dynamics effectively. Previous methods, reliant on simple transformations or generative adversarial networks (GANs) with inherent training instability, fall short in addressing these challenges. To bridge this gap, our research introduces an innovative approach that leverages Unconditional Diffusion Models with self-guidance mechanism for conditioning during inference to perform financial data augmentation. We incorporate technical indicators to help the diffusion model comprehend financial time series, and employ ensemble predictions for each context window to construct a stable output. We use refinement process which leverages the implicit probability density learned by the diffusion model serving as a prior, to iteratively improve the output. A filtration layer, based on EMA, RSI and accompanied by some randomness, is added to provide more control over the fidelity and diversity of the augmented dataset. We evaluate augmented dataset quality through metrics assessing affinity (indistinguishability from real data) and diversity (coverage of real data). These include: (1)Discriminative Score: measures classifier’s ability to differentiate between real and synthetic. (2) Mutual Information Coefficient: measures amount of shared information between the synthetic and real. (3) Variation in Information: Evaluates the diversity within the synthetic data. Findings indicate a direct link between these factors and model performance, with high-diversity and optimal fidelity datasets enhancing forecasting accuracy. Our augmentation method outperformed GANs, mixup-based, and transformation-based techniques in predictive accuracy, diversity, and closely aligned with real market dynamics. Finally, we evaluate a real life application, by employing a simple trading strategy using the downstream predictions and its back test on the test data showcased significantly better Return on Investment and Sharpe Ratio when augmentation is applied, indicating better risk-adjusted returns.
author2 Bo An
author_facet Bo An
Agarwala, Pratham
format Final Year Project
author Agarwala, Pratham
author_sort Agarwala, Pratham
title Enhancing downstream ML performance with unconditional diffusion models for return predictions
title_short Enhancing downstream ML performance with unconditional diffusion models for return predictions
title_full Enhancing downstream ML performance with unconditional diffusion models for return predictions
title_fullStr Enhancing downstream ML performance with unconditional diffusion models for return predictions
title_full_unstemmed Enhancing downstream ML performance with unconditional diffusion models for return predictions
title_sort enhancing downstream ml performance with unconditional diffusion models for return predictions
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/175255
_version_ 1814047181576863744