Programmatic policies for interpretable reinforcement learning using pre-trained models
Decision Trees (DTs) are widely used in machine learning due to their critical interpretability. However, training DTs in a Reinforcement Learning (RL) setting is challenging. In the project, we present a framework to improve the interpretability of reinforcement learning (RL) by generating prog...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181169 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | Decision Trees (DTs) are widely used in machine learning due to their critical
interpretability. However, training DTs in a Reinforcement Learning (RL) setting
is challenging.
In the project, we present a framework to improve the interpretability of
reinforcement learning (RL) by generating programmatic policies through large
language models (LLMs). Initially looking at decision trees within latent space
representation, we proceed to develop and apply a “code reflection” framework
in the Karel environment, which provides a controlled setting for evaluating task
performance. The framework leverages prompt engineering and code reflection
to optimize program synthesis, producing interpretable policies.
The report discusses the findings and concludes with recommendations for future
work at the end. |
---|