TrueGPT: can you privately extract algorithms from ChatGPT in tabular classification?

Recently, it has been shown that Large Language Models (LLMs) achieve impressive zero-shot classification on tabular data, revealing an internal algorithm ALLM without explicit training data. We predict that ALLM will become a standard for tabular data classification, replacing resource-intensive...

Full description

Saved in:
Bibliographic Details
Main Author: Soegeng, Hans Farrell
Other Authors: Thomas Peyrin
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/175642
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:Recently, it has been shown that Large Language Models (LLMs) achieve impressive zero-shot classification on tabular data, revealing an internal algorithm ALLM without explicit training data. We predict that ALLM will become a standard for tabular data classification, replacing resource-intensive custom ML models. However, LLM complexity hinders regulatory transparency. To address this, we introduce a method to approximate ALLM with human-interpretable binary feature rules Aerule. We utilize the TT-rules (Truth Table rules) model developed by Benamira et al., 2023 to extract the binary rules through the LLM inference of tabular datasets. Following the extraction and approximation processes, we set aside the LLM and exclusively rely on Aerule for inference. Our method is fully automatic. We validate the approach on 8 public tabular datasets, adding a user option to activate privacy-preserving feature to ensure owner data protection.