Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models
Leveraging large-scale datasets from open-source projects and advances in large language models, recent progress has led to sophisticated code models for key software engineering tasks, such as program repair and code completion. These models are trained on data from various sources, including publi...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9889 https://ink.library.smu.edu.sg/context/sis_research/article/10889/viewcontent/2310.01166v2.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-10889 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-108892025-01-02T09:08:43Z Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models YANG, Zhou ZHAO, Zhipeng WANG, Chenyu SHI, Jieke KIM, Dongsum HAN, Donggyun LO, David Leveraging large-scale datasets from open-source projects and advances in large language models, recent progress has led to sophisticated code models for key software engineering tasks, such as program repair and code completion. These models are trained on data from various sources, including public open-source projects like GitHub and private, confidential code from companies, raising significant privacy concerns. This paper investigates a crucial but unexplored question: What is the risk of membership information leakage in code models? Membership leakage refers to the vulnerability where an attacker can infer whether a specific data point was part of the training dataset. We present Gotcha , a novel membership inference attack method designed for code models, and evaluate its effectiveness on Java-based datasets. Gotcha simultaneously considers three key factors: model input, model output, and ground truth. Our ablation study confirms that each factor significantly enhances attack performance. Our ablation study confirms that each factor significantly enhances attack performance. Our investigation reveals a troubling finding: membership leakage risk is significantly elevated . While previous methods had accuracy close to random guessing, Gotcha achieves high precision, with a true positive rate of 0.95 and a low false positive rate of 0.10. We also demonstrate that the attacker's knowledge of the victim model (e.g., model architecture and pre-training data) affects attack success. Additionally, modifying decoding strategies can help reduce membership leakage risks. This research highlights the urgent need to better understand the privacy vulnerabilities of code models and develop strong countermeasures against these threats. 2024-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9889 info:doi/10.1109/TSE.2024.3482719 https://ink.library.smu.edu.sg/context/sis_research/article/10889/viewcontent/2310.01166v2.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Membership inference attack Privacy Large Langauge Models for code Code completion Information Security Numerical Analysis and Scientific Computing |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Membership inference attack Privacy Large Langauge Models for code Code completion Information Security Numerical Analysis and Scientific Computing |
spellingShingle |
Membership inference attack Privacy Large Langauge Models for code Code completion Information Security Numerical Analysis and Scientific Computing YANG, Zhou ZHAO, Zhipeng WANG, Chenyu SHI, Jieke KIM, Dongsum HAN, Donggyun LO, David Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models |
description |
Leveraging large-scale datasets from open-source projects and advances in large language models, recent progress has led to sophisticated code models for key software engineering tasks, such as program repair and code completion. These models are trained on data from various sources, including public open-source projects like GitHub and private, confidential code from companies, raising significant privacy concerns. This paper investigates a crucial but unexplored question: What is the risk of membership information leakage in code models? Membership leakage refers to the vulnerability where an attacker can infer whether a specific data point was part of the training dataset. We present Gotcha , a novel membership inference attack method designed for code models, and evaluate its effectiveness on Java-based datasets. Gotcha simultaneously considers three key factors: model input, model output, and ground truth. Our ablation study confirms that each factor significantly enhances attack performance. Our ablation study confirms that each factor significantly enhances attack performance. Our investigation reveals a troubling finding: membership leakage risk is significantly elevated . While previous methods had accuracy close to random guessing, Gotcha achieves high precision, with a true positive rate of 0.95 and a low false positive rate of 0.10. We also demonstrate that the attacker's knowledge of the victim model (e.g., model architecture and pre-training data) affects attack success. Additionally, modifying decoding strategies can help reduce membership leakage risks. This research highlights the urgent need to better understand the privacy vulnerabilities of code models and develop strong countermeasures against these threats. |
format |
text |
author |
YANG, Zhou ZHAO, Zhipeng WANG, Chenyu SHI, Jieke KIM, Dongsum HAN, Donggyun LO, David |
author_facet |
YANG, Zhou ZHAO, Zhipeng WANG, Chenyu SHI, Jieke KIM, Dongsum HAN, Donggyun LO, David |
author_sort |
YANG, Zhou |
title |
Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models |
title_short |
Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models |
title_full |
Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models |
title_fullStr |
Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models |
title_full_unstemmed |
Gotcha ! This model uses my code ! Evaluating membership leakage risks in code models |
title_sort |
gotcha ! this model uses my code ! evaluating membership leakage risks in code models |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/9889 https://ink.library.smu.edu.sg/context/sis_research/article/10889/viewcontent/2310.01166v2.pdf |
_version_ |
1821237275635220480 |