Unveiling memorization in code models
The availability of large-scale datasets, advanced architectures, and powerful computational resources have led to effective code models that automate diverse software engineering activities. The datasets usually consist of billions of lines of code from both open-source and private repositories. A...
Saved in:
Main Authors: | YANG, Zhou, ZHAO, Zhipeng, WANG, Chenyu, SHI, Jieke, KIM, Dongsun, HAN, DongGyun, LO, David |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9246 https://ink.library.smu.edu.sg/context/sis_research/article/10246/viewcontent/3597503.3639074.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Unveiling code pre-trained models: Investigating syntax and semantics capacities
by: MA, Wei, et al.
Published: (2024) -
Code coverage and postrelease defects: A large-scale study on open source projects
by: KOCHHAR, Pavneet Singh, et al.
Published: (2017) -
Code search is all you need? Improving code suggestions with code search
by: CHEN, Junkai, et al.
Published: (2024) -
Multi-modal API recommendation
by: IRSAN, Ivana Clairine, et al.
Published: (2023) -
Automatic code review by learning the revision of source code
by: SHI, Shu-Ting, et al.
Published: (2019)