On the transferability of pre-trained language models for low-resource programming languages

A recent study by Ahmed and Devanbu reported that using a corpus of code written in multilingual datasets to fine-tune multilingual Pre-trained Language Models (PLMs) achieves higher performance as opposed to using a corpus of code written in just one programming language. However, no analysis was m...

Full description

Saved in:
Bibliographic Details
Main Authors: CHEN, Fuxiang, FARD, Fatemeh H., LO, David, BRYKSIN, Timofey
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7693
https://ink.library.smu.edu.sg/context/sis_research/article/8696/viewcontent/On_the_transfer.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English