Pre-trained language models have achieved many prominent results in natural language processing. Since software engineering tasks include many natural language tasks, the application of pre-trained language models have received much attention in software engineering tasks. However, pre-training a large volume of source code requires a huge amount of computational resources and time. In this study, we propose an additional pre-training approach to a well-trained language model. Our initial results on mT5, multilingual T5 with an additional pretraining of Python code shows improved performance on multiple software engineering tasks including code generation, code summarisation, code repair, and error diagnosis.
Mihato Ueda Department of Informatics Education, Tokyo Gakugei Unversity, Yousuke Asano Graduate School of Education, Tokyo Gakugei Unversity, Hane Kondo Graduate School of Education, Tokyo Gakugei Unversity, Oh Sato Graduate School of Education, Tokyo Gakugei Unversity, Atsuo Hazeyama Tokyo Gakugei University
Anivesh Panjiyar ABV-Indian Institute of Information Technology and Management Gwalior, Debanjan Sadhya ABV-Indian Institute of Information Technology and Management Gwalior