PromptLink: Multi-template prompt learning with adversarial training for issue-commit link recovery
In recent years, Prompt Learning, based on pre-training, prompting, and prediction, has achieved significant success in natural language processing (NLP). The current issue-commit link recovery (ILR) method converts the ILR into a classification task using pre-trained language models (PLMs) and dedicated neural networks. However, due to inconsistencies between the ILR task and PLMs, these methods not fully leverage the semantic information in PLMs. To imitate the above problem, we make the first trial of the new paradigm to propose a Multi-template prompt learning method with adversarial training for issue-commit link recovery (PromptLink), which transforms the ILR task into a cloze task through the template. Specifically, a Multi-template PromptLink is designed to enhance the generalisation capability by integrating various templates and adopting adversarial training to mitigate the model overfitting. Experiments are conducted on six open-source projects and comprehensively evaluated across six commonly measures. The results show that PromptLink achieves an average F1 of 96.10%, Precision of 96.49%, Recall of 95.92%, MCC of 94.04%, AUC of 96.05%, and ACC of 98.15%, significantly outperforming existing state-of-the-art methods on all measures. Overall, PromptLink not only enhances performance and generalisation but also emerges new ideas and methods for future research. The source code of PromptLink is available at https://figshare.com/s/6130d42ff464c579cdec.
Thu 24 OctDisplayed time zone: Brussels, Copenhagen, Madrid, Paris change
16:00 - 17:30 | Machine learning for software engineeringESEM Technical Papers / ESEM Emerging Results, Vision and Reflection Papers Track / ESEM Journal-First Papers at Telensenyament (B3 Building - 1st Floor) Chair(s): Luigi Quaranta University of Bari, Italy | ||
16:00 20mFull-paper | A Transformer-based Approach for Augmenting Software Engineering Chatbots Datasets ESEM Technical Papers Ahmad Abdellatif University of Calgary, Khaled Badran Concordia University, Canada, Diego Costa Concordia University, Canada, Emad Shihab Concordia University | ||
16:20 20mFull-paper | Unsupervised and Supervised Co-learning for Comment-based Codebase Refining and its Application in Code Search ESEM Technical Papers Gang Hu School of Information Science & Engineering, Yunnan University, Xiaoqin Zeng School of Information Science & Engineering, Yunnan University, Wanlong Yu , Min Peng , YUAN Mengting School of Computer Science, Wuhan University, Wuhan, China, Liang Duan | ||
16:40 20mFull-paper | Good things come in three: Generating SO Post Titles with Pre-Trained Models, Self Improvement and Post Ranking ESEM Technical Papers Duc Anh Le Hanoi University of Science and Technology, Anh M. T. Bui Hanoi University of Science and Technology, Phuong T. Nguyen University of L’Aquila, Davide Di Ruscio University of L'Aquila Pre-print | ||
17:00 15mVision and Emerging Results | PromptLink: Multi-template prompt learning with adversarial training for issue-commit link recovery ESEM Emerging Results, Vision and Reflection Papers Track Yang Deng The School of Computer Science and Artificial Intelligence, Wuhan Textile University, Wuhan, China, Bangchao Wang Wuhan Textile University, Zhiyuan Zou The School of Computer Science and Artificial Intelligence, Wuhan Textile University, Wuhan, China, Luyao Ye The School of Computer Science and Artificial Intelligence, Wuhan Textile University, Wuhan, China | ||
17:15 15mJournal Early-Feedback | GPTSniffer: A CodeBERT-based classifier to detect source code written by ChatGPT ESEM Journal-First Papers Phuong T. Nguyen University of L’Aquila, Juri Di Rocco University of L'Aquila, Claudio Di Sipio University of l'Aquila, Riccardo Rubei University of L'Aquila, Davide Di Ruscio University of L'Aquila, Massimiliano Di Penta University of Sannio, Italy Link to publication DOI Pre-print |