What Do They Capture? - A Structural Analysis of Pre-Trained Language Models for Source Code
Thu 12 May 2022 04:20 - 04:25 at ICSE room 1-even hours - Machine Learning with and for SE 3 Chair(s): Antinisca Di Marco
Recently, many pre-trained language models for source code have been proposed to model the context of code and serve as a basis for downstream code intelligence tasks such as code completion, code search, and code summarization. Although current language models of code based on masked pre-training and Transformer have achieved promising results, there is still little progress regarding interpretability of existing pre-trained code models. It is not clear why these models work and what feature correlations they can capture. In this paper, we conduct a thorough structural analysis aiming to provide an interpretation of pre-trained language models for source code (e.g., CodeBERT, and GraphCodeBERT) from three distinctive perspectives: (1) attention analysis, (2) probing on the word embedding, and (3) syntax tree induction. Through comprehensive analysis, this paper reveals several insightful findings that may inspire future studies: (1) Attention aligns strongly with the syntax structure of code. (2) Pre-training language models of code can preserve the syntax structure of code in the intermediate representations of each Transformer layer. (3) The pre-trained models of code have the ability of reducing syntax trees of code. Theses findings suggest that it may be helpful to incorporate the syntax structure of code into the process of pre-training for better code representations.
Tue 10 MayDisplayed time zone: Eastern Time (US & Canada) change
Thu 12 MayDisplayed time zone: Eastern Time (US & Canada) change
04:00 - 05:00 | Machine Learning with and for SE 3Technical Track / Journal-First Papers / SEIP - Software Engineering in Practice at ICSE room 1-even hours Chair(s): Antinisca Di Marco University of L'Aquila | ||
04:00 5mTalk | In-IDE Code Generation from Natural Language: Promise and Challenges Journal-First Papers Frank Xu Carnegie Mellon University, Bogdan Vasilescu Carnegie Mellon University, USA, Graham Neubig Carnegie Mellon University | ||
04:05 5mTalk | Active Learning of Discriminative Subgraph Patterns for API Misuse Detection Journal-First Papers Pre-print Media Attached File Attached | ||
04:10 5mTalk | Dependency Tracking for Risk Mitigation in Machine Learning (ML) Systems SEIP - Software Engineering in Practice Xiwei (Sherry) Xu CSIRO Data61, Chen Wang CSIRO DATA61, Zhen Wang CSIRO Data61, Qinghua Lu CSIRO’s Data61, Liming Zhu CSIRO’s Data61; UNSW Media Attached | ||
04:15 5mTalk | DeepFD: Automated Fault Diagnosis and Localization for Deep Learning Programs Technical Track Jialun Cao Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Meiziniu LI Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Xiao Chen Huazhong University of Science and Technology, Ming Wen Huazhong University of Science and Technology, Yongqiang Tian University of Waterloo, Bo Wu MIT-IBM Watson AI Lab in Cambridge, Shing-Chi Cheung Hong Kong University of Science and Technology DOI Pre-print Media Attached | ||
04:20 5mTalk | What Do They Capture? - A Structural Analysis of Pre-Trained Language Models for Source Code Technical Track Yao Wan Huazhong University of Science and Technology, Wei Zhao Huazhong University of Science and Technology, Hongyu Zhang University of Newcastle, Yulei Sui University of Technology Sydney, Guandong Xu University of Technology, Sydney, Hai Jin Huazhong University of Science and Technology Pre-print Media Attached | ||
04:25 5mTalk | A Universal Data Augmentation Approach for Fault Localization Technical Track Huan Xie Chongqing University, Yan Lei School of Big Data & Software Engineering, Chongqing University, Meng Yan Chongqing University, Yue Yu College of Computer, National University of Defense Technology, Changsha 410073, China, Xin Xia Huawei Software Engineering Application Technology Lab, Xiaoguang Mao National University of Defense Technology DOI Pre-print Media Attached | ||
04:30 5mTalk | DeepState: Selecting Test Suites to Enhance the Robustness of Recurrent Neural Networks Technical Track Zixi Liu Nanjing University, Yang Feng Nanjing University, Yining Yin Nanjing University, China, Zhenyu Chen Nanjing University DOI Pre-print Media Attached |