Fri 22 Jul 2022 00:00 - 00:20 at ISSTA 2 - Session 1-10: Neural Networks, Learning, NLP A
Due to the labor-intensive nature of manual test oracle construction, various automated testing techniques have been proposed to enhance the reliability of Natural Language Processing (NLP) software. In theory, these techniques mutate an existing test case (\textit{e.g.}, a sentence with its label) and assume the generated one preserves an equivalent or similar semantic meaning and thus, the same label. However, in practice, many of the generated test cases fail to preserve similar semantic meaning and are unnatural (\textit{e.g.}, grammar errors), which leads to a high false alarm rate and unnatural test cases. Our evaluation study finds that 44% of the test cases generated by the state-of-the-art (SOTA) approaches are false alarms. These test cases require extensive manual checking effort, and instead of improving NLP software, they can even degrade NLP software when utilized in model training. To address this problem, we propose AEON for \textbf{A}utomatic \textbf{E}valuation \textbf{O}f \textbf{N}LP test cases. For each generated test case, it outputs scores based on semantic similarity and language naturalness. We employ AEON to evaluate test cases generated by four popular testing techniques on five datasets across three typical NLP tasks. The results show that AEON aligns the best with human judgment. In particular, AEON achieves the best average precision in detecting semantic inconsistent test cases, outperforming the best baseline metric by 10%. In addition, AEON also has the highest average precision of finding unnatural test cases, surpassing the baselines by more than 15%. Moreover, model training with test cases prioritized by AEON leads to models that are more accurate and robust, demonstrating AEON’s potential in improving NLP software.
Thu 21 JulDisplayed time zone: Seoul change
16:20 - 17:40 | |||
16:20 20mTalk | AEON: A Method for Automatic Evaluation of NLP Test Cases Technical Papers Jen-tse Huang The Chinese University of Hong Kong, Jianping Zhang The Chinese University of Hong Kong, Wenxuan Wang The Chinese University of Hong Kong, Pinjia He The Chinese University of Hong Kong, Shenzhen, Yuxin Su Sun Yat-sen University, Michael Lyu The Chinese University of Hong Kong DOI | ||
16:40 20mTalk | HybridRepair: Towards Annotation-Efficient Repair for Deep Learning Models Technical Papers DOI | ||
17:00 20mTalk | Improving Cross-Platform Binary Analysis using Representation Learning via Graph Alignment Technical Papers Geunwoo Kim University of California, Irvine, USA, Sanghyun Hong Oregon State University, Michael Franz University of California, Irvine, Dokyung Song Yonsei University, South Korea DOI | ||
17:20 20mTalk | Human-in-the-Loop Oracle Learning for Semantic Bugs in String Processing Programs Technical Papers Charaka Geethal Monash University, Thuan Pham The University of Melbourne, Aldeida Aleti Monash University, Marcel Böhme MPI-SP, Germany and Monash University, Australia DOI Pre-print |
Fri 22 JulDisplayed time zone: Seoul change
00:00 - 01:00 | |||
00:00 20mTalk | AEON: A Method for Automatic Evaluation of NLP Test Cases Technical Papers Jen-tse Huang The Chinese University of Hong Kong, Jianping Zhang The Chinese University of Hong Kong, Wenxuan Wang The Chinese University of Hong Kong, Pinjia He The Chinese University of Hong Kong, Shenzhen, Yuxin Su Sun Yat-sen University, Michael Lyu The Chinese University of Hong Kong DOI | ||
00:20 20mTalk | Simple Techniques Work Surprisingly Well for Neural Network Test Prioritization and Active Learning (Replicability Study) Technical Papers DOI Pre-print | ||
00:40 20mTalk | ε-weakened Robustness of Deep Neural Networks Technical Papers Pei Huang State Key Laboratory of Computer Science, Institution of Software, Chinese Academy of Sciences, Yuting Yang Institute of Computing Technology,Chinese Academy of Sciences; University of Chinese Academy of Sciences, Minghao Liu Institute of Software, Chinese Academy of Sciences, Fuqi Jia State Key Laboratory of Computer Science, Institution of Software, Chinese Academy of Sciences, Feifei Ma Institute of Software, Chinese Academy of Sciences, Jian Zhang Institute of Software at Chinese Academy of Sciences, China DOI |