Write a Blog >>
ICSE 2023
Sun 14 - Sat 20 May 2023 Melbourne, Australia
Wed 17 May 2023 14:52 - 15:00 at Meeting Room 104 - AI systems engineering Chair(s): Xin Peng

Deep neural network (DNN) models typically have many hyperparameters that can be configured to achieve optimal performance on a particular dataset. Practitioners usually tune the hyperparameters of their DNN models by training a number of trial models with different configurations of the hyperparameters, to find the optimal hyperparameter configuration that maximizes the training accuracy or minimizes the training loss. As such hyperparameter tuning usually focuses on the model accuracy or the loss function, it is not clear and remains under-explored how the process impacts other performance properties of DNN models, such as inference latency and model size. On the other hand, standard DNN models are often large in size and computing-intensive, prohibiting them from being directly deployed in resource-bounded environments such as mobile devices and Internet of Things (IoT) devices. To tackle this problem, various model optimization techniques (e.g., pruning or quantization) are proposed to make DNN models smaller and less computing-intensive so that they are better suited for resource-bounded environments. However, it is neither clear how the model optimization techniques impact other performance properties of DNN models such as inference latency and battery consumption, nor how the model optimization techniques impact the effect of hyperparameter tuning (i.e., the compounding effect). Therefore, in this paper, we perform a comprehensive study on four representative and widely-adopted DNN models, i.e., CNN image classification, Resnet-50, CNN text classification, and LSTM sentiment classification, to investigate how different DNN model hyperparameters affect the standard DNN models, as well as how the hyperparameter tuning combined with model optimization affect the optimized DNN models, in terms of various performance properties (e.g., inference latency or battery consumption). Our empirical results indicate that tuning specific hyperparameters has heterogeneous impact on the performance of DNN models across different models and different performance properties. In particular, although the top tuned DNN models usually have very similar accuracy, they may have significantly different performance in terms of other aspects (e.g., inference latency). We also observe that model optimization has a confounding effect on the impact of hyperparameters on DNN model performance. For example, two sets of hyperparameters may result in standard models with similar performance but their performance may become significantly different after they are optimized and deployed on the mobile device. Our findings highlight that practitioners can benefit from paying attention to a variety of performance properties and the confounding effect of model optimization when tuning and optimizing their DNN models.

Wed 17 May

Displayed time zone: Hobart change

13:45 - 15:15
13:45
15m
Talk
FedDebug: Systematic Debugging for Federated Learning Applications
Technical Track
Waris Gill Virginia Tech, Ali Anwar University of Minnesota, Muhammad Ali Gulzar Virginia Tech
14:00
15m
Talk
Practical and Efficient Model Extraction of Sentiment Analysis APIs
Technical Track
Weibin Wu Sun Yat-sen University, Jianping Zhang The Chinese University of Hong Kong, Victor Junqiu Wei The Hong Kong Polytechnic University, Xixian Chen Tencent, Zibin Zheng School of Software Engineering, Sun Yat-sen University, Irwin King The Chinese University of Hong Kong, Michael Lyu The Chinese University of Hong Kong
14:15
15m
Talk
CrossCodeBench: Benchmarking Cross-Task Generalization of Source Code Models
Technical Track
Changan Niu Software Institute, Nanjing University, Chuanyi Li Nanjing University, Vincent Ng Human Language Technology Research Institute, University of Texas at Dallas, Richardson, TX 75083-0688, Bin Luo Nanjing University
Pre-print
14:30
15m
Talk
Challenges in Adopting Artificial Intelligence Based User Input Verification Framework in Reporting Software Systems
SEIP - Software Engineering in Practice
Dong Jae Kim Concordia University, Tse-Hsun (Peter) Chen Concordia University, Steve Sporea , Andrei Toma ERA Environmental Management Solutions, Laura Weinkam , Sarah Sajedi ERA Environmental Management Solutions, Steve Sporea
14:45
7m
Talk
Towards Understanding Quality Challenges of the Federated Learning for Neural Networks: A First Look from the Lens of Robustness
Journal-First Papers
Amin Eslami Abyane University of Calgary, Derui Zhu Technical University of Munich, Roberto Souza University of Calgary, Lei Ma University of Alberta, Hadi Hemmati York University
14:52
7m
Talk
An Empirical Study of the Impact of Hyperparameter Tuning and Model Optimization on the Performance Properties of Deep Neural Networks
Journal-First Papers
Lizhi Liao Concordia University, Heng Li Polytechnique Montréal, Weiyi Shang University of Waterloo, Lei Ma University of Alberta
15:00
7m
Talk
Black-box Safety Analysis and Retraining of DNNs based on Feature Extraction and Clustering
Journal-First Papers
Mohammed Attaoui University of Luxembourg, Hazem FAHMY University of Luxembourg, Fabrizio Pastore University of Luxembourg, Lionel Briand University of Luxembourg; University of Ottawa
Link to publication Pre-print
15:07
7m
Talk
Iterative Assessment and Improvement of DNN Operational Accuracy
NIER - New Ideas and Emerging Results
Antonio Guerriero Università di Napoli Federico II, Roberto Pietrantuono Università di Napoli Federico II, Stefano Russo Università di Napoli Federico II
Pre-print