ICSE 2026
Sun 12 - Sat 18 April 2026 Rio de Janeiro, Brazil

This program is tentative and subject to change.

Thu 16 Apr 2026 11:15 - 11:30 at Oceania VII - Software Engineering for AI 4

Quantization reduces the precision of deep neural networks to lower model size and computational demands, but often at the expense of accuracy. Fully quantized models can suffer significant accuracy degradation, and resource-constrained hardware accelerators may not support all quantized operations. A common workaround is selective quantization, where only some layers are quantized while others remain at full precision. However, determining the optimal balance between accuracy and efficiency is a challenging task. To this direction, we propose SeQTO, a framework that enables selective quantization, deployment, and execution of ONNX models on diverse CPU and GPU devices, combined with profiling and multi-objective optimization. SeQTO generates selectively quantized models, deploys them across hardware accelerators, evaluates performance on metrics such as accuracy and size, applies Pareto Front-based objective minimization to identify optimal candidates, and provides visualization of results. We evaluated SeQTO on four ONNX models under two quantization settings across CPU and GPU devices. Our results show that SeQTO effectively identifies high-quality selectively quantized models, achieving up to $54.14$% lower accuracy loss while maintaining up to $98.18$% of size reduction compared to fully quantized models.

This program is tentative and subject to change.

Thu 16 Apr

Displayed time zone: Brasilia, Distrito Federal, Brazil change

11:00 - 12:30
11:00
15m
Talk
NeMo: A Neuron-level Modularizing-While-Training Approach for Decomposing DNN Models
Journal-first Papers
Xiaohan Bi Beihang University, Binhang Qi National University of Singapore, Hailong Sun Beihang University, Xiang Gao Beihang University, Yue Yu PengCheng Lab, Xiaojun Liang PengCheng Lab
11:15
15m
Talk
A Selective Quantization Tuner for ONNX Models
New Ideas and Emerging Results (NIER)
Nikolaos Louloudakis The University of Edinburgh, Ajitha Rajan The University of Edinburgh
11:30
15m
Paper
Green LLM Techniques in Action: How Effective Are Existing Techniques for Improving the Energy Efficiency of LLM-Based Applications in Industry?
SE In Practice (SEIP)
Pelin Rabia Kuran Vrije Universiteit Amsterdam, Rumbidzai Chitakunye Vrije Universiteit Amsterdam, Vincenzo Stoico Vrije Universiteit Amsterdam, Ilja Heitlager Schuberg Philis, Justus Bogner Vrije Universiteit Amsterdam
DOI Pre-print
11:45
15m
Talk
DNN Modularization via Activation-Driven Training
Research Track
Tuan Ngo University of Southern California, Abid Hassan University of Southern California, Saad Shafiq University of Southern California, Nenad Medvidović University of Southern California
12:00
15m
Talk
ModularEvo: Evolving Multi-Task Models via Neural Network Modularization and Composition
Research Track
Wenrui Long Beihang university, Binhang Qi Beihang University, Hailong Sun Beihang University, ZongZhen Yang Beihang University, Ruobing Zhao Beihang University, Xiang Gao Beihang University
12:15
15m
Talk
The Hidden Cost of Readability: How Code Formatting Silently Consumes Your LLM Budget
Research Track
Dangfeng Pan Monash University, Zhensu Sun Singapore Management University, cenyuan zhang Monash University, David Lo Singapore Management University, Xiaoning Du Monash University