Predicting Performance and Accuracy of Mixed-Precision Programs for Precision Tuning
A mixed-precision program is a floating-point program that utilizes different precisions for different operations, providing the opportunity of balancing the trade-off between accuracy and performance. Precision tuning aims to find a mixed-precision version of a program that improves its performance while maintaining a given accuracy. Unfortunately, existing precision tuning approaches are either limited to small-scale programs, or suffer from efficiency issues. In this paper, we propose FPLearner, a novel approach that addresses these limitations. Our insight is to leverage a Machine Learning based technique, Graph Neural Networks, to learn the representation of mixed-precision programs to predict their performance and accuracy. Such prediction models can then be used to accelerate the process of dynamic precision tuning by reducing the number of program runs. We create a dataset of mixed-precision programs from five diverse HPC applications for training our models, which achieve 96.34% F1 score in performance prediction and 97.03% F1 score in accuracy prediction. FPLearner improves the time efficiency of two dynamic precision tuners, Precimonious and HiFPTuner, by an average of 25.54% and up to 61.07% while achieving precision tuning results of comparable or better quality.
Fri 19 AprDisplayed time zone: Lisbon change
16:00 - 17:30 | LLM, NN and other AI technologies 7Software Engineering in Society / Software Engineering in Practice / Research Track / New Ideas and Emerging Results at Grande Auditório Chair(s): Vincent J. Hellendoorn Carnegie Mellon University | ||
16:00 15mTalk | Predicting Performance and Accuracy of Mixed-Precision Programs for Precision Tuning Research Track | ||
16:15 15mTalk | A Synthesis of Green Architectural Tactics for ML-Enabled Systems Software Engineering in Society Heli Järvenpää Vrije Universiteit Amsterdam, Patricia Lago Vrije Universiteit Amsterdam, Justus Bogner Vrije Universiteit Amsterdam, Grace Lewis Carnegie Mellon Software Engineering Institute, Henry Muccini University of L'Aquila, Italy, Ipek Ozkaya Carnegie Mellon University Pre-print | ||
16:30 15mTalk | Greening Large Language Models of Code Software Engineering in Society Jieke Shi Singapore Management University, Zhou Yang Singapore Management University, Hong Jin Kang UCLA, Bowen Xu North Carolina State University, Junda He Singapore Management University, David Lo Singapore Management University Pre-print Media Attached | ||
16:45 15mTalk | Lessons from Building CodeBuddy: A Contextualized AI Coding Assistant Software Engineering in Practice Gustavo Pinto Federal University of Pará (UFPA) and Zup Innovation, Cleidson de Souza Federal University of Pará Belém, João Batista Cordeiro Neto Federal University of Santa Catarina and Zup Innovation, Alberto de Souza Zup Innovation, Tarcísio Gotto Zup Innovation, Edward Monteiro StackSpot | ||
17:00 15mTalk | CodeFuse-13B: A Pretrained Multi-lingual Code Large Language Model Software Engineering in Practice | ||
17:15 7mTalk | Breaking the Silence: the Threats of Using LLMs in Software Engineering New Ideas and Emerging Results June Sallou Delft University of Technology, Thomas Durieux TU Delft, Annibale Panichella Delft University of Technology Pre-print |