Bridging Operator Semantic Inconsistencies: A Source-level Cross-framework Model Conversion Approach
Driven by the widespread use of deep learning (DL) frameworks, cross-framework model conversion supports a diverse ecosystem and facilitates efficient application development. However, interestingly, existing works commonly use intermediate representations (e.g., computation graphs or APIs) for cross-framework model conversion. Due to the lack of the ability to capture operator execution details compared to source code, it is difficult for these intermediate representations to bridge semantic inconsistencies in operators. These inconsistencies change operator behavior and potentially cause poor performance and errors in converted models.
Thus, we present the first comprehensive study to analyze features and dependencies of source code related to operator semantic inconsistencies, finding that 47% of sampled operators exhibit semantic inconsistencies across DL frameworks, and the related code snippets are distributed across layers of DL frameworks without inter-layer dependencies. These findings suggest that bridging operator semantic inconsistencies layer-by-layer is feasible. Based on the findings, we propose ModelX: a source-level cross-framework model conversion approach that bridges operator semantic inconsistencies by dynamically aligning related code snippets independently within each layer, focusing on model conversion at a much finer-grained level (i.e., source code) instead of intermediate representations. The large-scale experiments on the conversion from PyTorch to Paddle show that ModelX equivalently converts 624 out of 686 sampled PyTorch operators, and yields better performance over two state-of-the-art model conversion approaches and popular large language models. Furthermore, we achieve minimal metric gaps (avg. all under 3.4%) across 52 models in the three most popular application fields (i.e., vision, text, and audio), showing that ModelX is highly robust.
Tue 24 JunDisplayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
14:00 - 15:30 | LLM for SE 2Research Papers / Industry Papers / Ideas, Visions and Reflections at Cosmos Hall Chair(s): Jialun Cao Hong Kong University of Science and Technology | ||
14:00 20mTalk | Migrating Code At Scale With LLMs At Google Industry Papers Celal Ziftci Google, Stoyan Nikolov Google, Inc., Anna Sjovall Google, Inc., Bo Kim Google, Daniele Codecasa Google, Inc., Max Kim Google | ||
14:20 20mTalk | Integrating Large Language Models and Reinforcement Learning for Non-Linear Reasoning Research Papers DOI | ||
14:40 20mTalk | Smaller but Better: Self-Paced Knowledge Distillation for Lightweight yet Effective LCMs Research Papers Yujia Chen Harbin Institute of Technology, Shenzhen, Yang Ye Huawei Cloud Computing Technologies Co., Ltd., Zhongqi Li Huawei Cloud Computing Technologies Co., Ltd., Yuchi Ma Huawei Cloud Computing Technologies, Cuiyun Gao Harbin Institute of Technology, Shenzhen DOI | ||
15:00 10mTalk | Enabling Scalable Proactive Workspaces With Environment-Wide Context Ideas, Visions and Reflections Nick Bradley University of British Columbia, Thomas Fritz University of Zurich, Reid Holmes University of British Columbia | ||
15:10 20mTalk | Bridging Operator Semantic Inconsistencies: A Source-level Cross-framework Model Conversion Approach Research Papers Xingpei Li National University of Defense Technology, China, Yan Lei Chongqing University, Zhouyang Jia National University of Defense Technology, Yuanliang Zhang National University of Defense Technology, Haoran Liu National University of Defense Technology, Liqian Chen National University of Defense Technology, Wei Dong National University of Defense Technology, Shanshan Li National University of Defense Technology DOI |
This is the main event hall of Clarion Hotel, which will be used to host keynote talks and other plenary sessions. The FSE and ISSTA banquets will also happen in this room.
The room is just in front of the registration desk, on the other side of the main conference area. The large doors with numbers “1” and “2” provide access to the Cosmos Hall.