SGCR: A Specification-Grounded Framework for Trustworthy LLM Code Review
This program is tentative and subject to change.
Automating code review with Large Language Models (LLMs) shows immense promise, yet practical adoption is hampered by their lack of reliability, context-awareness, and control. To address this, we propose Specification-Grounded Code Review (SGCR), a framework that grounds LLMs in human-authored specifications to produce trustworthy and relevant feedback. SGCR features a novel dual-pathway architecture: an explicit path ensures deterministic compliance with predefined rules derived from these specifications, while an implicit path heuristically discovers and verifies issues beyond those rules. Deployed in a live industrial environment at HiThink Research, SGCR’s suggestions achieved a 42% developer adoption rate—a 90.9% relative improvement over a baseline LLM (22%). Our work demonstrates that specification-grounding is a powerful paradigm for bridging the gap between the generative power of LLMs and the rigorous reliability demands of software engineering.
This program is tentative and subject to change.
Mon 17 NovDisplayed time zone: Seoul change
16:00 - 17:00 | |||
16:00 10mTalk | SGCR: A Specification-Grounded Framework for Trustworthy LLM Code Review Industry Showcase Kai Wang HiThink Research, Bingcheng Mao HiThink Research, Shuai Jia HiThink Research, Yujie Ding HiThink Research, Dongming Han HiThink Research, Tianyi Ma HiThink Research, Bin Cao Zhejiang University of Technology | ||
16:10 10mTalk | What Types of Code Review Comments Do Developers Most Frequently Resolve? Industry Showcase Saul Goldman The University of Melbourne, Hong Yi Lin The University of Melbourne, Jirat Pasuksmit Atlassian, Patanamon Thongtanunam University of Melbourne, Kla Tantithamthavorn Monash University and Atlassian, Zhe Wang Institute of Computing Technology at Chinese Academy of Sciences; Zhongguancun Laboratory, Ruixiong Zhang Atlassian, Ali Behnaz Atlassian, Fan Jiang Atlassian, Michael Siers Atlassian, Ryan Jiang Atlassian, Mike Buller Atlassian, Minwoo Jeong Atlassian, Ming Wu Atlassian | ||
16:20 10mTalk | Vessel: A Taxonomy of Reproducibility Issues for Container Images NIER Track Kevin Pitstick Carnegie Mellon Software Engineering Institute, Alex Derr Carnegie Mellon Software Engineering Institute, Lihan Zhan Carnegie Mellon Software Engineering Institute, Sebastian Echeverria Carnegie Mellon Software Engineering Institute | ||
16:30 10mTalk | From Modules to Marketplaces: A Vision for Composable Capability Sharing Across Organizations NIER Track Wei-Ji Wang National Taiwan University & Chunghwa Telecom Laboratories | ||
16:40 10mTalk | Towards Automated Governance: A DSL for Human-Agent Collaboration in Software Projects NIER Track Adem Ait University of Luxembourg, Gwendal Jouneaux Luxembourg Institute of Science and Technology, Javier Luis Cánovas Izquierdo Universitat Oberta de Catalunya, Jordi Cabot Luxembourg Institute of Science and Technology Pre-print | ||
16:50 10mTalk | iCodeReviewer: Improving Secure Code Review with Mixture of Prompts Industry Showcase | ||