TCSE logo 
 Sigsoft logo
Sustainability badge
Sat 3 May 2025 09:10 - 10:10 at 214 - Opening / Keynote 1 / Paper Session 1 Chair(s): Zijian Wang

AI has rapidly transformed the developer landscape, moving from experimental code completion features to indispensable tools integrated into modern IDEs. The impact is undeniable, with significant portions of code now being generated with AI assistance, accelerating development workflows worldwide.

The progress doesn’t stop at code completion. The next frontier is the development of sophisticated AI agents within the IDE, capable of more complex tasks and deeper collaboration. These agents are intelligent systems designed to understand context, perform multi-step actions, and interact collaboratively with the developer. Building these hybrid human-AI systems presents unique challenges: designing intuitive UX, building auxiliary models, and steering the agent towards effective collaboration with the user.

In this talk, Federico will discuss the evolution of AI copilots within IDEs, from the initial release of GitHub Copilot to the emerging paradigm of AI agents in Cursor. We will explore the critical aspects of building successful hybrid human-AI development environments, sharing insights and lessons learned on tackling the technical challenges involved in creating the next generation of intelligent developer tools.

Federico Cassano is a research scientist at Cursor, where he works on CodeLLM training methodology and infrastructure. His research interests broadly include code generation, distributed training, and reinforcement learning.

Sat 3 May

Displayed time zone: Eastern Time (US & Canada) change

09:00 - 10:30
Opening / Keynote 1 / Paper Session 1LLM4Code at 214
Chair(s): Zijian Wang AWS AI Labs
09:00
10m
Day opening
Opening
LLM4Code
Lingming Zhang University of Illinois at Urbana-Champaign, Prem Devanbu University of California at Davis, Zijian Wang AWS AI Labs
09:10
60m
Keynote
Keynote 1: Building the Hybrid Human-AI Developer: From Code Completion to Agents (zoom talk)
LLM4Code
10:10
10m
Talk
Are Large Language Models Memorizing Bug Benchmarks?
LLM4Code
Daniel Ramos Carnegie Mellon University, Claudia Mamede Carnegie Mellon University, Kush Jain Carnegie Mellon University, Paulo Canelas Carnegie Mellon University, Catarina Gamboa Carnegie Mellon University, Claire Le Goues Carnegie Mellon University
10:20
10m
Talk
RepairBench: Leaderboard of Frontier Models for Program Repair
LLM4Code
André Silva KTH Royal Institute of Technology, Martin Monperrus KTH Royal Institute of Technology
:
:
:
: