ICSE 2026
Sun 12 - Sat 18 April 2026 Rio de Janeiro, Brazil

This program is tentative and subject to change.

Fri 17 Apr 2026 14:00 - 14:15 at Oceania VI - Education 8 Chair(s): Fabio Kon

Large Language Models (LLMs) are rapidly entering software engineering education. As students turn to them for Data Structures and Algorithms (DSA) tutoring, a critical question emerges: do they foster genuine problem-solving skills or merely supply polished answers? Moreover, their competence in DSA remains underexplored beyond benchmarks likely contaminated with training data. We evaluate four LLMs—GPT-4o, Claude 3.7 Sonnet, Llama 3.2, and DeepSeek R1—on 80 LeetCode Top Interview 150 problems and 100 recent Weekly Contest problems, assessing code correctness, maintainability, and explanatory reasoning. Results show that while Claude and DeepSeek perform best on harder problems, all models’ average success rates drop by 49% on novel problems, revealing a critical gap in translating theoretical insights into effective implementations. On the novel problems, DeepSeek attained the highest success rate (70%) with fewer prompt turns but produced the least maintainable code, whereas Claude solved fewer (50%) yet generated the most maintainable solutions, highlighting a trade-off between correctness and pedagogical value. These findings suggest that LLMs are capable reasoning partners yet insufficient for autonomous problem-solving. Thus, educators must carefully integrate these tools into curricula, emphasizing students’ critical reasoning and debugging skills, and develop assessments that either leverage or withstand LLM assistance. Our work contributes a rigorous, multidimensional evaluation framework and practical recommendations for AI-assisted learning in software engineering (SE) and computer science (CS) education, underscoring critical evaluation over reliance on automated solutions.

This program is tentative and subject to change.

Fri 17 Apr

Displayed time zone: Brasilia, Distrito Federal, Brazil change

14:00 - 15:30
Education 8Software Engineering Education and Training (SEET) at Oceania VI
Chair(s): Fabio Kon University of São Paulo
14:00
15m
Talk
Beyond Answer Engines: LLMs as Reasoning Partners in Data Structures and Algorithms Education
Software Engineering Education and Training (SEET)
Saad Zafar Khan University of Calgary, Desiree Leal University of Calgary, Lucas Valença University of Calgary, Ahmad Abdellatif University of Calgary, Mea Wang University of Calgary, Diwakar Krishnamurthy University of Calgary, Ronnie de Souza Santos University of Calgary
14:15
15m
Talk
The Boundary-Spanning Assistant: Understanding the Role and Usage patterns of LLMs in Project-Based Software Engineering
Software Engineering Education and Training (SEET)
Anh Nguyen-Duc University of South Eastern Norway, Kai-Kristian Kemell Tampere University, Aparna Chirumamilla NTNU
14:30
15m
Talk
An Experience Report on a Pedagogically Controlled, Curriculum-Constrained AI Tutor for SE Education
Software Engineering Education and Training (SEET)
Lucia Happe Karlsruhe Institute of Technology, Dominik Fuchß Karlsruhe Institute of Technology (KIT), Luca Hüttner Karlsruhe Institute of Technology (KIT), Kai Marquardt Karlsruhe Institute of Technology (KIT), Anne Koziolek Karlsruhe Institute of Technology
DOI Pre-print
14:45
15m
Talk
Enhancing Debugging Skills With AI-Powered Assistance: A Real-Time Tool for Debugging Support
Software Engineering Education and Training (SEET)
Elizaveta Artser JetBrains Research, Daniil Karol Researcher at Education Research at JetBrains Research, Anna Potriasaeva JetBrains Research, Aleksei Rostovskii JetBrains Research, Katsiaryna Dzialets JetBrains, Ekaterina Koshchenko JetBrains Research, Xiaotian Su ETH Zurich, April Wang ETH Zürich, Anastasiia Birillo JetBrains Research
15:00
15m
Talk
Learning to Program Alongside AI: Critical Thinking, AI Ethics, and Gendered Patterns of German Secondary School Students
Software Engineering Education and Training (SEET)
Isabella Graßl Technical University of Darmstadt
Hide past events