Beyond Answer Engines: LLMs as Reasoning Partners in Data Structures and Algorithms Education
This program is tentative and subject to change.
Large Language Models (LLMs) are rapidly entering software engineering education. As students turn to them for Data Structures and Algorithms (DSA) tutoring, a critical question emerges: do they foster genuine problem-solving skills or merely supply polished answers? Moreover, their competence in DSA remains underexplored beyond benchmarks likely contaminated with training data. We evaluate four LLMs—GPT-4o, Claude 3.7 Sonnet, Llama 3.2, and DeepSeek R1—on 80 LeetCode Top Interview 150 problems and 100 recent Weekly Contest problems, assessing code correctness, maintainability, and explanatory reasoning. Results show that while Claude and DeepSeek perform best on harder problems, all models’ average success rates drop by 49% on novel problems, revealing a critical gap in translating theoretical insights into effective implementations. On the novel problems, DeepSeek attained the highest success rate (70%) with fewer prompt turns but produced the least maintainable code, whereas Claude solved fewer (50%) yet generated the most maintainable solutions, highlighting a trade-off between correctness and pedagogical value. These findings suggest that LLMs are capable reasoning partners yet insufficient for autonomous problem-solving. Thus, educators must carefully integrate these tools into curricula, emphasizing students’ critical reasoning and debugging skills, and develop assessments that either leverage or withstand LLM assistance. Our work contributes a rigorous, multidimensional evaluation framework and practical recommendations for AI-assisted learning in software engineering (SE) and computer science (CS) education, underscoring critical evaluation over reliance on automated solutions.
This program is tentative and subject to change.
Fri 17 AprDisplayed time zone: Brasilia, Distrito Federal, Brazil change
14:00 - 15:30 | Education 8Software Engineering Education and Training (SEET) at Oceania VI Chair(s): Fabio Kon University of São Paulo | ||
14:00 15mTalk | Beyond Answer Engines: LLMs as Reasoning Partners in Data Structures and Algorithms Education Software Engineering Education and Training (SEET) Saad Zafar Khan University of Calgary, Desiree Leal University of Calgary, Lucas Valença University of Calgary, Ahmad Abdellatif University of Calgary, Mea Wang University of Calgary, Diwakar Krishnamurthy University of Calgary, Ronnie de Souza Santos University of Calgary | ||
14:15 15mTalk | The Boundary-Spanning Assistant: Understanding the Role and Usage patterns of LLMs in Project-Based Software Engineering Software Engineering Education and Training (SEET) Anh Nguyen-Duc University of South Eastern Norway, Kai-Kristian Kemell Tampere University, Aparna Chirumamilla NTNU | ||
14:30 15mTalk | An Experience Report on a Pedagogically Controlled, Curriculum-Constrained AI Tutor for SE Education Software Engineering Education and Training (SEET) Lucia Happe Karlsruhe Institute of Technology, Dominik Fuchß Karlsruhe Institute of Technology (KIT), Luca Hüttner Karlsruhe Institute of Technology (KIT), Kai Marquardt Karlsruhe Institute of Technology (KIT), Anne Koziolek Karlsruhe Institute of Technology DOI Pre-print | ||
14:45 15mTalk | Enhancing Debugging Skills With AI-Powered Assistance: A Real-Time Tool for Debugging Support Software Engineering Education and Training (SEET) Elizaveta Artser JetBrains Research, Daniil Karol Researcher at Education Research at JetBrains Research, Anna Potriasaeva JetBrains Research, Aleksei Rostovskii JetBrains Research, Katsiaryna Dzialets JetBrains, Ekaterina Koshchenko JetBrains Research, Xiaotian Su ETH Zurich, April Wang ETH Zürich, Anastasiia Birillo JetBrains Research | ||
15:00 15mTalk | Learning to Program Alongside AI: Critical Thinking, AI Ethics, and Gendered Patterns of German Secondary School Students Software Engineering Education and Training (SEET) Isabella Graßl Technical University of Darmstadt | ||