CAIN 2026
Sun 12 - Sat 18 April 2026 Rio de Janeiro, Brazil
co-located with ICSE 2026
Sun 12 Apr 2026 14:00 - 14:12 at Oceania X - Data, Transparency, and XAI Chair(s): Grace Lewis

Data annotation is a critical yet error-prone activity in developing AI-enabled perception systems (AIePS) for automated driving. Annotation quality directly affects AI model development performance, safety and reliability. Yet, there is little empirically grounded understanding of how annotation errors arise and propagate across the multi-organisational automotive supply chain. This study investigates the types, causes, and effects of data annotation errors through a multi-organisation case study involving six companies (e.g., technology and AI companies that develop software for ad- advanced driver assistance systems and self-driving cars) and four research institutes in Europe and the UK. We conducted 19 semi-structured interviews with 20 experts (≈ 50 hours of transcripts) and applied a six-phase thematic analysis. The resulting data annotation errors taxonomy identifies 18 recurring error types across three major data-quality dimensions: completeness ( attribute omission, missing feedback loop, privacy/compliance omission, edge-case omission, selection bias, sensor synchronisation issues), accuracy (wrong class label, bounding-box errors, granularity mismatch, insufficient guidance, bias-driven errors), and consistency (inter-annotator disagreement, ambiguous instructions, lack of purpose knowledge, misaligned hand-offs, limited review and logging, lack of frameworks and standards, cross-modality misalignment). For re- search rigour and triangulation, we further validated the proposed taxonomy of data annotation errors with industry practitioners, confirming its usefulness for data annotation errors root-cause analysis, supplier quality reviews, new projects onboarding, and data annotation guideline optimisation. Practitioners described it as a “failure-mode catalogue” comparable to FMEA (Failure Mode and Effects Analysis). By framing annotation quality as an AI-enabled system development lifecycle and supply-chain concern, this work advances SE4AI by providing a shared vocabulary, diagnostic check-list, and actionable guidance for trustworthy AIePS development.

Sun 12 Apr

Displayed time zone: Brasilia, Distrito Federal, Brazil change

14:00 - 15:30
Data, Transparency, and XAICAIN Program / Research Track / Industry Track at Oceania X
Chair(s): Grace Lewis Carnegie Mellon Software Engineering Institute
14:00
12m
Full-paper
Data Annotation Errors in AI-Enabled Perception System Development: A Multi-Organisation Case Study in the Automotive DomainFull Paper
Research Track
Hina Saeeda Chalmers University Sweden, Eric Knauss Chalmers | University of Gothenburg, Mazen Mohamad Chalmers | RISE - Research Institutes of Sweden, Tommy Johansson Kognic AB Sweden
14:12
8m
Short-paper
Model-Driven Engineering of Synthetic Data Pipelines for AI-Enabled Healthcare SystemsShort Paper
Research Track
Mukhtar Sani CEA List, France, Nicholas Matragkas Université Paris-Saclay, CEA, List., Nam-khanh Nguyen DILS/LSEA CEA LIST Palaiseau, France
14:20
12m
Full-paper
Data Leakage in Automotive Perception: Practitioners' InsightsFull Paper
Industry Track
Md Abu Ahammed Babu Volvo Cars AB, Sushant Kumar Pandey University of Groningen, The Netherlands, Darko Durisic , András Bálint , Miroslaw Staron Chalmers University of Technology and University of Gothenburg
Pre-print
14:32
12m
Full-paper
AIBoMGen: Generating an AI Bill of Materials for Secure, Transparent, and Compliant Model TrainingFull Paper
Research Track
Wiebe Vandendriessche Ghent University, imec, Jordi Thijsman Ghent University, imec, Laurens D'hooge Ghent University, imec, Bruno Volckaert Ghent University, imec, Merlijn Sebrechts Ghent University, imec
Pre-print
14:44
8m
Short-paper
Leveraging Domain Requirements in Concept Based Models via Differentiable Fuzzy LogicShort Paper
Research Track
Eik Reichmann Humboldt-Universität zu Berlin, Joao Paulo Costa de Araujo Humboldt-Universität zu Berlin, Lars Grunske Humboldt-Universität zu Berlin
14:52
12m
Full-paper
Distilling Ensemble Intelligence into Explainable Anomaly Detection ModelsFull Paper
Research Track
Ashish Rauniyar SINTEF Digital, Norway, Erik Johannes Husom SINTEF Digital, Sagar Sen
15:04
26m
Live Q&A
Joint Q&A (Data, Transparency, and XAI)
CAIN Program