ICSE 2026
Sun 12 - Sat 18 April 2026 Rio de Janeiro, Brazil

This program is tentative and subject to change.

Fri 17 Apr 2026 14:15 - 14:30 at Oceania V - Human and Social Aspects 12 Chair(s): Alexander Serebrenik

Aim. The code review team at Meta is continuously improving the code review process. In this work, we report on three randomized controlled experimental trials to improve code reviewer recommendation.

Method. To evaluate the recommenders, we conduct three A/B tests which are a type of randomized controlled experimental trial. The unit is either the code diff (Meta’s term for a pull-request) or all the diffs that an author creates during the experimental period. We set goal metrics, i.e. those we expect to improve, and guardrail metrics, those that we do not want to negatively impact, i.e. analogous to safety metrics in medical trials. We test the outcomes using a t-test, Wilcoxon test, or Fisher test depending on the type of data.

Expt 1. We developed a new recommender, RevRecV2, based on features that had been successfully used in the literature and that could be calculated with low latency. In an A/B test on 82k diffs in Spring of 2022, we found that the new recommender was more accurate and had lower latency. The new recommender did not impact the amount of time a diff was under review. The results allowed us to roll-out the recommender in Summer of 2022 to all of Meta.

Expt 2. Reviewer workload is not evenly distributed, our goal was to reduce the workload of top reviewers. Based on the literature, and using historical data, we conducted backtests to determine the best measure of reviewer workload. We then ran an A/B test on 28k diff authors in Winter 2023 on a workload balanced recommender, RevRecWL. Our A/B test led to mixed results. When a low workload reviewer had reasonable expertise, authors selected them, however, the top recommended low workload reviewer was often not selected. There was no impact on our guardrail metrics of the amount of time to perform a review. This workload balancing replaced the recommender from the first experiment as the recommender in production at Meta.

Expt 3. Engineers at Meta often select a team rather than an individual reviewer to review a diff. We suspected the bystander effect might be slowing down reviews of these diffs because no single individual was assigned the review. On diffs that only had a team assigned, we randomly selected one of the top three recommended reviewers to review the diff with BystanderRecRnd. We conducted an A/B test on 12.5k authors in Spring 2023 and found a large decrease in the amount of time it took for diffs to be reviewed. We did not find that reviewers rushed reviews. The results were strong enough to roll this recommender out to all diffs that only have a team assigned for review.

Implications. Aside from the direct findings from our work, our findings suggest there can be a discrepancy between historical back-testing and A/B test experimental findings, and that more A/B tests are necessary to test recommenders in production. Outcome measures beyond accuracy are important. This is especially true in understanding how recommenders change a reviewer’s workload. We also see that the latency in displaying a recommendation can have a large impact on how often authors select recommendations making the reporting of latency an important metric for future work.

This program is tentative and subject to change.

Fri 17 Apr

Displayed time zone: Brasilia, Distrito Federal, Brazil change

14:00 - 15:30
Human and Social Aspects 12Research Track / Journal-first Papers at Oceania V
Chair(s): Alexander Serebrenik Eindhoven University of Technology
14:00
15m
Talk
What Could Possibly Go Wrong: Undesirable Patterns in Collective Development
Journal-first Papers
Ekaterina Koshchenko JetBrains Research, Mikhail Evtikhiev JetBrains Research, Vladimir Kovalenko JetBrains Research
14:15
15m
Talk
Improving Code Reviewer Recommendation: Accuracy, Latency, Workload, and Bystanders
Journal-first Papers
Peter C Rigby Meta / Concordia University, Seth Rogers Meta, Sadruddin Saleem Meta Platforms, Inc., Parth Suresh Meta Platforms, Inc., Daniel Suskin Meta Platforms, Inc., Patrick Riggs Meta, Chandra Sekhar Maddila Meta Platforms, Inc., Nachiappan Nagappan Meta Platforms, Inc., Audris Mockus University of Tennessee
14:30
15m
Talk
With Great Power Comes Great Responsibility: The Role of Software Engineers
Journal-first Papers
Stefanie Betz Furtwangen University & LUT University, Birgit Penzenstadler Chalmers Tekniska Högskola and Gothenburg University and Lappenranta University of Technology
14:45
15m
Talk
Staying or Leaving? How Job Satisfaction, Embeddedness and Antecedents Predict Turnover Intentions of Software Professionals
Research Track
Miikka Kuutila Dalhousie University, Paul Ralph Dalhousie University, Huilian Sophie Qiu Northwestern University, Ronnie de Souza Santos University of Calgary, Morakot Choetkiertikul Mahidol University, Thailand, Amin Milani Fard New York Institute of Technology, Rana Alkadhi King Saud University, Xavier Devroey University of Namur, Gregorio Robles Universidad Rey Juan Carlos, Hideaki Hata Shinshu University, Sebastian Baltes Heidelberg University, Hera Arif Dalhousie University, Vladimir Kovalenko JetBrains Research, Shalini Chakraborty University of Bayreuth, Eray Tüzün Bilkent University, Gianisa Adisaputri Dalhousie University
Pre-print
15:00
15m
Talk
“I need to learn better searching tactics for privacy policy laws.” Investigating Software Developers’ Behavior When Using Sources on Privacy Issues
Research Track
Stefan Albert Horstmann Ruhr University Bochum, Sandy Hong Ruhr University Bochum, Maziar Niazian Ruhr University Bochum, Cristiana Santos Utrecht University, Alena Naiakshina University of Bonn
Pre-print
15:15
15m
Talk
What’s in a Software Engineering Job Posting?
Research Track
Marvin Wyrich Saarland University, Lloyd Montgomery University of Hamburg, Germany
Pre-print
Hide past events