Write a Blog >>
Thu 21 Jan 2016 15:10 - 15:35 at Grand Bay North - Track 1: Learning and verification Chair(s): David Monniaux

Inductive invariants can be robustly synthesized using a learning model where the teacher is a program verifier who instructs the learner through concrete program configurations, classified as positive, negative, and implications. We propose the first learning algorithms in this model with implication counterexamples that are based on machine learning techniques. We implement the learners and an appropriate teacher, and show that the resulting invariant synthesis is efficient and convergent for a large suite of programs.

Thu 21 Jan

Displayed time zone: Guadalajara, Mexico City, Monterrey change

14:20 - 16:00
Track 1: Learning and verificationResearch Papers at Grand Bay North
Chair(s): David Monniaux CNRS, VERIMAG
14:20
25m
Talk
Combining Static Analysis with Probabilistic Models to Enable Market-Scale Analysis
Research Papers
Damien Octeau University of Wisconsin and Pennsylvania State University, Somesh Jha University of Wisconsin, Madison, Matthew Dering Pennsylvania State University, Patrick McDaniel Pennsylvania State University, Alexandre Bartel Technical University Darmstadt, Hongyu Li Rice University, Jacques Klein University of Luxembourg, Yves Le Traon University of Luxembourg
Media Attached
14:45
25m
Talk
Abstraction Refinement Guided by a Learnt Probabilistic Model
Research Papers
Radu Grigore University of Oxford, Hongseok Yang University of Oxford, UK
Media Attached
15:10
25m
Talk
Learning Invariants using Decision Trees and Implication Counterexamples
Research Papers
Pranav Garg University of Illinois at Urbana-Champaign, Daniel Neider University of Illinois at Urbana-Champaign, P. Madhusudan University of Illinois at Urbana-Champaign, Dan Roth University of Illinois at Urbana-Champaign
Media Attached
15:35
25m
Talk
Symbolic Abstract Data Type Inference
Research Papers
Michael Emmi IMDEA Software Institute, Constantin Enea LIAFA, Université Paris Diderot
Media Attached