Wed 5 Jun 2024 09:00 - 10:00 at Ballroom - Keynote Talk #2 Chair(s): Divya Gopinath

Providing safety guarantees for autonomous systems is difficult as these systems operate in complex environments that require the use of learning-enabled components, such as deep neural networks (DNNs) for visual perception. DNNs are hard to analyze due to their size (they can have thousands or millions of parameters), lack of formal specifications (DNNs are typically learnt from labeled data, in the absence of any formal requirements), and sensitivity to small changes in the environment. We present compositional techniques for the formal verification of safety properties of such autonomous systems. The main idea is to abstract the hard-to-analyze components of the autonomous system, such as DNN-based perception and environmental dynamics, with either probabilistic or worst-case abstractions. This makes the system amenable to formal analysis using off-the-shelf model checking tools, enabling the derivation of specifications for the behavior of the abstracted components such that system safety is guaranteed. The derived specifications can be used as run-time monitors deployed on the DNN outputs. We illustrate the idea in a case study from the autonomous airplane domain.

Wed 5 Jun

Displayed time zone: Pacific Time (US & Canada) change

09:00 - 10:00
Keynote Talk #2NFM 2024 at Ballroom
Chair(s): Divya Gopinath NASA Ames (KBR Inc.)
09:00
60m
Keynote
Formal Verification and Run-time Monitoring for Learning-Enabled Autonomous SystemsKeynote
NFM 2024
Corina S. Pasareanu Carnegie Mellon University Silicon Valley, NASA Ames Research Center