Optimizing decision making in concolic execution using reinforcement learning
This paper presents an improvement to a new opensource testing tool capable of performing concolic execution on x86 binaries. The novelty is to use a reinforcement learning solution that reduces the number of symbolically executed states. It does so by learning a set of models that predict how efficiently it would be to change the conditions at various branch points. Thus, we first reinterpret the state-of-the-art concolic execution algorithm as a typical reinforcement learning environment, then we build estimation models used to prune states that do not look promising. The architecture of the base model is a Deep Q-Network used inside an LSTM that captures the patterns from the ordered set of branch points (path) resulted by executing the application under test with different inputs generated at runtime (experiments). Various reward functions can give automatic feedback from the concolic execution environment to define different policies. These are customizable in our open-source implementation, such that users can define their custom test targets.
Sat 24 Oct Times are displayed in time zone: Greenwich Mean Time : Lisbon change
|16:00 - 16:30|
Leen LambersHasso-Plattner-Institut, Universität Potsdam , Sven SchneiderHasso-Plattner-Institut, Universität Potsdam , Marcel WeisgutHasso-Plattner-Institut, Universität PotsdamLink to publication DOI
|16:30 - 17:00|
Ciprian PaduraruUniversity of Bucharest , Alin StefanescuUniversity of Bucharest , Miruna Gabriela PaduraruUniversity of BucharestLink to publication DOI
|17:00 - 17:30|