Cargando…

Deep Statistical Model Checking

Neural networks (NN) are taking over ever more decisions thus far taken by humans, even though verifiable system-level guarantees are far out of reach. Neither is the verification technology available, nor is it even understood what a formal, meaningful, extensible, and scalable testbed might look l...

Descripción completa

Detalles Bibliográficos
Autores principales: Gros, Timo P., Hermanns, Holger, Hoffmann, Jörg, Klauck, Michaela, Steinmetz, Marcel
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7281856/
http://dx.doi.org/10.1007/978-3-030-50086-3_6
_version_ 1783544013019676672
author Gros, Timo P.
Hermanns, Holger
Hoffmann, Jörg
Klauck, Michaela
Steinmetz, Marcel
author_facet Gros, Timo P.
Hermanns, Holger
Hoffmann, Jörg
Klauck, Michaela
Steinmetz, Marcel
author_sort Gros, Timo P.
collection PubMed
description Neural networks (NN) are taking over ever more decisions thus far taken by humans, even though verifiable system-level guarantees are far out of reach. Neither is the verification technology available, nor is it even understood what a formal, meaningful, extensible, and scalable testbed might look like for such a technology. The present paper is a modest attempt to improve on both the above aspects. We present a family of formal models that contain basic features of automated decision making contexts and which can be extended with further orthogonal features, ultimately encompassing the scope of autonomous driving. Due to the possibility to model random noise in the decision actuation, each model instance induces a Markov decision process (MDP) as verification object. The NN in this context has the duty to actuate (near-optimal) decisions. From the verification perspective, the externally learnt NN serves as a determinizer of the MDP, the result being a Markov chain which as such is amenable to statistical model checking. The combination of a MDP and a NN encoding the action policy is central to what we call “deep statistical model checking” (DSMC). While being a straightforward extension of statistical model checking, it enables to gain deep insight into questions like “how high is the NN-induced safety risk?”, “how good is the NN compared to the optimal policy?” (obtained by model checking the MDP), or “does further training improve the NN?”. We report on an implementation of DSMC inside The Modest Toolset in combination with externally learnt NNs, demonstrating the potential of DSMC on various instances of the model family.
format Online
Article
Text
id pubmed-7281856
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-72818562020-06-09 Deep Statistical Model Checking Gros, Timo P. Hermanns, Holger Hoffmann, Jörg Klauck, Michaela Steinmetz, Marcel Formal Techniques for Distributed Objects, Components, and Systems Article Neural networks (NN) are taking over ever more decisions thus far taken by humans, even though verifiable system-level guarantees are far out of reach. Neither is the verification technology available, nor is it even understood what a formal, meaningful, extensible, and scalable testbed might look like for such a technology. The present paper is a modest attempt to improve on both the above aspects. We present a family of formal models that contain basic features of automated decision making contexts and which can be extended with further orthogonal features, ultimately encompassing the scope of autonomous driving. Due to the possibility to model random noise in the decision actuation, each model instance induces a Markov decision process (MDP) as verification object. The NN in this context has the duty to actuate (near-optimal) decisions. From the verification perspective, the externally learnt NN serves as a determinizer of the MDP, the result being a Markov chain which as such is amenable to statistical model checking. The combination of a MDP and a NN encoding the action policy is central to what we call “deep statistical model checking” (DSMC). While being a straightforward extension of statistical model checking, it enables to gain deep insight into questions like “how high is the NN-induced safety risk?”, “how good is the NN compared to the optimal policy?” (obtained by model checking the MDP), or “does further training improve the NN?”. We report on an implementation of DSMC inside The Modest Toolset in combination with externally learnt NNs, demonstrating the potential of DSMC on various instances of the model family. 2020-05-13 /pmc/articles/PMC7281856/ http://dx.doi.org/10.1007/978-3-030-50086-3_6 Text en © IFIP International Federation for Information Processing 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Gros, Timo P.
Hermanns, Holger
Hoffmann, Jörg
Klauck, Michaela
Steinmetz, Marcel
Deep Statistical Model Checking
title Deep Statistical Model Checking
title_full Deep Statistical Model Checking
title_fullStr Deep Statistical Model Checking
title_full_unstemmed Deep Statistical Model Checking
title_short Deep Statistical Model Checking
title_sort deep statistical model checking
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7281856/
http://dx.doi.org/10.1007/978-3-030-50086-3_6
work_keys_str_mv AT grostimop deepstatisticalmodelchecking
AT hermannsholger deepstatisticalmodelchecking
AT hoffmannjorg deepstatisticalmodelchecking
AT klauckmichaela deepstatisticalmodelchecking
AT steinmetzmarcel deepstatisticalmodelchecking