Cargando…
Deep Statistical Model Checking
Neural networks (NN) are taking over ever more decisions thus far taken by humans, even though verifiable system-level guarantees are far out of reach. Neither is the verification technology available, nor is it even understood what a formal, meaningful, extensible, and scalable testbed might look l...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7281856/ http://dx.doi.org/10.1007/978-3-030-50086-3_6 |
_version_ | 1783544013019676672 |
---|---|
author | Gros, Timo P. Hermanns, Holger Hoffmann, Jörg Klauck, Michaela Steinmetz, Marcel |
author_facet | Gros, Timo P. Hermanns, Holger Hoffmann, Jörg Klauck, Michaela Steinmetz, Marcel |
author_sort | Gros, Timo P. |
collection | PubMed |
description | Neural networks (NN) are taking over ever more decisions thus far taken by humans, even though verifiable system-level guarantees are far out of reach. Neither is the verification technology available, nor is it even understood what a formal, meaningful, extensible, and scalable testbed might look like for such a technology. The present paper is a modest attempt to improve on both the above aspects. We present a family of formal models that contain basic features of automated decision making contexts and which can be extended with further orthogonal features, ultimately encompassing the scope of autonomous driving. Due to the possibility to model random noise in the decision actuation, each model instance induces a Markov decision process (MDP) as verification object. The NN in this context has the duty to actuate (near-optimal) decisions. From the verification perspective, the externally learnt NN serves as a determinizer of the MDP, the result being a Markov chain which as such is amenable to statistical model checking. The combination of a MDP and a NN encoding the action policy is central to what we call “deep statistical model checking” (DSMC). While being a straightforward extension of statistical model checking, it enables to gain deep insight into questions like “how high is the NN-induced safety risk?”, “how good is the NN compared to the optimal policy?” (obtained by model checking the MDP), or “does further training improve the NN?”. We report on an implementation of DSMC inside The Modest Toolset in combination with externally learnt NNs, demonstrating the potential of DSMC on various instances of the model family. |
format | Online Article Text |
id | pubmed-7281856 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72818562020-06-09 Deep Statistical Model Checking Gros, Timo P. Hermanns, Holger Hoffmann, Jörg Klauck, Michaela Steinmetz, Marcel Formal Techniques for Distributed Objects, Components, and Systems Article Neural networks (NN) are taking over ever more decisions thus far taken by humans, even though verifiable system-level guarantees are far out of reach. Neither is the verification technology available, nor is it even understood what a formal, meaningful, extensible, and scalable testbed might look like for such a technology. The present paper is a modest attempt to improve on both the above aspects. We present a family of formal models that contain basic features of automated decision making contexts and which can be extended with further orthogonal features, ultimately encompassing the scope of autonomous driving. Due to the possibility to model random noise in the decision actuation, each model instance induces a Markov decision process (MDP) as verification object. The NN in this context has the duty to actuate (near-optimal) decisions. From the verification perspective, the externally learnt NN serves as a determinizer of the MDP, the result being a Markov chain which as such is amenable to statistical model checking. The combination of a MDP and a NN encoding the action policy is central to what we call “deep statistical model checking” (DSMC). While being a straightforward extension of statistical model checking, it enables to gain deep insight into questions like “how high is the NN-induced safety risk?”, “how good is the NN compared to the optimal policy?” (obtained by model checking the MDP), or “does further training improve the NN?”. We report on an implementation of DSMC inside The Modest Toolset in combination with externally learnt NNs, demonstrating the potential of DSMC on various instances of the model family. 2020-05-13 /pmc/articles/PMC7281856/ http://dx.doi.org/10.1007/978-3-030-50086-3_6 Text en © IFIP International Federation for Information Processing 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Gros, Timo P. Hermanns, Holger Hoffmann, Jörg Klauck, Michaela Steinmetz, Marcel Deep Statistical Model Checking |
title | Deep Statistical Model Checking |
title_full | Deep Statistical Model Checking |
title_fullStr | Deep Statistical Model Checking |
title_full_unstemmed | Deep Statistical Model Checking |
title_short | Deep Statistical Model Checking |
title_sort | deep statistical model checking |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7281856/ http://dx.doi.org/10.1007/978-3-030-50086-3_6 |
work_keys_str_mv | AT grostimop deepstatisticalmodelchecking AT hermannsholger deepstatisticalmodelchecking AT hoffmannjorg deepstatisticalmodelchecking AT klauckmichaela deepstatisticalmodelchecking AT steinmetzmarcel deepstatisticalmodelchecking |