Cargando…

MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty

Increasingly, safety-critical systems include artificial intelligence and machine learning components (i.e., learning-enabled components (LECs)). However, when behavior is learned in a training environment that fails to fully capture real-world phenomena, the response of an LEC to untrained phenomen...

Descripción completa

Detalles Bibliográficos
Autores principales: Langford, Michael Austin, Chan, Kenneth H., Fleck, Jonathon Emil, McKinley, Philip K., Cheng, Betty H. C.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer Berlin Heidelberg 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10024308/
https://www.ncbi.nlm.nih.gov/pubmed/37363107
http://dx.doi.org/10.1007/s10270-023-01090-9
_version_ 1784909074335268864
author Langford, Michael Austin
Chan, Kenneth H.
Fleck, Jonathon Emil
McKinley, Philip K.
Cheng, Betty H. C.
author_facet Langford, Michael Austin
Chan, Kenneth H.
Fleck, Jonathon Emil
McKinley, Philip K.
Cheng, Betty H. C.
author_sort Langford, Michael Austin
collection PubMed
description Increasingly, safety-critical systems include artificial intelligence and machine learning components (i.e., learning-enabled components (LECs)). However, when behavior is learned in a training environment that fails to fully capture real-world phenomena, the response of an LEC to untrained phenomena is uncertain and therefore cannot be assured as safe. Automated methods are needed for self-assessment and adaptation to decide when learned behavior can be trusted. This work introduces a model-driven approach to manage self-adaptation of a learning-enabled system (LES) to account for run-time contexts for which the learned behavior of LECs cannot be trusted. The resulting framework enables an LES to monitor and evaluate goal models at run time to determine whether or not LECs can be expected to meet functional objectives and enables system adaptation accordingly. Using this framework enables stakeholders to have more confidence that LECs are used only in contexts comparable to those validated at design time.
format Online
Article
Text
id pubmed-10024308
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Springer Berlin Heidelberg
record_format MEDLINE/PubMed
spelling pubmed-100243082023-03-21 MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty Langford, Michael Austin Chan, Kenneth H. Fleck, Jonathon Emil McKinley, Philip K. Cheng, Betty H. C. Softw Syst Model Special Section Paper Increasingly, safety-critical systems include artificial intelligence and machine learning components (i.e., learning-enabled components (LECs)). However, when behavior is learned in a training environment that fails to fully capture real-world phenomena, the response of an LEC to untrained phenomena is uncertain and therefore cannot be assured as safe. Automated methods are needed for self-assessment and adaptation to decide when learned behavior can be trusted. This work introduces a model-driven approach to manage self-adaptation of a learning-enabled system (LES) to account for run-time contexts for which the learned behavior of LECs cannot be trusted. The resulting framework enables an LES to monitor and evaluate goal models at run time to determine whether or not LECs can be expected to meet functional objectives and enables system adaptation accordingly. Using this framework enables stakeholders to have more confidence that LECs are used only in contexts comparable to those validated at design time. Springer Berlin Heidelberg 2023-03-18 /pmc/articles/PMC10024308/ /pubmed/37363107 http://dx.doi.org/10.1007/s10270-023-01090-9 Text en © The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2023, Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Special Section Paper
Langford, Michael Austin
Chan, Kenneth H.
Fleck, Jonathon Emil
McKinley, Philip K.
Cheng, Betty H. C.
MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
title MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
title_full MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
title_fullStr MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
title_full_unstemmed MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
title_short MoDALAS: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
title_sort modalas: addressing assurance for learning-enabled autonomous systems in the face of uncertainty
topic Special Section Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10024308/
https://www.ncbi.nlm.nih.gov/pubmed/37363107
http://dx.doi.org/10.1007/s10270-023-01090-9
work_keys_str_mv AT langfordmichaelaustin modalasaddressingassuranceforlearningenabledautonomoussystemsinthefaceofuncertainty
AT chankennethh modalasaddressingassuranceforlearningenabledautonomoussystemsinthefaceofuncertainty
AT fleckjonathonemil modalasaddressingassuranceforlearningenabledautonomoussystemsinthefaceofuncertainty
AT mckinleyphilipk modalasaddressingassuranceforlearningenabledautonomoussystemsinthefaceofuncertainty
AT chengbettyhc modalasaddressingassuranceforlearningenabledautonomoussystemsinthefaceofuncertainty