Cargando…
Deep Reinforcement Learning and the Type IIA Landscape
<!--HTML-->An artificial intelligence agent known as an asynchronous advantage actor-critic is utilized to explore type IIA compactifications with intersecting D6-branes. By reinforcement learning, the agent's performance in satisfying string consistency conditions, and finding Standard M...
Autor principal: | |
---|---|
Lenguaje: | eng |
Publicado: |
2019
|
Materias: | |
Acceso en línea: | http://cds.cern.ch/record/2682616 |
_version_ | 1780963173821579264 |
---|---|
author | Nelson, Brent |
author_facet | Nelson, Brent |
author_sort | Nelson, Brent |
collection | CERN |
description | <!--HTML-->An artificial intelligence agent known as an asynchronous advantage actor-critic is utilized to explore type IIA compactifications with intersecting D6-branes. By reinforcement learning, the agent's performance in satisfying string consistency conditions, and finding Standard Model like configurations, is significantly improved. In one case, we demonstrate that the agent learns a human-derived strategy for finding consistent string models. In another case, where no human-derived strategy exists, the agent learns a genuinely new strategy that achieves the same goal twice as efficiently per unit time. |
id | cern-2682616 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2019 |
record_format | invenio |
spelling | cern-26826162022-11-02T22:21:37Zhttp://cds.cern.ch/record/2682616engNelson, BrentDeep Reinforcement Learning and the Type IIA LandscapeString Phenomenology 2019Conferences & Workshops<!--HTML-->An artificial intelligence agent known as an asynchronous advantage actor-critic is utilized to explore type IIA compactifications with intersecting D6-branes. By reinforcement learning, the agent's performance in satisfying string consistency conditions, and finding Standard Model like configurations, is significantly improved. In one case, we demonstrate that the agent learns a human-derived strategy for finding consistent string models. In another case, where no human-derived strategy exists, the agent learns a genuinely new strategy that achieves the same goal twice as efficiently per unit time.oai:cds.cern.ch:26826162019 |
spellingShingle | Conferences & Workshops Nelson, Brent Deep Reinforcement Learning and the Type IIA Landscape |
title | Deep Reinforcement Learning and the Type IIA Landscape |
title_full | Deep Reinforcement Learning and the Type IIA Landscape |
title_fullStr | Deep Reinforcement Learning and the Type IIA Landscape |
title_full_unstemmed | Deep Reinforcement Learning and the Type IIA Landscape |
title_short | Deep Reinforcement Learning and the Type IIA Landscape |
title_sort | deep reinforcement learning and the type iia landscape |
topic | Conferences & Workshops |
url | http://cds.cern.ch/record/2682616 |
work_keys_str_mv | AT nelsonbrent deepreinforcementlearningandthetypeiialandscape AT nelsonbrent stringphenomenology2019 |