Cargando…

The Principles of Deep Learning Theory

<!--HTML--><p>Deep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will...

Descripción completa

Detalles Bibliográficos
Autor principal: Roberts, Dan
Lenguaje:eng
Publicado: 2022
Materias:
Acceso en línea:http://cds.cern.ch/record/2809263
_version_ 1780973137468325888
author Roberts, Dan
author_facet Roberts, Dan
author_sort Roberts, Dan
collection CERN
description <!--HTML--><p>Deep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will encompass both understanding the statistics of initialized deep networks and determining the training dynamics of such an ensemble when learning from data. Borrowing from the "effective theory" framework of physics and developing a perturbative 1/n expansion around the limit of infinite hidden-layer width, we will find a principle of sparsity that will let us describe effectively-deep networks of practical large-but-finite-width networks.</p> <p>This talk is based on a book, "The Principles of Deep Learning Theory," co-authored with Sho Yaida and based on research also in collaboration with Boris Hanin. It will be published this month by Cambridge University Press.</p>
id cern-2809263
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2022
record_format invenio
spelling cern-28092632022-11-02T22:05:13Zhttp://cds.cern.ch/record/2809263engRoberts, DanThe Principles of Deep Learning TheoryThe Principles of Deep Learning TheoryTH String Theory Seminar<!--HTML--><p>Deep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will encompass both understanding the statistics of initialized deep networks and determining the training dynamics of such an ensemble when learning from data. Borrowing from the "effective theory" framework of physics and developing a perturbative 1/n expansion around the limit of infinite hidden-layer width, we will find a principle of sparsity that will let us describe effectively-deep networks of practical large-but-finite-width networks.</p> <p>This talk is based on a book, "The Principles of Deep Learning Theory," co-authored with Sho Yaida and based on research also in collaboration with Boris Hanin. It will be published this month by Cambridge University Press.</p>oai:cds.cern.ch:28092632022
spellingShingle TH String Theory Seminar
Roberts, Dan
The Principles of Deep Learning Theory
title The Principles of Deep Learning Theory
title_full The Principles of Deep Learning Theory
title_fullStr The Principles of Deep Learning Theory
title_full_unstemmed The Principles of Deep Learning Theory
title_short The Principles of Deep Learning Theory
title_sort principles of deep learning theory
topic TH String Theory Seminar
url http://cds.cern.ch/record/2809263
work_keys_str_mv AT robertsdan theprinciplesofdeeplearningtheory
AT robertsdan principlesofdeeplearningtheory