Cargando…
The Principles of Deep Learning Theory
<!--HTML--><p>Deep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will...
Autor principal: | |
---|---|
Lenguaje: | eng |
Publicado: |
2022
|
Materias: | |
Acceso en línea: | http://cds.cern.ch/record/2809263 |
_version_ | 1780973137468325888 |
---|---|
author | Roberts, Dan |
author_facet | Roberts, Dan |
author_sort | Roberts, Dan |
collection | CERN |
description | <!--HTML--><p>Deep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will encompass both understanding the statistics of initialized deep networks and determining the training dynamics of such an ensemble when learning from data. Borrowing from the "effective theory" framework of physics and developing a perturbative 1/n expansion around the limit of infinite hidden-layer width, we will find a principle of sparsity that will let us describe effectively-deep networks of practical large-but-finite-width networks.</p>
<p>This talk is based on a book, "The Principles of Deep Learning Theory," co-authored with Sho Yaida and based on research also in collaboration with Boris Hanin. It will be published this month by Cambridge University Press.</p> |
id | cern-2809263 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2022 |
record_format | invenio |
spelling | cern-28092632022-11-02T22:05:13Zhttp://cds.cern.ch/record/2809263engRoberts, DanThe Principles of Deep Learning TheoryThe Principles of Deep Learning TheoryTH String Theory Seminar<!--HTML--><p>Deep learning is an exciting approach to modern artificial intelligence based on artificial neural networks. The goal of this talk is to provide a blueprint — using tools from physics — for theoretically analyzing deep neural networks of practical relevance. This task will encompass both understanding the statistics of initialized deep networks and determining the training dynamics of such an ensemble when learning from data. Borrowing from the "effective theory" framework of physics and developing a perturbative 1/n expansion around the limit of infinite hidden-layer width, we will find a principle of sparsity that will let us describe effectively-deep networks of practical large-but-finite-width networks.</p> <p>This talk is based on a book, "The Principles of Deep Learning Theory," co-authored with Sho Yaida and based on research also in collaboration with Boris Hanin. It will be published this month by Cambridge University Press.</p>oai:cds.cern.ch:28092632022 |
spellingShingle | TH String Theory Seminar Roberts, Dan The Principles of Deep Learning Theory |
title | The Principles of Deep Learning Theory |
title_full | The Principles of Deep Learning Theory |
title_fullStr | The Principles of Deep Learning Theory |
title_full_unstemmed | The Principles of Deep Learning Theory |
title_short | The Principles of Deep Learning Theory |
title_sort | principles of deep learning theory |
topic | TH String Theory Seminar |
url | http://cds.cern.ch/record/2809263 |
work_keys_str_mv | AT robertsdan theprinciplesofdeeplearningtheory AT robertsdan principlesofdeeplearningtheory |