Cargando…
Designing rotationally invariant neural networks from PDEs and variational methods
Partial differential equation models and their associated variational energy formulations are often rotationally invariant by design. This ensures that a rotation of the input results in a corresponding rotation of the output, which is desirable in applications such as image analysis. Convolutional...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9352643/ https://www.ncbi.nlm.nih.gov/pubmed/35941960 http://dx.doi.org/10.1007/s40687-022-00339-x |
_version_ | 1784762692855136256 |
---|---|
author | Alt, Tobias Schrader, Karl Weickert, Joachim Peter, Pascal Augustin, Matthias |
author_facet | Alt, Tobias Schrader, Karl Weickert, Joachim Peter, Pascal Augustin, Matthias |
author_sort | Alt, Tobias |
collection | PubMed |
description | Partial differential equation models and their associated variational energy formulations are often rotationally invariant by design. This ensures that a rotation of the input results in a corresponding rotation of the output, which is desirable in applications such as image analysis. Convolutional neural networks (CNNs) do not share this property, and existing remedies are often complex. The goal of our paper is to investigate how diffusion and variational models achieve rotation invariance and transfer these ideas to neural networks. As a core novelty, we propose activation functions which couple network channels by combining information from several oriented filters. This guarantees rotation invariance within the basic building blocks of the networks while still allowing for directional filtering. The resulting neural architectures are inherently rotationally invariant. With only a few small filters, they can achieve the same invariance as existing techniques which require a fine-grained sampling of orientations. Our findings help to translate diffusion and variational models into mathematically well-founded network architectures and provide novel concepts for model-based CNN design. |
format | Online Article Text |
id | pubmed-9352643 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-93526432022-08-06 Designing rotationally invariant neural networks from PDEs and variational methods Alt, Tobias Schrader, Karl Weickert, Joachim Peter, Pascal Augustin, Matthias Res Math Sci Research Partial differential equation models and their associated variational energy formulations are often rotationally invariant by design. This ensures that a rotation of the input results in a corresponding rotation of the output, which is desirable in applications such as image analysis. Convolutional neural networks (CNNs) do not share this property, and existing remedies are often complex. The goal of our paper is to investigate how diffusion and variational models achieve rotation invariance and transfer these ideas to neural networks. As a core novelty, we propose activation functions which couple network channels by combining information from several oriented filters. This guarantees rotation invariance within the basic building blocks of the networks while still allowing for directional filtering. The resulting neural architectures are inherently rotationally invariant. With only a few small filters, they can achieve the same invariance as existing techniques which require a fine-grained sampling of orientations. Our findings help to translate diffusion and variational models into mathematically well-founded network architectures and provide novel concepts for model-based CNN design. Springer International Publishing 2022-08-04 2022 /pmc/articles/PMC9352643/ /pubmed/35941960 http://dx.doi.org/10.1007/s40687-022-00339-x Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Research Alt, Tobias Schrader, Karl Weickert, Joachim Peter, Pascal Augustin, Matthias Designing rotationally invariant neural networks from PDEs and variational methods |
title | Designing rotationally invariant neural networks from PDEs and variational methods |
title_full | Designing rotationally invariant neural networks from PDEs and variational methods |
title_fullStr | Designing rotationally invariant neural networks from PDEs and variational methods |
title_full_unstemmed | Designing rotationally invariant neural networks from PDEs and variational methods |
title_short | Designing rotationally invariant neural networks from PDEs and variational methods |
title_sort | designing rotationally invariant neural networks from pdes and variational methods |
topic | Research |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9352643/ https://www.ncbi.nlm.nih.gov/pubmed/35941960 http://dx.doi.org/10.1007/s40687-022-00339-x |
work_keys_str_mv | AT alttobias designingrotationallyinvariantneuralnetworksfrompdesandvariationalmethods AT schraderkarl designingrotationallyinvariantneuralnetworksfrompdesandvariationalmethods AT weickertjoachim designingrotationallyinvariantneuralnetworksfrompdesandvariationalmethods AT peterpascal designingrotationallyinvariantneuralnetworksfrompdesandvariationalmethods AT augustinmatthias designingrotationallyinvariantneuralnetworksfrompdesandvariationalmethods |