Cargando…

Compact Neural Architecture Designs by Tensor Representations

We propose a framework of tensorial neural networks (TNNs) extending existing linear layers on low-order tensors to multilinear operations on higher-order tensors. TNNs have three advantages over existing networks: First, TNNs naturally apply to higher-order data without flattening, which preserves...

Descripción completa

Detalles Bibliográficos
Autores principales: Su, Jiahao, Li, Jingling, Liu, Xiaoyu, Ranadive, Teresa, Coley, Christopher, Tuan, Tai-Ching, Huang, Furong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8959219/
https://www.ncbi.nlm.nih.gov/pubmed/35355829
http://dx.doi.org/10.3389/frai.2022.728761
_version_ 1784677102343159808
author Su, Jiahao
Li, Jingling
Liu, Xiaoyu
Ranadive, Teresa
Coley, Christopher
Tuan, Tai-Ching
Huang, Furong
author_facet Su, Jiahao
Li, Jingling
Liu, Xiaoyu
Ranadive, Teresa
Coley, Christopher
Tuan, Tai-Ching
Huang, Furong
author_sort Su, Jiahao
collection PubMed
description We propose a framework of tensorial neural networks (TNNs) extending existing linear layers on low-order tensors to multilinear operations on higher-order tensors. TNNs have three advantages over existing networks: First, TNNs naturally apply to higher-order data without flattening, which preserves their multi-dimensional structures. Second, compressing a pre-trained network into a TNN results in a model with similar expressive power but fewer parameters. Finally, TNNs interpret advanced compact designs of network architectures, such as bottleneck modules and interleaved group convolutions. To learn TNNs, we derive their backpropagation rules using a novel suite of generalized tensor algebra. With backpropagation, we can either learn TNNs from scratch or pre-trained models using knowledge distillation. Experiments on VGG, ResNet, and Wide-ResNet demonstrate that TNNs outperform the state-of-the-art low-rank methods on a wide range of backbone networks and datasets.
format Online
Article
Text
id pubmed-8959219
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-89592192022-03-29 Compact Neural Architecture Designs by Tensor Representations Su, Jiahao Li, Jingling Liu, Xiaoyu Ranadive, Teresa Coley, Christopher Tuan, Tai-Ching Huang, Furong Front Artif Intell Artificial Intelligence We propose a framework of tensorial neural networks (TNNs) extending existing linear layers on low-order tensors to multilinear operations on higher-order tensors. TNNs have three advantages over existing networks: First, TNNs naturally apply to higher-order data without flattening, which preserves their multi-dimensional structures. Second, compressing a pre-trained network into a TNN results in a model with similar expressive power but fewer parameters. Finally, TNNs interpret advanced compact designs of network architectures, such as bottleneck modules and interleaved group convolutions. To learn TNNs, we derive their backpropagation rules using a novel suite of generalized tensor algebra. With backpropagation, we can either learn TNNs from scratch or pre-trained models using knowledge distillation. Experiments on VGG, ResNet, and Wide-ResNet demonstrate that TNNs outperform the state-of-the-art low-rank methods on a wide range of backbone networks and datasets. Frontiers Media S.A. 2022-03-08 /pmc/articles/PMC8959219/ /pubmed/35355829 http://dx.doi.org/10.3389/frai.2022.728761 Text en Copyright © 2022 Su, Li, Liu, Ranadive, Coley, Tuan and Huang. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Su, Jiahao
Li, Jingling
Liu, Xiaoyu
Ranadive, Teresa
Coley, Christopher
Tuan, Tai-Ching
Huang, Furong
Compact Neural Architecture Designs by Tensor Representations
title Compact Neural Architecture Designs by Tensor Representations
title_full Compact Neural Architecture Designs by Tensor Representations
title_fullStr Compact Neural Architecture Designs by Tensor Representations
title_full_unstemmed Compact Neural Architecture Designs by Tensor Representations
title_short Compact Neural Architecture Designs by Tensor Representations
title_sort compact neural architecture designs by tensor representations
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8959219/
https://www.ncbi.nlm.nih.gov/pubmed/35355829
http://dx.doi.org/10.3389/frai.2022.728761
work_keys_str_mv AT sujiahao compactneuralarchitecturedesignsbytensorrepresentations
AT lijingling compactneuralarchitecturedesignsbytensorrepresentations
AT liuxiaoyu compactneuralarchitecturedesignsbytensorrepresentations
AT ranadiveteresa compactneuralarchitecturedesignsbytensorrepresentations
AT coleychristopher compactneuralarchitecturedesignsbytensorrepresentations
AT tuantaiching compactneuralarchitecturedesignsbytensorrepresentations
AT huangfurong compactneuralarchitecturedesignsbytensorrepresentations