Cargando…

Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks

We develop biologically plausible training mechanisms for self-supervised learning (SSL) in deep networks. Specifically, by biologically plausible training we mean (i) all updates of weights are based on current activities of pre-synaptic units and current, or activity retrieved from short term memo...

Descripción completa

Detalles Bibliográficos
Autores principales: Tang, Mufeng, Yang, Yibo, Amit, Yali
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8977509/
https://www.ncbi.nlm.nih.gov/pubmed/35386856
http://dx.doi.org/10.3389/fncom.2022.789253
_version_ 1784680779731697664
author Tang, Mufeng
Yang, Yibo
Amit, Yali
author_facet Tang, Mufeng
Yang, Yibo
Amit, Yali
author_sort Tang, Mufeng
collection PubMed
description We develop biologically plausible training mechanisms for self-supervised learning (SSL) in deep networks. Specifically, by biologically plausible training we mean (i) all updates of weights are based on current activities of pre-synaptic units and current, or activity retrieved from short term memory of post synaptic units, including at the top-most error computing layer, (ii) complex computations such as normalization, inner products and division are avoided, (iii) asymmetric connections between units, and (iv) most learning is carried out in an unsupervised manner. SSL with a contrastive loss satisfies the third condition as it does not require labeled data and it introduces robustness to observed perturbations of objects, which occur naturally as objects or observers move in 3D and with variable lighting over time. We propose a contrastive hinge based loss whose error involves simple local computations satisfying (ii), as opposed to the standard contrastive losses employed in the literature, which do not lend themselves easily to implementation in a network architecture due to complex computations involving ratios and inner products. Furthermore, we show that learning can be performed with one of two more plausible alternatives to backpropagation that satisfy conditions (i) and (ii). The first is difference target propagation (DTP), which trains network parameters using target-based local losses and employs a Hebbian learning rule, thus overcoming the biologically implausible symmetric weight problem in backpropagation. The second is layer-wise learning, where each layer is directly connected to a layer computing the loss error. The layers are either updated sequentially in a greedy fashion (GLL) or in random order (RLL), and each training stage involves a single hidden layer network. Backpropagation through one layer needed for each such network can either be altered with fixed random feedback weights (RF) or using updated random feedback weights (URF) as in Amity's study 2019. Both methods represent alternatives to the symmetric weight issue of backpropagation. By training convolutional neural networks (CNNs) with SSL and DTP, GLL or RLL, we find that our proposed framework achieves comparable performance to standard BP learning downstream linear classifier evaluation of the learned embeddings.
format Online
Article
Text
id pubmed-8977509
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-89775092022-04-05 Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks Tang, Mufeng Yang, Yibo Amit, Yali Front Comput Neurosci Neuroscience We develop biologically plausible training mechanisms for self-supervised learning (SSL) in deep networks. Specifically, by biologically plausible training we mean (i) all updates of weights are based on current activities of pre-synaptic units and current, or activity retrieved from short term memory of post synaptic units, including at the top-most error computing layer, (ii) complex computations such as normalization, inner products and division are avoided, (iii) asymmetric connections between units, and (iv) most learning is carried out in an unsupervised manner. SSL with a contrastive loss satisfies the third condition as it does not require labeled data and it introduces robustness to observed perturbations of objects, which occur naturally as objects or observers move in 3D and with variable lighting over time. We propose a contrastive hinge based loss whose error involves simple local computations satisfying (ii), as opposed to the standard contrastive losses employed in the literature, which do not lend themselves easily to implementation in a network architecture due to complex computations involving ratios and inner products. Furthermore, we show that learning can be performed with one of two more plausible alternatives to backpropagation that satisfy conditions (i) and (ii). The first is difference target propagation (DTP), which trains network parameters using target-based local losses and employs a Hebbian learning rule, thus overcoming the biologically implausible symmetric weight problem in backpropagation. The second is layer-wise learning, where each layer is directly connected to a layer computing the loss error. The layers are either updated sequentially in a greedy fashion (GLL) or in random order (RLL), and each training stage involves a single hidden layer network. Backpropagation through one layer needed for each such network can either be altered with fixed random feedback weights (RF) or using updated random feedback weights (URF) as in Amity's study 2019. Both methods represent alternatives to the symmetric weight issue of backpropagation. By training convolutional neural networks (CNNs) with SSL and DTP, GLL or RLL, we find that our proposed framework achieves comparable performance to standard BP learning downstream linear classifier evaluation of the learned embeddings. Frontiers Media S.A. 2022-03-21 /pmc/articles/PMC8977509/ /pubmed/35386856 http://dx.doi.org/10.3389/fncom.2022.789253 Text en Copyright © 2022 Tang, Yang and Amit. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Tang, Mufeng
Yang, Yibo
Amit, Yali
Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
title Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
title_full Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
title_fullStr Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
title_full_unstemmed Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
title_short Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
title_sort biologically plausible training mechanisms for self-supervised learning in deep networks
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8977509/
https://www.ncbi.nlm.nih.gov/pubmed/35386856
http://dx.doi.org/10.3389/fncom.2022.789253
work_keys_str_mv AT tangmufeng biologicallyplausibletrainingmechanismsforselfsupervisedlearningindeepnetworks
AT yangyibo biologicallyplausibletrainingmechanismsforselfsupervisedlearningindeepnetworks
AT amityali biologicallyplausibletrainingmechanismsforselfsupervisedlearningindeepnetworks