Cargando…

Information Bottleneck Classification in Extremely Distributed Systems

We present a new decentralized classification system based on a distributed architecture. This system consists of distributed nodes, each possessing their own datasets and computing modules, along with a centralized server, which provides probes to classification and aggregates the responses of node...

Descripción completa

Detalles Bibliográficos
Autores principales: Ullmann, Denis, Rezaeifar, Shideh, Taran, Olga, Holotyak, Taras, Panos, Brandon, Voloshynovskiy, Slava
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7711965/
https://www.ncbi.nlm.nih.gov/pubmed/33287005
http://dx.doi.org/10.3390/e22111237
_version_ 1783618263597449216
author Ullmann, Denis
Rezaeifar, Shideh
Taran, Olga
Holotyak, Taras
Panos, Brandon
Voloshynovskiy, Slava
author_facet Ullmann, Denis
Rezaeifar, Shideh
Taran, Olga
Holotyak, Taras
Panos, Brandon
Voloshynovskiy, Slava
author_sort Ullmann, Denis
collection PubMed
description We present a new decentralized classification system based on a distributed architecture. This system consists of distributed nodes, each possessing their own datasets and computing modules, along with a centralized server, which provides probes to classification and aggregates the responses of nodes for a final decision. Each node, with access to its own training dataset of a given class, is trained based on an auto-encoder system consisting of a fixed data-independent encoder, a pre-trained quantizer and a class-dependent decoder. Hence, these auto-encoders are highly dependent on the class probability distribution for which the reconstruction distortion is minimized. Alternatively, when an encoding–quantizing–decoding node observes data from different distributions, unseen at training, there is a mismatch, and such a decoding is not optimal, leading to a significant increase of the reconstruction distortion. The final classification is performed at the centralized classifier that votes for the class with the minimum reconstruction distortion. In addition to the system applicability for applications facing big-data communication problems and or requiring private classification, the above distributed scheme creates a theoretical bridge to the information bottleneck principle. The proposed system demonstrates a very promising performance on basic datasets such as MNIST and FasionMNIST.
format Online
Article
Text
id pubmed-7711965
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-77119652021-02-24 Information Bottleneck Classification in Extremely Distributed Systems Ullmann, Denis Rezaeifar, Shideh Taran, Olga Holotyak, Taras Panos, Brandon Voloshynovskiy, Slava Entropy (Basel) Article We present a new decentralized classification system based on a distributed architecture. This system consists of distributed nodes, each possessing their own datasets and computing modules, along with a centralized server, which provides probes to classification and aggregates the responses of nodes for a final decision. Each node, with access to its own training dataset of a given class, is trained based on an auto-encoder system consisting of a fixed data-independent encoder, a pre-trained quantizer and a class-dependent decoder. Hence, these auto-encoders are highly dependent on the class probability distribution for which the reconstruction distortion is minimized. Alternatively, when an encoding–quantizing–decoding node observes data from different distributions, unseen at training, there is a mismatch, and such a decoding is not optimal, leading to a significant increase of the reconstruction distortion. The final classification is performed at the centralized classifier that votes for the class with the minimum reconstruction distortion. In addition to the system applicability for applications facing big-data communication problems and or requiring private classification, the above distributed scheme creates a theoretical bridge to the information bottleneck principle. The proposed system demonstrates a very promising performance on basic datasets such as MNIST and FasionMNIST. MDPI 2020-10-30 /pmc/articles/PMC7711965/ /pubmed/33287005 http://dx.doi.org/10.3390/e22111237 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ullmann, Denis
Rezaeifar, Shideh
Taran, Olga
Holotyak, Taras
Panos, Brandon
Voloshynovskiy, Slava
Information Bottleneck Classification in Extremely Distributed Systems
title Information Bottleneck Classification in Extremely Distributed Systems
title_full Information Bottleneck Classification in Extremely Distributed Systems
title_fullStr Information Bottleneck Classification in Extremely Distributed Systems
title_full_unstemmed Information Bottleneck Classification in Extremely Distributed Systems
title_short Information Bottleneck Classification in Extremely Distributed Systems
title_sort information bottleneck classification in extremely distributed systems
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7711965/
https://www.ncbi.nlm.nih.gov/pubmed/33287005
http://dx.doi.org/10.3390/e22111237
work_keys_str_mv AT ullmanndenis informationbottleneckclassificationinextremelydistributedsystems
AT rezaeifarshideh informationbottleneckclassificationinextremelydistributedsystems
AT taranolga informationbottleneckclassificationinextremelydistributedsystems
AT holotyaktaras informationbottleneckclassificationinextremelydistributedsystems
AT panosbrandon informationbottleneckclassificationinextremelydistributedsystems
AT voloshynovskiyslava informationbottleneckclassificationinextremelydistributedsystems