Cargando…

End-to-end Sinkhorn Autoencoder with Noise Generator

In this work, we propose a novel end-to-end Sinkhorn Autoencoder with a noise generator for efficient data collection simulation. Simulating processes that aim at collecting experimental data is crucial for multiple real-life applications, including nuclear medicine, astronomy, and high energy physi...

Descripción completa

Detalles Bibliográficos
Autores principales: Deja, Kamil, Dubiński, Jan, Nowak, Piotr, Wenzel, Sandro, Spurek, Przemysław, Trzciński, Tomasz
Lenguaje:eng
Publicado: 2020
Materias:
Acceso en línea:https://dx.doi.org/10.1109/ACCESS.2020.3048622
http://cds.cern.ch/record/2749256
_version_ 1780969031011926016
author Deja, Kamil
Dubiński, Jan
Nowak, Piotr
Wenzel, Sandro
Spurek, Przemysław
Trzciński, Tomasz
author_facet Deja, Kamil
Dubiński, Jan
Nowak, Piotr
Wenzel, Sandro
Spurek, Przemysław
Trzciński, Tomasz
author_sort Deja, Kamil
collection CERN
description In this work, we propose a novel end-to-end Sinkhorn Autoencoder with a noise generator for efficient data collection simulation. Simulating processes that aim at collecting experimental data is crucial for multiple real-life applications, including nuclear medicine, astronomy, and high energy physics. Contemporary methods, such as Monte Carlo algorithms, provide high-fidelity results at a price of high computational cost. Multiple attempts are taken to reduce this burden, e.g. using generative approaches based on Generative Adversarial Networks or Variational Autoencoders. Although such methods are much faster, they are often unstable in training and do not allow sampling from an entire data distribution. To address these shortcomings, we introduce a novel method dubbed end-to-end Sinkhorn Autoencoder, that leverages the Sinkhorn algorithm to explicitly align distribution of encoded real data examples and generated noise. More precisely, we extend autoencoder architecture by adding a deterministic neural network trained to map noise from a known distribution onto autoencoder latent space representing data distribution. We optimise the entire model jointly. Our method outperforms co mpeting approaches on a challenging dataset of simulation data from Zero Degree Calorimeters of ALICE experiment in LHC. as well as standard benchmarks, such as MNIST and CelebA.
id cern-2749256
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2020
record_format invenio
spelling cern-27492562023-06-29T03:27:29Zdoi:10.1109/ACCESS.2020.3048622http://cds.cern.ch/record/2749256engDeja, KamilDubiński, JanNowak, PiotrWenzel, SandroSpurek, PrzemysławTrzciński, TomaszEnd-to-end Sinkhorn Autoencoder with Noise Generatorstat.MLMathematical Physics and Mathematicscs.LGComputing and ComputersIn this work, we propose a novel end-to-end Sinkhorn Autoencoder with a noise generator for efficient data collection simulation. Simulating processes that aim at collecting experimental data is crucial for multiple real-life applications, including nuclear medicine, astronomy, and high energy physics. Contemporary methods, such as Monte Carlo algorithms, provide high-fidelity results at a price of high computational cost. Multiple attempts are taken to reduce this burden, e.g. using generative approaches based on Generative Adversarial Networks or Variational Autoencoders. Although such methods are much faster, they are often unstable in training and do not allow sampling from an entire data distribution. To address these shortcomings, we introduce a novel method dubbed end-to-end Sinkhorn Autoencoder, that leverages the Sinkhorn algorithm to explicitly align distribution of encoded real data examples and generated noise. More precisely, we extend autoencoder architecture by adding a deterministic neural network trained to map noise from a known distribution onto autoencoder latent space representing data distribution. We optimise the entire model jointly. Our method outperforms co mpeting approaches on a challenging dataset of simulation data from Zero Degree Calorimeters of ALICE experiment in LHC. as well as standard benchmarks, such as MNIST and CelebA.In this work, we propose a novel end-to-end sinkhorn autoencoder with noise generator for efficient data collection simulation. Simulating processes that aim at collecting experimental data is crucial for multiple real-life applications, including nuclear medicine, astronomy and high energy physics. Contemporary methods, such as Monte Carlo algorithms, provide high-fidelity results at a price of high computational cost. Multiple attempts are taken to reduce this burden, e.g. using generative approaches based on Generative Adversarial Networks or Variational Autoencoders. Although such methods are much faster, they are often unstable in training and do not allow sampling from an entire data distribution. To address these shortcomings, we introduce a novel method dubbed end-to-end Sinkhorn Autoencoder, that leverages sinkhorn algorithm to explicitly align distribution of encoded real data examples and generated noise. More precisely, we extend autoencoder architecture by adding a deterministic neural network trained to map noise from a known distribution onto autoencoder latent space representing data distribution. We optimise the entire model jointly. Our method outperforms competing approaches on a challenging dataset of simulation data from Zero Degree Calorimeters of ALICE experiment in LHC. as well as standard benchmarks, such as MNIST and CelebA.arXiv:2006.06704oai:cds.cern.ch:27492562020-06-11
spellingShingle stat.ML
Mathematical Physics and Mathematics
cs.LG
Computing and Computers
Deja, Kamil
Dubiński, Jan
Nowak, Piotr
Wenzel, Sandro
Spurek, Przemysław
Trzciński, Tomasz
End-to-end Sinkhorn Autoencoder with Noise Generator
title End-to-end Sinkhorn Autoencoder with Noise Generator
title_full End-to-end Sinkhorn Autoencoder with Noise Generator
title_fullStr End-to-end Sinkhorn Autoencoder with Noise Generator
title_full_unstemmed End-to-end Sinkhorn Autoencoder with Noise Generator
title_short End-to-end Sinkhorn Autoencoder with Noise Generator
title_sort end-to-end sinkhorn autoencoder with noise generator
topic stat.ML
Mathematical Physics and Mathematics
cs.LG
Computing and Computers
url https://dx.doi.org/10.1109/ACCESS.2020.3048622
http://cds.cern.ch/record/2749256
work_keys_str_mv AT dejakamil endtoendsinkhornautoencoderwithnoisegenerator
AT dubinskijan endtoendsinkhornautoencoderwithnoisegenerator
AT nowakpiotr endtoendsinkhornautoencoderwithnoisegenerator
AT wenzelsandro endtoendsinkhornautoencoderwithnoisegenerator
AT spurekprzemysław endtoendsinkhornautoencoderwithnoisegenerator
AT trzcinskitomasz endtoendsinkhornautoencoderwithnoisegenerator