Cargando…

Hyperparameter Optimization for Deep Learning Models Using High Performance Computing

<!--HTML--><h2><strong>Abstract</strong></h2><p><br>In the past decade, Machine Learning (ML), and especially Deep Learning (DL), has outperformed traditional rule-based algorithms on a wide variety of tasks, such as for instance image recognition, object de...

Descripción completa

Detalles Bibliográficos
Autor principal: Wulff, Eric
Lenguaje:eng
Publicado: 2023
Materias:
Acceso en línea:http://cds.cern.ch/record/2865457
_version_ 1780978043714535424
author Wulff, Eric
author_facet Wulff, Eric
author_sort Wulff, Eric
collection CERN
description <!--HTML--><h2><strong>Abstract</strong></h2><p><br>In the past decade, Machine Learning (ML), and especially Deep Learning (DL), has outperformed traditional rule-based algorithms on a wide variety of tasks, such as for instance image recognition, object detection and natural language processing. In CoE RAISE, we have additionally seen that ML can unlock new potential in fields such as high energy physics (HEP), remote sensing, seismic imaging, additive manufacturing, and acoustics. Training DL models, however, is no trivial task, especially if the model is large and have many hyperparameters (HP). To tackle this challenge, Hyperparameter Optimization (HPO) can be used to systematically explore the search space of possible HP configurations and, paired with the computing power of modern High Performance Computing (HPC) systems, it can drastically speed up the process of improving DL models. The aim of this talk is to introduce HPO and the major challenges data scientists face when tuning their models, as well as to give some examples from a HEP use-case where large-scale HPO on HPC systems was successfully applied.</p><h2><strong>Bio</strong></h2><p><br>Eric Wulff has a MSc in Engineering Physics from Lund University and is a fellow in the IT department at CERN. He is the Task Leader for the use-case on LHC collision event reconstruction at the European Center of Excellence in Exascale Computing (CoE RAISE). His experience includes large-scale distributed training and hyperparameter optimization of AI models on supercomputers as well as using quantum computing for DL-based algorithms. Prior to joining CERN, Eric was a Machine Learning Engineer at Axis Communications, where he worked on object detection and video analytics using DL techniques.</p>
id cern-2865457
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2023
record_format invenio
spelling cern-28654572023-07-19T20:00:35Zhttp://cds.cern.ch/record/2865457engWulff, EricHyperparameter Optimization for Deep Learning Models Using High Performance ComputingHyperparameter Optimization for Deep Learning Models Using High Performance ComputingCERN openlab summer student lecture programme<!--HTML--><h2><strong>Abstract</strong></h2><p><br>In the past decade, Machine Learning (ML), and especially Deep Learning (DL), has outperformed traditional rule-based algorithms on a wide variety of tasks, such as for instance image recognition, object detection and natural language processing. In CoE RAISE, we have additionally seen that ML can unlock new potential in fields such as high energy physics (HEP), remote sensing, seismic imaging, additive manufacturing, and acoustics. Training DL models, however, is no trivial task, especially if the model is large and have many hyperparameters (HP). To tackle this challenge, Hyperparameter Optimization (HPO) can be used to systematically explore the search space of possible HP configurations and, paired with the computing power of modern High Performance Computing (HPC) systems, it can drastically speed up the process of improving DL models. The aim of this talk is to introduce HPO and the major challenges data scientists face when tuning their models, as well as to give some examples from a HEP use-case where large-scale HPO on HPC systems was successfully applied.</p><h2><strong>Bio</strong></h2><p><br>Eric Wulff has a MSc in Engineering Physics from Lund University and is a fellow in the IT department at CERN. He is the Task Leader for the use-case on LHC collision event reconstruction at the European Center of Excellence in Exascale Computing (CoE RAISE). His experience includes large-scale distributed training and hyperparameter optimization of AI models on supercomputers as well as using quantum computing for DL-based algorithms. Prior to joining CERN, Eric was a Machine Learning Engineer at Axis Communications, where he worked on object detection and video analytics using DL techniques.</p>oai:cds.cern.ch:28654572023
spellingShingle CERN openlab summer student lecture programme
Wulff, Eric
Hyperparameter Optimization for Deep Learning Models Using High Performance Computing
title Hyperparameter Optimization for Deep Learning Models Using High Performance Computing
title_full Hyperparameter Optimization for Deep Learning Models Using High Performance Computing
title_fullStr Hyperparameter Optimization for Deep Learning Models Using High Performance Computing
title_full_unstemmed Hyperparameter Optimization for Deep Learning Models Using High Performance Computing
title_short Hyperparameter Optimization for Deep Learning Models Using High Performance Computing
title_sort hyperparameter optimization for deep learning models using high performance computing
topic CERN openlab summer student lecture programme
url http://cds.cern.ch/record/2865457
work_keys_str_mv AT wulfferic hyperparameteroptimizationfordeeplearningmodelsusinghighperformancecomputing