EP-IT Data Science Seminars

Hyperparameter Optimization for Deep Learning Models Using High Performance Computing

by Eric Wulff (CERN)

Europe/Zurich
40/S2-C01 - Salle Curie (CERN)

40/S2-C01 - Salle Curie

CERN

115
Show room on map
Description

In the past decade, Machine Learning (ML), and in particular Deep Learning (DL), has outperformed traditional rule-based algorithms on a wide variety of tasks, such as for instance image recognition, object detection and natural language processing. In CoE RAISE, we have additionally seen that ML can unlock new potential in fields such as high energy physics (HEP), remote sensing, seismic imaging, additive manufacturing, and acoustics. Training DL models, however, is no trivial task, especially if the model is large and have many tunable hyperparameters. To tackle this challenge, Hyperparameter Optimization (HPO) can be used to systematically explore the search space of possible hyperparameter configurations and, paired with the computing power of modern High Performance Computing (HPC) systems, it can drastically speed up the process of improving DL models. The aim of this talk is to give an introduction to HPO and the major challenges data scientists face when tuning their models, as well as to give some examples from a HEP use-case where large-scale HPO on HPC systems was successfully applied.

Eric Wulff is a fellow in the IT department at CERN and Task Leader for the use-case on LHC collision event reconstruction at the European Center of Excellence in Exascale Computing (CoE RAISE). His experience includes large-scale distributed training and hyperparameter optimization of DL algorithms on supercomputers and using quantum computing for ML/DL-based algorithms. Prior to joining CERN, Eric was a Machine Learning Engineer at Axis Communications, where he worked on object detection and video analytics using deep learning techniques.

Coffee will be served at 10:30.

Organised by

M. Girone, M. Elsing, L. Moneta, M. Pierini