Skip to main content
SHARE
Publication

Distributed Training and Optimization of Neural Networks...

by Jean-roch Vlimant, Junqi Yin
Publication Type
Book Chapter
Publication Date
Page Numbers
243 to 264
Publisher Name
World Scientific Publishing Co Pte Ltd
Publisher Location
Toh Tuck Link, Singapore

Deep learning models are yielding increasingly better performances thanks to multiple factors. To be successful, model may have large number of parameters or complex architectures and be trained on large dataset. This leads to large requirements on computing resource and turn around time, even more so when hyperparameter optimization is done (e.g. search over model architectures). While this is a challenge that goes beyond particle physics, we review the various ways to do the necessary computations in parallel, and put it in the context of high-energy physics.