• Medientyp: Elektronischer Konferenzbericht
  • Titel: Short Paper: Accelerating Hyperparameter Optimization Algorithms with Mixed Precision
  • Beteiligte: Aach, Marcel [Verfasser:in]; Sarma, Rakesh [Verfasser:in]; Inanc, Eray [Verfasser:in]; Riedel, Morris [Verfasser:in]; Lintermann, Andreas [Verfasser:in]
  • Erschienen: ACM New York, NY, USA, 2023
  • Erschienen in: ACM New York, NY, USA 1776–1779 (2023). doi:10.1145/3624062.3624259 ; SC-W 2023: Workshops of The International Conference on High Performance Computing, Network, Storage, and Analysis, SC 2023, Denver, CO, USA, 2023-11-12 - 2023-11-17
  • Sprache: Englisch
  • DOI: https://doi.org/10.1145/3624062.3624259; https://doi.org/10.34734/FZJ-2023-04518
  • Entstehung:
  • Anmerkungen: Diese Datenquelle enthält auch Bestandsnachweise, die nicht zu einem Volltext führen.
  • Beschreibung: Hyperparameter Optimization (HPO) of Neural Networks (NNs) is a computationally expensive procedure. On accelerators, such as NVIDIA Graphics Processing Units (GPUs) equipped with Tensor Cores, it is possible to speed-up the NN training by reducing the precision of some of the NN parameters, also referred to as mixed precision training. This paper investigates the performance of three popular HPO algorithms in terms of the achieved speed-up and model accuracy, utilizing early stopping, Bayesian, and genetic optimization approaches, in combination with mixed precision functionalities. The benchmarks are performed on 64 GPUs in parallel on three datasets: two from the vision and one from the Computational Fluid Dynamics domain. The results show that larger speed-ups can be achieved for mixed compared to full precision HPO if the checkpoint frequency is kept low. In addition to the reduced runtime, small gains in generalization performance on the test set are observed.
  • Zugangsstatus: Freier Zugang