TY - JOUR
T1 - Distributed hybrid quantum-classical performance prediction for hyperparameter optimization
AU - Wulff, Eric
AU - Garcia Amboage, Juan Pablo
AU - Aach, Marcel
AU - Gislason, Thorsteinn Eli
AU - Ingolfsson, Thorsteinn Kristinn
AU - Ingolfsson, Tomas Kristinn
AU - Pasetto, Edoardo
AU - Delilbasic, Amer
AU - Riedel, Morris
AU - Sarma, Rakesh
AU - Girone, Maria
AU - Lintermann, Andreas
N1 - Publisher Copyright: © The Author(s) 2024.
PY - 2024/12
Y1 - 2024/12
N2 - Hyperparameter optimization (HPO) of neural networks is a computationally expensive procedure, which requires a large number of different model configurations to be trained. To reduce such costs, this work presents a distributed, hybrid workflow, that runs the training of the neural networks on multiple graphics processing units (GPUs) on a classical supercomputer, while predicting the configurations’ performance with quantum-trained support vector regression (QT-SVR) on a quantum annealer (QA). The workflow is shown to run on up to 50 GPUs and a QA at the same time, completely automating the communication between the classical and the quantum systems. The approach is evaluated extensively on several benchmarking datasets from the computer vision (CV), high-energy physics (HEP), and natural language processing (NLP) domains. Empirical results show that resource costs for performing HPO can be reduced by up to 9% when using the hybrid workflow with performance prediction, compared to using a plain HPO algorithm without performance prediction. Additionally, the workflow obtains similar and in some cases even better accuracy of the final hyperparameter configuration, when combining multiple heuristically obtained predictions from the QA, compared to using just a single classically obtained prediction. The results highlight the potential of hybrid quantum-classical machine learning algorithms. The workflow code is made available open-source to foster adoption in the community.
AB - Hyperparameter optimization (HPO) of neural networks is a computationally expensive procedure, which requires a large number of different model configurations to be trained. To reduce such costs, this work presents a distributed, hybrid workflow, that runs the training of the neural networks on multiple graphics processing units (GPUs) on a classical supercomputer, while predicting the configurations’ performance with quantum-trained support vector regression (QT-SVR) on a quantum annealer (QA). The workflow is shown to run on up to 50 GPUs and a QA at the same time, completely automating the communication between the classical and the quantum systems. The approach is evaluated extensively on several benchmarking datasets from the computer vision (CV), high-energy physics (HEP), and natural language processing (NLP) domains. Empirical results show that resource costs for performing HPO can be reduced by up to 9% when using the hybrid workflow with performance prediction, compared to using a plain HPO algorithm without performance prediction. Additionally, the workflow obtains similar and in some cases even better accuracy of the final hyperparameter configuration, when combining multiple heuristically obtained predictions from the QA, compared to using just a single classically obtained prediction. The results highlight the potential of hybrid quantum-classical machine learning algorithms. The workflow code is made available open-source to foster adoption in the community.
KW - Distributed computing
KW - Hyperband
KW - Hyperparameter optimization
KW - Quantum annealing
UR - https://www.scopus.com/pages/publications/85203866977
U2 - 10.1007/s42484-024-00198-5
DO - 10.1007/s42484-024-00198-5
M3 - Article
SN - 2524-4906
VL - 6
JO - Quantum Machine Intelligence
JF - Quantum Machine Intelligence
IS - 2
M1 - 59
ER -