TY - GEN
T1 - Optimizing Distributed Deep Learning in Heterogeneous Computing Platforms for Remote Sensing Data Classification
AU - Moreno-Alvarez, Sergio
AU - Paoletti, Mercedes E.
AU - Rico, Juan A.
AU - Cavallaro, Gabriele
AU - Haut, Juan M.
N1 - Publisher Copyright: © 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Applications from Remote Sensing (RS) unveiled unique challenges to Deep Learning (DL) due to the high volume and complexity of their data. On the one hand, deep neural network architectures have the capability to automatically ex-tract informative features from RS data. On the other hand, these models have massive amounts of tunable parameters, re-quiring high computational capabilities. Distributed DL with data parallelism on High-Performance Computing (HPC) sys-tems have proved necessary in dealing with the demands of DL models. Nevertheless, a single HPC system can be al-ready highly heterogeneous and include different computing resources with uneven processing power. In this context, a standard data parallelism strategy does not partition the data efficiently according to the available computing resources. This paper proposes an alternative approach to compute the gradient, which guarantees that the contribution to the gradi-ent calculation is proportional to the processing speed of each DL model's replica. The experimental results are obtained in a heterogeneous HPC system with RS data and demon-strate that the proposed approach provides a significant training speed up and gain in the global accuracy compared to one of the state-of-the-art distributed DL framework.
AB - Applications from Remote Sensing (RS) unveiled unique challenges to Deep Learning (DL) due to the high volume and complexity of their data. On the one hand, deep neural network architectures have the capability to automatically ex-tract informative features from RS data. On the other hand, these models have massive amounts of tunable parameters, re-quiring high computational capabilities. Distributed DL with data parallelism on High-Performance Computing (HPC) sys-tems have proved necessary in dealing with the demands of DL models. Nevertheless, a single HPC system can be al-ready highly heterogeneous and include different computing resources with uneven processing power. In this context, a standard data parallelism strategy does not partition the data efficiently according to the available computing resources. This paper proposes an alternative approach to compute the gradient, which guarantees that the contribution to the gradi-ent calculation is proportional to the processing speed of each DL model's replica. The experimental results are obtained in a heterogeneous HPC system with RS data and demon-strate that the proposed approach provides a significant training speed up and gain in the global accuracy compared to one of the state-of-the-art distributed DL framework.
KW - Convolutional neural networks
KW - Deep Learning
KW - Heterogeneous platforms
KW - High performance computing
KW - Remote sensing classification
UR - https://www.scopus.com/pages/publications/85140410050
U2 - 10.1109/IGARSS46834.2022.9883762
DO - 10.1109/IGARSS46834.2022.9883762
M3 - Conference contribution
T3 - International Geoscience and Remote Sensing Symposium (IGARSS)
SP - 2726
EP - 2729
BT - IGARSS 2022 - 2022 IEEE International Geoscience and Remote Sensing Symposium
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2022
Y2 - 17 July 2022 through 22 July 2022
ER -