Deep Learning applications are pervasive today, and efficient strategies are designed to reduce the computational time and resource demand of the training process. The Distributed Deep Learning (DDL) paradigm yields a significant speed-up by partitioning the training into multiple, parallel tasks. The Ray framework supports DDL applications exploiting data parallelism by enhancing the scalability with minimal user effort. This work aims at evaluating the performance of DDL training applications, by profiling their execution on a Ray cluster and developing Machine Learning-based models to predict the training time when changing the dataset size, the number of parallel workers and the amount of computational resources. Such performance-prediction models are crucial to forecast computational resources usage and costs in Cloud environments. Experimental results prove that our models achieve average prediction errors between 3 and 15% for both interpolation and extrapolation, thus demonstrating their applicability to unforeseen scenarios.

Filippini, F., Lublinsky, B., De Bayser, M., Ardagna, D. (2023). Performance Models for Distributed Deep Learning Training Jobs on Ray. In 2023 49th Euromicro Conference on Software Engineering and Advanced Applications (SEAA) (pp.30-35). Institute of Electrical and Electronics Engineers Inc. [10.1109/SEAA60479.2023.00014].

Performance Models for Distributed Deep Learning Training Jobs on Ray

Filippini F.;
2023

Abstract

Deep Learning applications are pervasive today, and efficient strategies are designed to reduce the computational time and resource demand of the training process. The Distributed Deep Learning (DDL) paradigm yields a significant speed-up by partitioning the training into multiple, parallel tasks. The Ray framework supports DDL applications exploiting data parallelism by enhancing the scalability with minimal user effort. This work aims at evaluating the performance of DDL training applications, by profiling their execution on a Ray cluster and developing Machine Learning-based models to predict the training time when changing the dataset size, the number of parallel workers and the amount of computational resources. Such performance-prediction models are crucial to forecast computational resources usage and costs in Cloud environments. Experimental results prove that our models achieve average prediction errors between 3 and 15% for both interpolation and extrapolation, thus demonstrating their applicability to unforeseen scenarios.
paper
Distributed training; Performance models; Ray;
English
49th Euromicro Conference on Software Engineering and Advanced Applications, SEAA 2023 - 06-08 September 2023
2023
2023 49th Euromicro Conference on Software Engineering and Advanced Applications (SEAA)
9798350342352
2023
30
35
open
Filippini, F., Lublinsky, B., De Bayser, M., Ardagna, D. (2023). Performance Models for Distributed Deep Learning Training Jobs on Ray. In 2023 49th Euromicro Conference on Software Engineering and Advanced Applications (SEAA) (pp.30-35). Institute of Electrical and Electronics Engineers Inc. [10.1109/SEAA60479.2023.00014].
File in questo prodotto:
File Dimensione Formato  
Filippini et al-2023-SEAA-AAM.pdf

accesso aperto

Tipologia di allegato: Author’s Accepted Manuscript, AAM (Post-print)
Licenza: Licenza open access specifica dell’editore
Dimensione 6.11 MB
Formato Adobe PDF
6.11 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/601085
Citazioni
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
Social impact