
Comparison Distributed and Parallel Machine Learning: Evidence from Models, Principles and Application Scenarios
- 1 Department of Communication Engineering, Wuhan University of Science and Technology, Wuhan, China
* Author to whom correspondence should be addressed.
Abstract
Contemporarily, the demand for processing large-scale data has been rapidly increasing, prompting continuous advancements in the field of machine learning. This study examines the state of parallel and distributed machine learning at present, both of which aim to enhance computational efficiency when dealing with large datasets and demonstrate promising applications across various domains. As the scale of data escalates, traditional single-machine learning methods are becoming increasingly inadequate, which lead to the emergence of parallel and distributed machine learning. These approaches enable substantial computations to be performed more efficiently through the collaboration of multi-core CPUs or multiple computing nodes. This research conducts an in-depth analysis of the inherent challenges associated with these methods, including data transmission latency, synchronization requirements, and user privacy concerns. Ultimately, this research emphasizes the tremendous potential of both methods for future applications, a potential that is bolstered by ongoing advancements in hardware and algorithm optimization. These results provide valuable insights for practitioners in the field and offers guidance for future research directions.
Keywords
Parallel and distributed machine learning, data scalability, hardware and algorithm optimization.
[1]. Fradkov A 2020 Early history of machine learning IFAC-PapersOnLine vol 53(2) pp 1385-1390
[2]. Tyagi A, Kukreja S, Nair M M and Tyagi A K 2022 Machine Learning: Past Present and Future School of Computer Science and Engineering Vellore Institute of Technology; Terna Engineering College University of Mumbai
[3]. Park D J, Park M W, Lee H, Kim Y J, Kim Y and Park Y H 2021 Development of machine learning model for diagnostic disease prediction based on laboratory tests Scientific Reports vol 11(1) p 7567
[4]. Salman S A, Dheyab S A, Salih Q M and Hammood W A 2023 Parallel machine learning algorithms Mesopotamian Journal of Big Data vol 12 pp 12–15
[5]. Wang J, Hei H, Zheng Y, et al 2024 Five-site water models for ice and liquid water generated by a series–parallel machine learning strategy Journal of Chemical Theory and Computation American Chemical Society vol 19 p 6
[6]. Verbraeken J, Wolting M, Katzy J, Kloppenburg J, Verbelen T and Rellermeyer J S 2020 A survey on distributed machine learning ACM Computing Surveys (CSUR) vol 53(2) pp 1-33
[7]. Muscinelli E, Shinde S S and Tarchi D 2022 Overview of distributed machine learning techniques for 6G networks Algorithms vol 15(6) p 210
[8]. Hoang T M, Le Thi T L and Quy N M 2023 A novel distributed machine learning model to detect attacks on edge computing network Journal of Advanced Information Technology vol 14(1) pp 153–159
[9]. Miao Q H, Lv Y S, Huang M, Wang X and Wang F Y 2023 Parallel learning: Overview and perspective for computational learning across Syn2Real and Sim2Real IEEE/CAA Journal of Automatica Sinica vol 10(3) pp 603–631
[10]. Chatterjee B 2024 Distributed machine learning Proceedings of the 25th International Conference on Distributed Computing and Networking pp 4-7
[11]. Srinivasan K, Coble N, Hamlin J, Antonsen T, Ott E and Girvan M 2022 Parallel machine learning for forecasting the dynamics of complex networks Physical Review Letters vol 128(16) p 164101
[12]. Imakura A and Sakurai T 2020 Data collaboration analysis framework using centralization of individual intermediate representations for distributed data sets ASCE-ASME Journal of Risk and Uncertainty in Engineering Systems Part A: Civil Engineering vol 6(2)
[13]. Bogdanova A, Imakura A and Sakurai T 2023 DC-SHAP method for consistent explainability in privacy-preserving distributed machine learning Human-Centric Intelligent Systems vol 3 pp 197–210
Cite this article
Zhang,Z. (2024). Comparison Distributed and Parallel Machine Learning: Evidence from Models, Principles and Application Scenarios. Applied and Computational Engineering,102,155-161.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 2nd International Conference on Machine Learning and Automation
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).