
Machine learning for privacy-preserving: Approaches, challenges and discussion
- 1 Donghua University
* Author to whom correspondence should be addressed.
Abstract
Currently, advanced technologies such as big data, artificial intelligence and machine learning are undergoing rapid development. However, the emergence of cybersecurity and privacy leakage problems has resulted in serious implications. This paper discusses the current state of privacy security issues in the field of machine learning in a comprehensive manner. During machine training, training models often unconsciously extract and record private information from raw data, and in addition, third-party attackers are interested in maliciously extracting private information from raw data. This paper first provides a quick introduction to the validation criterion in privacy-preserving strategies, based on which algorithms can account for and validate the privacy leakage problem during machine learning. The paper then describes different privacy-preserving strategies based mainly on federation learning that focus on Differentially Private Federated Averaging and Privacy-Preserving Asynchronous Federated Learning Mechanism and provides an analysis and discussion of their advantages and disadvantages. By improving the original machine learning methods, such as improving the parameter values and limiting the range of features, the possibility of privacy leakage during machine learning is successfully reduced. However, the different privacy-preserving strategies are mainly limited to changing the parameters of the original model training method, which leads to limitations in the training method, such as reduced efficiency or difficulty in training under certain conditions.
Keywords
machine learning, cyber security, artificial intelligence
[1]. ENISA 2022 ENISA Threat Landscape [J/OL] https://www.enisa.europa.eu/publications/enisa-threat-landscape-2022
[2]. Goodfellow I J Shlens J Szegedy C 2014 Explaining and harnessing adversarial examples arXiv preprint arXiv:1412.6572, 2014.
[3]. Biggio B Roli F 2018 Wild patterns: Ten years after the rise of adversarial machine learning Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security pp 2154-2156
[4]. Xue M Yuan C Wu H et al. 2020 Machine learning security: Threats, countermeasures, and evaluations IEEE Access 8 74720-74742
[5]. Li J 2018 Cyber security meets artificial intelligence: a survey Frontiers of Information Technology & Electronic Engineering 19(12) pp 1462-1474
[6]. Carlini N Liu C Erlingsson Ú et al. 2019 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks USENIX Security Symposium 267
[7]. Lyu L He X Li Y 2020 Differentially private representation for nlp: Formal guarantee and an empirical study on privacy and fairness arXiv preprint arXiv:2010.01285
[8]. Ramaswamy S Thakkar O Mathews R et al. 2020 Training production language models without memorizing user data arXiv preprint arXiv:2009.10031
[9]. Xu R Baracaldo N Joshi J 2021 Privacy-preserving machine learning: Methods, challenges and directions arXiv preprint arXiv:2108.04417
[10]. Lu X Liao Y Lio P et al. 2020 Privacy-preserving asynchronous federated learning mechanism for edge network computing IEEE Access 8: pp 48970-48981
[11]. Dwork C Roth A 2014 The algorithmic foundations of differential privacy Foundations and Trends® in Theoretical Computer Science 9(3–4) pp 211-407
[12]. McMahan B Moore E Ramage D et al. 2017 Communication-efficient learning of deep networks from decentralized data Artificial intelligence and statistics PMLR pp 1273-1282
[13]. Kairouz P McMahan H B Avent B et al. 2021 Advances and open problems in federated learning Foundations and Trends® in Machine Learning 14(1–2) pp 1-210
[14]. Shi W Cao J Zhang Q et al. 2016 Edge computing: Vision and challenges IEEE internet of things journal 3(5): pp 637-646
[15]. Wu Y et al. 2017 Big data and computational intelligence in networking CRC Press
[16]. Shokri R Stronati M Song C et al. 2017 Membership inference attacks against machine learning models 2017 IEEE symposium on security and privacy (SP) 2017 3-18
[17]. Salem A Zhang Y Humbert M et al. 2018 Ml-leaks: Model and data independent membership inference attacks and defenses on machine learning models arXiv preprint arXiv:1806.01246
Cite this article
Pan,Z. (2023). Machine learning for privacy-preserving: Approaches, challenges and discussion. Applied and Computational Engineering,18,23-27.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 5th International Conference on Computing and Data Science
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).