
ChatGPT: Technology Frontiers and Cybersecurity Challenges
- 1 School of software, Taiyuan University of Technology, Jinzhong, 030600, China
- 2 School of Computer Science and Technology, Anhui University, Hefei, 230601, China
- 3 Aliyun School of Big Data, Changzhou University, Changzhou 213000, China
- 4 Faculty of Engineering, Bristol, The University of Bristol, Bristol BS8 1QU, United Kingdom
* Author to whom correspondence should be addressed.
Abstract
This paper examines the evolution and application of OpenAI's advanced conversational AI, ChatGPT, particularly within the domain of cybersecurity. With an architecture built on the Transformer model, ChatGPT demonstrates significant capabilities in language understanding and generation. It leverages vast datasets, ranging from social media posts to technical documents, ensuring the model adapts to diverse fields and maintains compliance with privacy and security regulations. The paper explores ChatGPT's role in network security, highlighting its proficiency in threat detection, vulnerability assessment, and incident response, essential as regulations like GDPR and CCPA become more stringent. Furthermore, the study addresses potential security risks associated with AI, such as phishing and misinformation, and discusses mitigation strategies through advanced training techniques like adversarial training and multi-task learning. A novel variational autoencoder (VAE)-based method, T-VAE, is introduced, offering enhanced generalization capabilities across different tasks and scenarios. The findings suggest that while ChatGPT has made significant strides in cybersecurity applications, continuous improvements in model robustness and adaptability are necessary to mitigate emerging threats and adapt to evolving digital landscapes.
Keywords
ChatGPT, Cybersecurity, Proficiency.
[1]. Barkan A, Zhao Z, Wang J 2021 Gradient-based methods for explaining transformer decisions Journal of Machine Learning Research 22(57) 1–30
[2]. Brown T, Mann B, Ryder N, Subbiah M, Kaplan J, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A, Agarwal S, Herbert-Voss A, Krueger G, Henighan T, Child R, Ramesh A, Ziegler D, Wu J, Winter C, Amodei D 2020 Language models are few-shot learners Advances in Neural Information Processing Systems 33 1877–1901
[3]. Carlini N, Liu C, Erlingsson Ú, Kos J, Song D 2021 Extracting training data from large language models Proceedings of the 2021 USENIX Security Symposium
[4]. Che T, Liu J, Zhou Y, Ren J, Zhou J, Sheng V, Dai H 2023 Federated learning of large language models with parameter-efficient prompt tuning and adaptive optimization Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
[5]. Chen M, Tworek J, Jun H, Yuan Q, de Oliveira Pinto H P, Kaplan J, Edwards H, Burda Y, Joseph N, Brockman G, Ray A, Puri R, Krueger G, Petrov M, Gauthier J, Plappert M, Brundage M, Clark J, Ziegler D 2021 Evaluating large language models trained on code arXiv preprint arXiv:2107.03374
[6]. Cui Y, Zhang Z, Zhou J 2022 Backdoor attacks on large language models: A survey and defense strategies Proceedings of the 2022 International Conference on Computational Intelligence and Security
[7]. Fantozzi P, Naldi M 2024 The explainability of transformers: Current status and directions Computers 13(4) 92
[8]. Federated LLM Position Paper 2023 Federated large language model: A position paper arXiv preprint arXiv:2307.08925
[9]. Huang P, Liu Q, Wu T 2022 Privacy-preserving federated learning in large language models Journal of Privacy and Data Protection 14(2) 120–133
[10]. Jin Y, Dobry A, Wang L 2023 Advances in large language models for healthcare Artificial Intelligence Review 36(5) 789–805
[11]. Kim G, Yoo J, Kang S 2023 Efficient federated learning with pre-trained large language model using several adapter mechanisms Mathematics 11(21) 4479
[12]. Lin C-Y 2004 ROUGE: A package for automatic evaluation of summaries Text Summarization Branches Out 74–81
[13]. Papineni K, Roukos S, Ward T, Zhu W J 2002 BLEU: A method for automatic evaluation of machine translation Proceedings of the 40th Annual Meeting on Association for Computational Linguistics 311–318
[14]. Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I 2019 Language models are unsupervised multitask learners OpenAI Blog 1(8) 9
[15]. Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, Zhou Y, Li W, Liu P J 2020 Exploring the limits of transfer learning with a unified text-to-text transformer Journal of Machine Learning Research 21(140) 1–67
[16]. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A N, Kaiser Ł, Polosukhin I 2017 Attention is all you need Advances in Neural Information Processing Systems 30 5998–6008
[17]. Zhang Y, Sun S, Galley M, Chen Y-C, Brockett C, Gao X, Gao J, Dolan B 2020 Dialogpt: Large-scale generative pre-training for conversational response generation Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations 270–278
[18]. Zhao Z, Wang H, Liu Z 2023 Efficient computation and green AI in large language models Proceedings of the ACM Conference on AI and Sustainability
Cite this article
Mao,X.;Xu,S.;Yang,G.;Yang,Y. (2024). ChatGPT: Technology Frontiers and Cybersecurity Challenges. Applied and Computational Engineering,110,65-70.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of CONF-MLA 2024 Workshop: Securing the Future: Empowering Cyber Defense with Machine Learning and Deep Learning
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).