The Current Status and Trends of the Application of Large Language Models in the Field of Education

Research Article
Open access

The Current Status and Trends of the Application of Large Language Models in the Field of Education

Haiyan Shang 1*
  • 1 School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China    
  • *corresponding author 2023210982@stu.hit.edu.cn
Published on 21 February 2025 | https://doi.org/10.54254/2755-2721/2025.21088
ACE Vol.135
ISSN (Print): 2755-273X
ISSN (Online): 2755-2721
ISBN (Print): 978-1-83558-959-5
ISBN (Online): 978-1-83558-960-1

Abstract

In recent years, with the development of Natural Language Processing (NLP), Significant progress has been made in Large Language Models (LLMs) based on Deep Learning algorithms, which have been widely applied in multiple fields. In the field of education, LLMs has shown great potential for application in multiple aspects, such as providing real-time feedback, lowering educational barriers, promoting personalized learning, etc. Researchers gradually realized that using LLMs can effectively address many pain points in current education. This article provides an overview of the related technologies of LLMs, analyzed the application of some technologies in education, proposed the advantages of applying LLMs in education, and analyzed the challenges faced by LLMs in educational applications and the existing response plans, finally discussed the future development trends of LLMs in the field of education. The research shows that Large Language Models have excellent application effects and broad application prospects in education, and are expected to become an important tool to promote the process of education intelligence and personalization in the future education system.

Keywords:

Large Language Models, Education, Artificial Intelligence, ChatGPT

Shang,H. (2025). The Current Status and Trends of the Application of Large Language Models in the Field of Education. Applied and Computational Engineering,135,115-121.
Export citation

1. Introduction

In recent years, LLMs have become an important breakthrough for humans in the field of natural language, and they have received widespread attention as an important means of Natural Language Processing (NLP). For example, OpenAl's GPT series, Google's BERT and PalM models have had a profound impact on the field of Artificial Intelligence(AI) and even various industries through their powerful text generation and understanding capabilities. In the field of education, the introduction of LLMs is gradually changing the traditional teaching methods and promoting innovation in educational technology.

In traditional educational methods, teachers are at the center of the classroom and bear the important responsibility of educating and guiding students. However, with the diversification and personalized development of educational needs, traditional education methods are facing many challenges, such as uneven distribution of educational resources in different regions and difficulty in meeting students' differentiated needs. LLMs as a deep learning based NLP technique, have powerful abilities in understanding and generating natural language through massive data training, personalized learning support and intelligent teaching evaluation can be provided, playing an increasingly important role in the field of teaching. However, the use of LLMs may also bring new challenges, such as security and privacy issues, bias and unfairness issues, etc. How to face these challenges has become a problem that must be overcome for the LLMs to play an important role in the field of education.

Alqahtani et al.'s research suggests that AI can assist educators in providing educational support and personalized learning by analyzing data on student performance and behavior, identifying areas where students may encounter difficulties, and providing personalized improvement suggestions [1]. Cao et proposed an intelligent tutoring system based on a LLMs, aimed at providing personalized feedback based on students' learning progress and understanding. The research results indicate that the system exhibits significant advantages in improving students' autonomous learning ability and learning outcomes [2]. Mizumoto et studied the application of LLMs in automated essay grading and evaluated their reliability and accuracy in grading. The results indicate that the Automated Essay Scoring (AES) based on GPT has certain accuracy and reliability, and can provide valuable support for human evaluation. In addition, the analysis also indicates that utilizing language features can improve the accuracy of scoring [3]. These studies demonstrate various innovative applications of LLMs in education, providing new perspectives for the transformation of the education industry.

In this article, the author will explore the current application status of LLMs from the technical foundation and advantages of LLMs, and explore the challenges and development prospects they face in application. Through in-depth analysis of these aspects, this article aims to analyze the current situation and impact of LLMs in education.

2. Technical Overview of LLMs

2.1. Transformer Architecture and Text Generation

The basic technology of the LLMs - Transform architecture, and its Self-Attention Mechanism provides strong support for language understanding tasks in the field of education. In education, LLMs can understand and process students' language input at multiple levels, such as answering students' questions, correcting language errors, and generating text content. The Self-Attention Mechanism can also capture semantic connections between sentences, paragraphs, and longer texts, allowing the model to efficiently process educational content and support various educational applications including teaching, question answering, and text generation [4,5].

For example, in language learning and writing teaching, LLMs can analyze students' compositions and provide specific feedback suggestions, including optimization of grammar, vocabulary use, sentence structure, etc., to help students improve their writing skills. In addition, LLMs can also identify students' weak links in language learning and provide personalized tutoring advice.

2.2. The Application of Strategies Fine-Tuning and Pre-training in Education

The Pre-training and Fine-Tuning strategies of LLMs have significant advantages in the field of education. Through the Pre-training of massive educational resources and domain specific corpus, LLMs can show strong ability in understanding basic education knowledge, academic research, discipline knowledge, etc. The pre trained LLMs can provide natural language understanding and generation services for different disciplines, such as automatically generating test questions, writing drafts of academic papers, scoring students' answers, etc [6].

For example, the Fine-Tuning LLMs can play an important role in specific disciplines (such as mathematics, history, literature, etc.), and use the teaching materials, questions and answers in specific fields for directional training to help students better understand the knowledge system of the discipline. Based on the intelligent system of LLMs, teachers can use the Fine-Tuning model to effectively correct homework and carry out personalized teaching, so as to further improve teaching efficiency and students' learning outcomes [7].

2.3. Multimodal Ability and Educational Interaction

With the expansion of LLMs in the multimodal field, the rapid development of educational applications also benefits from the cross domain ability of the model. For example, LLMs can process text and image, voice, video and other multimodal data, providing richer interactive ways in classroom teaching. Teachers can use Multimodal LLMs to create more vivid and creative teaching materials, such as generating illustrations through text description, or combining video content for classroom presentation [8].

In the virtual education environment, the Multimodal ability of LLMs can also enhance learners' sense of immersion. For example, students can have a dialogue with the virtual teacher through voice, and carry out interactive learning in the form of images, animation, video and other media. This cross modal education experience will promote students' more comprehensive and profound understanding of learning content.

3. Technical Overview of LLMs

3.1. Advantages of LLMs in Education

LLMs can provide students with 24-hour continuous real-time feedback and guidance, which is difficult to achieve in traditional education. When students finish their homework or ask questions, LLMs can immediately answer or provide feedback to help students correct mistakes in time. Research shows that timely feedback can significantly improve students' learning motivation and achievement [9]. For example, LLMs can not only provide correct answers, but also gradually show the problem-solving process and explain the logic of each step, so as to help students deepen their understanding of knowledge [10].

Taking the assessment of writing assignments as an example, LLMs can automatically analyze the articles submitted by students, identify problems in grammar, structure, logic and other aspects, and provide targeted suggestions for improvement. This kind of instant feedback can not only help students correct errors when they occur, but also encourage them to actively correct and optimize their works. In addition, LLMs can intelligently analyze students' learning trends according to their long-term performance, provide personalized learning reports, and help teachers better understand students' growth trajectory and needs [11].

In addition, LLMs can provide students with personalized counseling experience. When students ask questions, the model can not only adjust the answers according to their previous learning history and current performance, but also guide students to gradually master knowledge by repeatedly asking and answering questions. Compared with the traditional classroom counseling limited by time and manpower, LLMs can provide customized counseling services for each student, greatly enhancing the flexibility and autonomy of learning [12].

3.2. Lowering the Threshold of Educational Resources

The uneven distribution of educational resources is an important challenge for the education system. As a technical tool, LLMs can effectively reduce the threshold of educational resources, especially in areas where educational resources are scarce. Through digitization and automation, LLMs can help a large number of students to obtain high-quality education, no matter where they are. For example, through the intelligent education platform, students can access high-quality teaching content, simulated tests, learning materials, etc. at any time, which are usually generated or assisted by LLMs. This makes the access to educational resources more popular and convenient [13].

In remote areas, the shortage of teachers and educational resources has seriously restricted students' learning opportunities. Through the online education platform and the intelligent tutoring system based on LLMs, students can obtain high-quality educational tutoring and resources even without high-level teachers, helping them fill the gap caused by the imbalance of educational resources [14].

3.3. Promote Personalized Learning

Personalized learning is one of the core issues of education reform, which aims to provide customized learning experience according to students' learning progress and cognitive characteristics. The powerful natural language understanding ability of LLMs makes it have unique advantages in personalized education. LLMs can identify students' mastery of knowledge by analyzing their input, and provide customized learning resources and suggestions according to their needs.

For example, LLMs can intelligently recommend appropriate learning materials according to students' homework performance, or adjust the depth and style of explanation according to students' questions and expressions. Through this dynamic adaptation, LLMs can provide targeted help at different learning stages to help students learn effectively in the "Zone of Proximal Development" [15]. This adaptability is not only limited to the push of knowledge points, but also includes the adjustment of learning strategies and the optimization of learning methods.

In addition, the research shows that LLMs can greatly reduce the burden of teachers in personalized education through its automated interaction, so that teachers can focus more on the depth and creativity of teaching [16].

4. Challenges and Responses

4.1. Data Privacy and Security Issues

LLMs needs a lot of student data for training and optimization, including students' learning behavior, habits, grades and other information. However, the collection and use of these data involve privacy protection issues, especially for students' data. Data leakage or misuse may pose a serious threat to students' personal privacy.

At present, researchers are developing data encryption technology and privacy protection mechanism to ensure the security of student data. For example, Federated Learning is a distributed learning method, which allows the model to be trained on the user side without uploading sensitive data to the server, thus reducing the risk of data leakage [17]. In addition, privacy preserving machine learning technologies (such as differential Privacy) are also applied to ensure that personal information will not be disclosed during model training [18]. These technologies provide theoretical support and technical guarantee for the safe use of LLMs.

4.2. Prejudice and Unfairness

LLMs's training data usually comes from a wide range of Internet content. Therefore, these data may contain biased or unfair views, and the model may inadvertently spread these biases when generating content. For example, LLMS may make unfair inferences or decisions based on factors such as gender, race or social background, which may lead to unfair evaluation or recommendation in educational applications.

Researchers are studying how to reduce and eliminate the bias in LLMs. For example, equity researchers have proposed a variety of de bias techniques, such as filtering training data, adjusting the output of the model using de bias algorithms, or simulating a more fair decision-making process by generating confrontation networks [19]. Some studies also suggest that the risk of model bias can be effectively reduced by introducing diversified training data sources and regularly checking the fairness of the model [20].

4.3. Depth and Accuracy of Language Comprehension

Although LLMs perform well in language understanding and generation, the depth and accuracy of the model are still limited in some complex educational situations. For example, LLMs may fail to provide completely correct answers due to missing information or reasoning errors when processing long and complex articles, especially when involving multi-step reasoning or interdisciplinary knowledge [21].

In order to meet this challenge, the reasoning ability of LLMs is rapidly developed by improving the model architecture. In recent years, Transformer architectures (such as Bert and GPT Series) have made breakthroughs in understanding and generating languages, but the performance of these models in complex reasoning tasks still needs to be improved. Researchers have proposed a scheme combining Graph Neural Networks (GNNS) and knowledge graphs to enhance the model reasoning ability [22]. By introducing external knowledge base and structured information into model training, LLMs can more accurately reason and understand when answering complex questions. In addition, the training of Multi-task Learning model also helps to improve the generalization ability of the model, so that it can deal with various types of educational tasks.

4.4. The Interaction Between Teachers and Students

Although LLMs can provide personalized guidance and feedback, its interaction with students still lacks the emotional resonance and teaching flexibility that human teachers have. In the process of education, emotional support and social interaction are crucial for students' learning motivation and psychological development [23]. Therefore, LLMs' pure machine feedback may not completely replace the role of human teachers, especially in teaching situations that require emotional care and personalized attention.

At present, research in the field of Affective Computing is committed to enabling AI systems to recognize and respond to users' emotional and psychological states [24]. Some studies have begun to explore how to make LLMs not only provide feedback at the knowledge level, but also show more emotional understanding through tone, wording and other ways. In addition, some education platforms try to provide students with a more interactive learning experience by combining LLMs with virtual teaching assistants and online teachers.

5. Conclusion

This paper discusses the application status and trends of the LLMs in education, analyzes the development of the LLMs, the application of the LLMs in education and the challenges they face, the application of the LLMs in the field of education is bringing far-reaching changes to the educational model, learning experience and educational equity. From intelligent tutoring and personalized learning to academic writing and language translation, the LLMs provide unprecedented tools for students and teachers, making educational resources more abundant, flexible and accessible. However, despite its huge potential, researchers still need to be aware of the challenges that may be brought about in the application of technology, including data privacy, model bias and the impact on the role of teachers.

Looking forward to the future, the development of LLMs in education will be more diversified and in-depth. On the one hand, with the continuous iteration and optimization of technology, it is expected that the model will be more intelligent, able to understand and adapt to the personalized learning needs of different students, and provide more accurate and effective learning support. On the other hand, the popularization and development of educational technology may further narrow the education gap between urban and rural areas and regions, and provide more people with high-quality education opportunities.

Future education may not only be the interaction between human teachers and students, but also the collaborative learning between human and AI model. In this mode, teachers will no longer be a single knowledge transmitter, but more likely to become learning guides, critical thinking cultivators and practitioners of moral and humanistic care. As a part of the education system, the LLMs will provide students with customized learning programs. At the same time, it also needs to maintain continuous attention and Reflection on humanistic and ethical issues.

With the continuous progress of technology, researchers can foresee a more intelligent, personalized and fair education future. In this future, the LLMs and artificial intelligence will not only be the provider of learning tools, but also an important driving force for educational innovation and social progress.


References

[1]. Alqahtani, T., Badreldin, H. A., Alrashed, M., Alshaya, A. I., Alghamdi, S. S., bin Saleh, K., ... & Albekairy, A. M. (2023). The emergent role of artificial intelligence, natural learning processing, and large language models in higher education and research. Research in Social and Administrative Pharmacy, 19(8), 1236-1242.

[2]. Cao, C. (2023). Leveraging large language model and story-based gamification in intelligent tutoring system to scaffold introductory programming courses: A design-based research study. arXiv preprint arXiv:2302.12834.

[3]. Mizumoto, A., & Eguchi, M. (2023). Exploring the potential of using an AI language model for automated essay scoring. Research Methods in Applied Linguistics, 2(2), 100050.

[4]. Mo, Y., Qin, H., Dong, Y., Zhu, Z., & Li, Z. (2024). Large language model (llm) ai text generation detection based on transformer deep learning algorithm. arXiv preprint arXiv:2405.06652.

[5]. Vaswani, A. (2017). Attention is all you need. Advances in Neural Information Processing Systems.

[6]. Zoph, B., Ghiasi, G., Lin, T. Y., Cui, Y., Liu, H., Cubuk, E. D., & Le, Q. (2020). Rethinking pre-training and self-training. Advances in neural information processing systems, 33, 3833-3845.

[7]. Han, Z., Gao, C., Liu, J., Zhang, J., & Zhang, S. Q. (2024). Parameter-efficient fine-tuning for large models: A comprehensive survey. arXiv preprint arXiv:2403.14608.

[8]. Yu, W., Yang, Z., Li, L., Wang, J., Lin, K., Liu, Z., ... & Wang, L. (2023). Mm-vet: Evaluating large multimodal models for integrated capabilities. arXiv preprint arXiv:2308.02490.

[9]. Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81-112.

[10]. Liu, S., Zhang, S., & Wei, Z. (2020). A survey on personalized learning: Theories, technologies, and practices. Educational Research Review, 32, 100389.

[11]. Liu, Y., Wang, J., & Chen, Z. (2023). Automated formative feedback for language learners using large language models. Computers & Education, 98, 79-95.

[12]. Roll, I., & Wylie, R. (2016). Learning analytics in education. International Journal of Artificial Intelligence in Education, 26(4), 594-616.

[13]. Harry, A. (2023). Role of AI in Education. Interdiciplinary Journal and Hummanity (INJURITY), 2(3), 260-268.

[14]. Rakha, N. A. (2023). Revolution in learning through digitization: How technology is changing the landscape of education. International Journal of Cyber Law, 1(3).

[15]. Hausfather, S. J. (1996). Vygotsky and schooling: Creating a social context for learning. Action in teacher education, 18(2), 1-10.

[16]. Li, X., Zheng, Y., & Zhang, Y. (2023). AI-driven personalized learning: A framework for learning analytics and adaptive systems. Journal of Educational Technology & Society, 26(3), 21-35.

[17]. McMahan, B., et al. (2017). "Communication-efficient learning of deep networks from decentralized data." Proceedings of AISTATS 2017.

[18]. Dwork, C., et al. (2014). "The Algorithmic Foundations of Differential Privacy." Foundations and Trends in Theoretical Computer Science, 9(3–4), 211-407.

[19]. Zhao, X., et al. (2017). "Reducing gender bias in language models." Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.

[20]. Binns, R. D., et al. (2018). "The impact of fairness on machine learning and its application in education." Journal of Artificial Intelligence in Education, 28(3), 345-368.

[21]. Brown, T. B., et al. (2020). "Language models are few-shot learners." Proceedings of NeurIPS 2020.

[22]. Zhou, L., et al. (2021). "Incorporating knowledge graphs into transformer models for complex question answering." Proceedings of the 2021 Conference on Neural Information Processing Systems.

[23]. Liu, Y., et al. (2020). "Emotion recognition in learning: A survey." Educational Technology & Society, 23(2), 1-15.

[24]. Wang, Y., Song, W., Tao, W., Liotta, A., Yang, D., Li, X., ... & Zhang, W. (2022). A systematic review on affective computing: Emotion models, databases, and recent advances. Information Fusion, 83, 19-52.


Cite this article

Shang,H. (2025). The Current Status and Trends of the Application of Large Language Models in the Field of Education. Applied and Computational Engineering,135,115-121.

Data availability

The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.

Disclaimer/Publisher's Note

The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

About volume

Volume title: Proceedings of the 3rd International Conference on Mechatronics and Smart Systems

ISBN:978-1-83558-959-5(Print) / 978-1-83558-960-1(Online)
Editor:Mian Umer Shafiq
Conference website: https://2025.confmss.org/
Conference date: 16 June 2025
Series: Applied and Computational Engineering
Volume number: Vol.135
ISSN:2755-2721(Print) / 2755-273X(Online)

© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license. Authors who publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See Open access policy for details).

References

[1]. Alqahtani, T., Badreldin, H. A., Alrashed, M., Alshaya, A. I., Alghamdi, S. S., bin Saleh, K., ... & Albekairy, A. M. (2023). The emergent role of artificial intelligence, natural learning processing, and large language models in higher education and research. Research in Social and Administrative Pharmacy, 19(8), 1236-1242.

[2]. Cao, C. (2023). Leveraging large language model and story-based gamification in intelligent tutoring system to scaffold introductory programming courses: A design-based research study. arXiv preprint arXiv:2302.12834.

[3]. Mizumoto, A., & Eguchi, M. (2023). Exploring the potential of using an AI language model for automated essay scoring. Research Methods in Applied Linguistics, 2(2), 100050.

[4]. Mo, Y., Qin, H., Dong, Y., Zhu, Z., & Li, Z. (2024). Large language model (llm) ai text generation detection based on transformer deep learning algorithm. arXiv preprint arXiv:2405.06652.

[5]. Vaswani, A. (2017). Attention is all you need. Advances in Neural Information Processing Systems.

[6]. Zoph, B., Ghiasi, G., Lin, T. Y., Cui, Y., Liu, H., Cubuk, E. D., & Le, Q. (2020). Rethinking pre-training and self-training. Advances in neural information processing systems, 33, 3833-3845.

[7]. Han, Z., Gao, C., Liu, J., Zhang, J., & Zhang, S. Q. (2024). Parameter-efficient fine-tuning for large models: A comprehensive survey. arXiv preprint arXiv:2403.14608.

[8]. Yu, W., Yang, Z., Li, L., Wang, J., Lin, K., Liu, Z., ... & Wang, L. (2023). Mm-vet: Evaluating large multimodal models for integrated capabilities. arXiv preprint arXiv:2308.02490.

[9]. Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81-112.

[10]. Liu, S., Zhang, S., & Wei, Z. (2020). A survey on personalized learning: Theories, technologies, and practices. Educational Research Review, 32, 100389.

[11]. Liu, Y., Wang, J., & Chen, Z. (2023). Automated formative feedback for language learners using large language models. Computers & Education, 98, 79-95.

[12]. Roll, I., & Wylie, R. (2016). Learning analytics in education. International Journal of Artificial Intelligence in Education, 26(4), 594-616.

[13]. Harry, A. (2023). Role of AI in Education. Interdiciplinary Journal and Hummanity (INJURITY), 2(3), 260-268.

[14]. Rakha, N. A. (2023). Revolution in learning through digitization: How technology is changing the landscape of education. International Journal of Cyber Law, 1(3).

[15]. Hausfather, S. J. (1996). Vygotsky and schooling: Creating a social context for learning. Action in teacher education, 18(2), 1-10.

[16]. Li, X., Zheng, Y., & Zhang, Y. (2023). AI-driven personalized learning: A framework for learning analytics and adaptive systems. Journal of Educational Technology & Society, 26(3), 21-35.

[17]. McMahan, B., et al. (2017). "Communication-efficient learning of deep networks from decentralized data." Proceedings of AISTATS 2017.

[18]. Dwork, C., et al. (2014). "The Algorithmic Foundations of Differential Privacy." Foundations and Trends in Theoretical Computer Science, 9(3–4), 211-407.

[19]. Zhao, X., et al. (2017). "Reducing gender bias in language models." Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.

[20]. Binns, R. D., et al. (2018). "The impact of fairness on machine learning and its application in education." Journal of Artificial Intelligence in Education, 28(3), 345-368.

[21]. Brown, T. B., et al. (2020). "Language models are few-shot learners." Proceedings of NeurIPS 2020.

[22]. Zhou, L., et al. (2021). "Incorporating knowledge graphs into transformer models for complex question answering." Proceedings of the 2021 Conference on Neural Information Processing Systems.

[23]. Liu, Y., et al. (2020). "Emotion recognition in learning: A survey." Educational Technology & Society, 23(2), 1-15.

[24]. Wang, Y., Song, W., Tao, W., Liotta, A., Yang, D., Li, X., ... & Zhang, W. (2022). A systematic review on affective computing: Emotion models, databases, and recent advances. Information Fusion, 83, 19-52.