
Neural Machine Translation (NMT): Deep learning approaches through Neural Network Models
- 1 Shandong University, Shandong, China
* Author to whom correspondence should be addressed.
Abstract
This paper explores the significant advancements in Neural Machine Translation (NMT) models, focusing on the impact of different architectures, training methodologies, and optimization techniques on translation quality. The study contrasts the performance of Recurrent Neural Networks (RNNs), Convolutional Neural Networks (CNNs), and the Transformer model, highlighting the superior capabilities of the Transformer in handling long-range dependencies and providing contextually accurate translations. Key optimization techniques, such as learning rate scheduling, dropout regularization, and gradient clipping, are discussed in detail, emphasizing their roles in enhancing model performance and training efficiency. Furthermore, the paper presents a comparative analysis of NMT and traditional Statistical Machine Translation (SMT) systems, showcasing NMT's superior BLEU scores and fluency. The application of model distillation is also examined, demonstrating how smaller models can achieve high performance with reduced computational resources. These findings underscore the transformative potential of NMT in achieving state-of-the-art translation quality and efficiency.
Keywords
Neural Machine Translation, Transformer Model, Recurrent Neural Networks, Convolutional Neural Networks.
[1]. Ranathunga, Surangika, et al. "Neural machine translation for low-resource languages: A survey." ACM Computing Surveys 55.11 (2023): 1-37.
[2]. Klimova, Blanka, et al. "Neural machine translation in foreign language teaching and learning: a systematic review." Education and Information Technologies 28.1 (2023): 663-682.
[3]. Giovannotti, Patrizio. "Evaluating machine translation quality with conformal predictive distributions." Conformal and Probabilistic Prediction with Applications. PMLR, 2023.
[4]. Fernandes, Patrick, et al. "Scaling laws for multilingual neural machine translation." International Conference on Machine Learning. PMLR, 2023.
[5]. Xu, Hongfei. "Transformer-based NMT: modeling, training and implementation." (2021).
[6]. Reheman, Abudurexiti, et al. "Prompting neural machine translation with translation memories." Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 37. No. 11. 2023.
[7]. Gehring, Jonas, et al. "Convolutional sequence to sequence learning." International conference on machine learning. PMLR, 2017.
Cite this article
Hu,J. (2024). Neural Machine Translation (NMT): Deep learning approaches through Neural Network Models. Applied and Computational Engineering,82,93-99.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 2nd International Conference on Machine Learning and Automation
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).