1. Introduction
Typhoons, as intense tropical cyclones occurring in the Western Pacific, pose severe threats to coastal regions and maritime activities. These weather systems are characterized by extremely strong winds, heavy precipitation, and complex three-dimensional structures including the eye, eyewall, and spiral rainbands [1]. The formation and evolution of typhoons involve complex atmospheric dynamics, making accurate prediction and monitoring challenging yet essential for disaster preparedness and mitigation.
Traditional typhoon analysis has relied heavily on numerical weather prediction models, statistical methods, and subjective interpretation techniques such as the Dvorak method [2]. While these approaches have provided valuable insights, they often suffer from limitations including computational constraints, subjective bias, and insufficient utilization of the increasing volume of satellite observations [3]. The emergence of deep learning technologies, combined with the proliferation of multi-modal remote sensing data, offers unprecedented opportunities to enhance typhoon monitoring capabilities [4].
Recent advances in satellite technology have provided access to high-resolution, multi-temporal observations from various sensors including visible/infrared imagers, microwave radiometers, and scatterometers . The integration of these diverse data sources through sophisticated deep learning frameworks enables more comprehensive understanding of typhoon characteristics and behavior. This technological evolution has sparked significant interest in developing automated, objective, and accurate methods for typhoon detection, tracking, and intensity estimation [5].
Research theme and questions—This review examines how deep learning combined with multimodal satellite data supports three operational tasks: (1) detection and structural delineation, (2) tracking/nowcasting, and (3) intensity estimation.
2. Multi-modal remote sensing data for typhoon monitoring
2.1. Satellite-based observations
Modern typhoon monitoring relies extensively on satellite-based remote sensing systems that provide continuous, global coverage of atmospheric conditions. Geostationary satellites such as Himawari-8/9, GOES-16/17, and Fengyun-4 series offer high temporal resolution observations (typically 10-15 minutes) that are crucial for tracking rapidly evolving typhoon systems [6]. These satellites carry advanced imaging instruments capable of capturing data across multiple spectral channels, from visible light to thermal infrared wavelengths.
The Advanced Baseline Imager (ABI) on GOES-16/17 and the Advanced Himawari Imager (AHI) on Himawari-8/9 represent significant technological advances, providing 16 spectral channels with spatial resolutions ranging from 0.5 to 2 kilometers [7]. These high-resolution observational data enable detailed analysis of typhoon structures, encompassing key features such as eyewall dynamics, convective patterns, and cloud-top characteristics—all of which are critical for accurate typhoon intensity assessment.
Polar-orbiting satellites complement geostationary observations by providing higher spatial resolution data and access to additional spectral regions. The Moderate Resolution Imaging Spectroradiometer (MODIS) aboard Terra and Aqua satellites offers detailed visible and infrared imagery [8], while the Visible Infrared Imaging Radiometer Suite (VIIRS) on Suomi NPP and NOAA-20 provides enhanced low-light imaging capabilities for nighttime typhoon monitoring .
2.2. Microwave remote sensing
Microwave sensors play a crucial role in typhoon observation due to their ability to penetrate clouds and provide information about precipitation, atmospheric temperature and humidity profiles, and ocean surface conditions. The Advanced Microwave Scanning Radiometer 2 (AMSR2) aboard GCOM-W1 provides valuable data on sea surface temperature, precipitation rates, and atmospheric water vapor content [9].
Scatterometer data, derived from instruments such as the Advanced Scatterometer (ASCAT) and Oceansat Scatterometer (OSCAT), enables direct measurements of ocean surface wind vectors—information that is indispensable for precise typhoon analysis. These measurements complement model-derived wind fields and provide ground truth for validation of intensity estimates derived from other sensors.
The Global Precipitation Measurement (GPM) mission, successor to the Tropical Rainfall Measuring Mission (TRMM), provides critical precipitation observations through its dual-frequency precipitation radar and multichannel microwave radiometer [10]. These observations are particularly valuable for understanding typhoon structure and intensity changes, as precipitation patterns are closely linked to storm dynamics.
3. Deep learning approaches for typhoon detection
3.1. Convolutional Neural Networks for feature extraction
Convolutional Neural Networks (CNNs) have emerged as the dominant approach for automated typhoon detection from satellite imagery. Unlike traditional methods that rely on hand-crafted features and threshold-based algorithms, CNNs can automatically learn hierarchical representations of typhoon characteristics directly from raw satellite data.
Early applications of CNNs to typhoon detection adapted existing computer vision architectures such as AlexNet, VGG, and ResNet [11]. These networks demonstrated superior performance compared to traditional methods in identifying typhoon presence and location within satellite images. The ability of CNNs to capture multi-scale features makes them particularly well-suited for typhoon analysis, as these systems exhibit characteristic patterns across different spatial scales.
More recent developments have focused on designing specialized CNN architectures optimized for meteorological applications [12]. These custom networks often incorporate domain knowledge about typhoon structure and behavior, leading to improved detection accuracy and reduced false positive rates. Attention mechanisms have been integrated into CNN architectures to focus on the most relevant regions of satellite imagery, such as the typhoon eye and spiral rainbands.
3.2. Object detection frameworks
The evolution of object detection frameworks has provided powerful tools for typhoon identification and localization. Two-stage detectors such as R-CNN, Fast R-CNN, and Faster R-CNN have been successfully applied to typhoon detection tasks, offering high accuracy in both detection and localization [13]. These methods first generate region proposals and then classify and refine the detected objects.
Single-stage detectors, such as You Only Look Once (YOLO) and Single Shot MultiBox Detector (SSD), exhibit distinct advantages in computational efficiency while sustaining competitive levels of accuracy [14]. These frameworks are particularly valuable for operational applications where real-time processing is required. The RetinaNet architecture, with its focal loss function, has shown promising results in addressing the class imbalance problem common in typhoon detection tasks .
Recent advances in transformer-based architectures, such as DETR (Detection Transformer), represent a paradigm shift in object detection methodology [15]. These approaches treat object detection as a direct set prediction problem, potentially offering advantages in handling complex typhoon structures and multiple storm systems within a single image.
3.3. Semantic segmentation for typhoon structure analysis
Semantic segmentation techniques enable pixel-level classification of satellite imagery, providing detailed information about typhoon structure and components. U-Net, originally developed for biomedical image segmentation, has been successfully adapted for typhoon analysis due to its ability to preserve spatial details while capturing global context [16].
The DeepLab series of networks, particularly DeepLabv3+ with its encoder-decoder architecture and atrous convolution, has demonstrated excellent performance in segmenting typhoon components such as the eye, eyewall, and rainbands. These detailed segmentations provide valuable input for subsequent intensity estimation and structure analysis.
4. Temporal modeling for typhoon tracking and prediction
4.1. Recurrent Neural Networks for sequential analysis
Typhoon tracking and intensity prediction inherently involve temporal sequences, making Recurrent Neural Networks (RNNs) and their variants particularly suitable for these tasks. Long Short-Term Memory (LSTM) networks address the vanishing gradient problem in traditional RNNs and can effectively model long-term dependencies in typhoon evolution .
Gated Recurrent Units (GRUs) offer a simplified alternative to LSTMs while maintaining comparable performance in many applications [17]. The reduced computational complexity of GRUs makes them attractive for operational forecasting systems where computational efficiency is crucial.
Bidirectional RNNs leverage both past and future information in the sequence, which can be valuable for typhoon analysis when the complete temporal sequence is available. This approach proves particularly valuable for typhoon reanalysis studies and for advancing the understanding of typhoon development processes.
4.2. Spatio-temporal fusion models
The combination of spatial and temporal modeling capabilities has led to the development of hybrid architectures that can simultaneously capture typhoon structure and evolution. ConvLSTM networks integrate convolutional operations within LSTM cells, enabling the model to learn spatio-temporal representations directly from satellite image sequences [18].
Three-dimensional CNNs provide another approach to spatio-temporal modeling by treating time as an additional dimension . These networks can capture temporal relationships while preserving spatial structure, making them well-suited for analyzing typhoon development patterns.
4.3. Attention mechanisms and transformer architectures
Attention mechanisms have brought about a paradigm shift in sequence modeling, as they enable models to selectively focus on the most relevant segments of the input sequence. In typhoon applications, attention can help models identify critical development phases or important structural features that influence future evolution.
Transformer architectures, based entirely on attention mechanisms, have shown remarkable success in various sequence modeling tasks. For typhoon prediction, transformers can model complex temporal dependencies and capture global relationships in the data without the limitations of recurrent architectures [19].
5. Multi-modal data fusion strategies
5.1. Early and late fusion approaches
Multi-modal data fusion represents a critical aspect of modern typhoon monitoring systems. Early fusion approaches concatenate features from different sensors at the input level, allowing deep learning models to learn joint representations across modalities [20]. This approach is computationally efficient but may not optimally exploit the unique characteristics of each data source.
Late fusion strategies combine predictions from separate models trained on individual modalities. While this approach allows for specialized processing of each data type, it may miss important cross-modal interactions that could improve overall performance.
5.2. Cross-modal attention and adaptive fusion
Advanced fusion architectures employ cross-modal attention mechanisms that enable different modalities to influence each other's feature representations. These approaches can automatically learn the relative importance of different data sources for specific tasks and conditions.
Adaptive fusion networks adjust the contribution of different modalities based on data quality, availability, and relevance to the current prediction task [21]. This flexibility is particularly valuable in operational settings where sensor data may be incomplete or of varying quality.
6. Challenges and future directions
6.1. Current technical challenges
Despite significant progress, several challenges remain in applying deep learning to typhoon monitoring. Data quality issues, including cloud contamination, sensor calibration differences, and temporal coverage gaps, continue to affect model performance [22]. The relatively small number of extreme typhoon events in training datasets poses challenges for developing robust models that can handle rare but critical cases.
Model interpretability remains a significant concern for operational applications, where forecasters need to understand and trust model predictions. The "black box" nature of deep learning models can hinder acceptance and integration into existing forecasting workflows.
6.2. Emerging technologies and methodologies
Self-supervised learning approaches show promise for reducing dependence on labeled training data by learning useful representations from unlabeled satellite observations [23]. This is particularly valuable given the extensive archives of satellite data that lack detailed annotations.
Few-shot learning techniques could address the challenge of rare typhoon events by enabling models to learn from limited examples [21]. Transfer learning approaches that leverage knowledge from related meteorological phenomena or different geographical regions may also improve model robustness (Yosinski et al., 2014).
Physics-informed neural networks represent an exciting direction that incorporates physical constraints and meteorological knowledge directly into deep learning architectures [22]. These approaches could improve model reliability and extrapolation capabilities beyond training data distributions.
7. Conclusion
The application of deep learning to typhoon monitoring using multi-modal remote sensing data represents a significant advancement in meteorological science and operational forecasting. These technologies have demonstrated superior performance in automated feature extraction, temporal modeling, and multi-source data integration compared to traditional methods.
Key achievements include the development of robust CNN architectures for typhoon detection and structure analysis, sophisticated temporal models for tracking and prediction, and effective fusion strategies for integrating diverse observation sources. These advances have improved the accuracy, objectivity, and computational efficiency of typhoon monitoring systems.
However, challenges remain in areas such as model interpretability, handling of rare events, and integration with existing operational frameworks. Future research should focus on developing more robust and interpretable models, incorporating physical constraints, and advancing real-time processing capabilities.
The continued evolution of satellite technology, computational resources, and deep learning methodologies promises further improvements in typhoon monitoring and prediction capabilities. As these technologies mature, they will play an increasingly important role in protecting lives and property from typhoon-related hazards, contributing to more resilient and prepared coastal communities worldwide.
References
[1]. Emanuel, K. (2003). Tropical cyclones. Annual Review of Earth and Planetary Sciences, 31(1), 75-104.
[2]. Velden, C., Harper, B., Wells, F., Beven II, J. L., Zehr, R., Olander, T., ... & McCrone, P. (2006). The Dvorak tropical cyclone intensity estimation technique: A satellite-based method that has endured for over 30 years. Bulletin of the American Meteorological Society, 87(9), 1195-1210.
[3]. Kossin, J. P., Knapp, K. R., Vimont, D. J., Murnane, R. J., & Harper, B. A. (2007). A globally consistent reanalysis of hurricane variability and trends. Geophysical Research Letters, 34(4).
[4]. Reichstein, M., Camps-Valls, G., Stevens, B., Jung, M., Denzler, J., Carvalhais, N., & Prabhat. (2019). Deep learning and process understanding for data-driven Earth system science. Nature, 566(7743), 195-204.
[5]. Lee, J., Im, J., Cha, D. H., Park, H., & Sim, S. (2019). Tropical cyclone intensity estimation based on satellite infrared images using deep learning. Remote Sensing, 11(9), 1032.
[6]. Schmit, T. J., Griffith, P., Gunshor, M. M., Daniels, J. M., Goodman, S. J., & Lebair, W. J. (2017). A closer look at the ABI on the GOES-R series. Bulletin of the American Meteorological Society, 98(4), 681-698.
[7]. Schmit, T. J., Gunshor, M. M., Menzel, W. P., Gurka, J. J., Li, J., & Bachmeier, A. S. (2005). Introducing the next-generation advanced baseline imager on GOES-R. Bulletin of the American Meteorological Society, 86(8), 1079-1096.
[8]. King, M. D., Menzel, W. P., Kaufman, Y. J., Tanré, D., Gao, B. C., Platnick, S., ... & Hubanks, P. A. (2003). Cloud and aerosol properties, precipitable water, and profiles of temperature and water vapor from MODIS. IEEE Transactions on Geoscience and Remote Sensing, 41(2), 442-458.
[9]. Imaoka, K., Kachi, M., Kasahara, M., Ito, N., Nakagawa, K., & Oki, T. (2010). Instrument performance and calibration of AMSR-E and AMSR2. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, 38(8), 13-16.
[10]. Skofronick-Jackson, G., Petersen, W. A., Berg, W., Kidd, C., Stocker, E. F., Kirschbaum, D. B., ... & Wilheit, T. (2017). The Global Precipitation Measurement (GPM) mission for science and society. Bulletin of the American Meteorological Society, 98(8), 1679-1695.
[11]. Pradhan, R., Aygun, R. S., Maskey, M., Ramachandran, R., & Cecil, D. J. (2018). Tropical cyclone intensity estimation using a deep convolutional neural network. IEEE Transactions on Image Processing, 27(2), 692-702.
[12]. Chen, R., Wang, X., Zhang, W., Zhu, X., Li, A., & Yang, C. (2019). A hybrid CNN-LSTM model for typhoon formation forecasting. Geoinformatica, 23(3), 375-396.
[13]. Girshick, R. (2015). Fast r-cnn. In Proceedings of the IEEE international conference on computer vision (pp. 1440-1448).
[14]. Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788).
[15]. Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., & Zagoruyko, S. (2020). End-to-end object detection with transformers. In European conference on computer vision (pp. 213-229). Springer.
[16]. Zhang, L., Liu, Y., & Liu, Y. (2019). Improved U-Net for building extraction from high resolution remote sensing imagery. In 2019 IEEE International Geoscience and Remote Sensing Symposium (pp. 3187-3190). IEEE.
[17]. Cho, K., Van Merriënboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., & Bengio, Y. (2014). Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv: 1406.1078.
[18]. Xingjian, S. H. I., Chen, Z., Wang, H., Yeung, D. Y., Wong, W. K., & Woo, W. C. (2015). Convolutional LSTM network: A machine learning approach for precipitation nowcasting. Advances in Neural Information Processing Systems, 28, 802-810.
[19]. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: a survey. Philosophical Transactions of the Royal Society A, 379(2194), 20200209.
[20]. Atrey, P. K., Hossain, M. A., El Saddik, A., & Kankanhalli, M. S. (2010). Multimodal fusion for multimedia analysis: a survey. Multimedia Systems, 16(6), 345-379.
[21]. Wang, Y., Yao, Q., Kwok, J. T., & Ni, L. M. (2020). Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys, 53(3), 1-34.
[22]. Zhang, W., Jiang, L., Chen, P., & Zhang, L. (2021). A survey on deep learning based typhoon trajectory and intensity prediction. Neurocomputing, 458, 301-313.
[23]. Jing, L., & Tian, Y. (2020). Self-supervised visual feature learning with deep neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(11), 4037-4058.
Cite this article
Yu,K. (2025). A Review of Typhoon Detection, Tracking and Intensity Estimation Using Deep Learning and Multi-modal Remote Sensing Data. Applied and Computational Engineering,191,52-58.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of CONF-MLA 2025 Symposium: Intelligent Systems and Automation: AI Models, IoT, and Robotic Algorithms
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).
References
[1]. Emanuel, K. (2003). Tropical cyclones. Annual Review of Earth and Planetary Sciences, 31(1), 75-104.
[2]. Velden, C., Harper, B., Wells, F., Beven II, J. L., Zehr, R., Olander, T., ... & McCrone, P. (2006). The Dvorak tropical cyclone intensity estimation technique: A satellite-based method that has endured for over 30 years. Bulletin of the American Meteorological Society, 87(9), 1195-1210.
[3]. Kossin, J. P., Knapp, K. R., Vimont, D. J., Murnane, R. J., & Harper, B. A. (2007). A globally consistent reanalysis of hurricane variability and trends. Geophysical Research Letters, 34(4).
[4]. Reichstein, M., Camps-Valls, G., Stevens, B., Jung, M., Denzler, J., Carvalhais, N., & Prabhat. (2019). Deep learning and process understanding for data-driven Earth system science. Nature, 566(7743), 195-204.
[5]. Lee, J., Im, J., Cha, D. H., Park, H., & Sim, S. (2019). Tropical cyclone intensity estimation based on satellite infrared images using deep learning. Remote Sensing, 11(9), 1032.
[6]. Schmit, T. J., Griffith, P., Gunshor, M. M., Daniels, J. M., Goodman, S. J., & Lebair, W. J. (2017). A closer look at the ABI on the GOES-R series. Bulletin of the American Meteorological Society, 98(4), 681-698.
[7]. Schmit, T. J., Gunshor, M. M., Menzel, W. P., Gurka, J. J., Li, J., & Bachmeier, A. S. (2005). Introducing the next-generation advanced baseline imager on GOES-R. Bulletin of the American Meteorological Society, 86(8), 1079-1096.
[8]. King, M. D., Menzel, W. P., Kaufman, Y. J., Tanré, D., Gao, B. C., Platnick, S., ... & Hubanks, P. A. (2003). Cloud and aerosol properties, precipitable water, and profiles of temperature and water vapor from MODIS. IEEE Transactions on Geoscience and Remote Sensing, 41(2), 442-458.
[9]. Imaoka, K., Kachi, M., Kasahara, M., Ito, N., Nakagawa, K., & Oki, T. (2010). Instrument performance and calibration of AMSR-E and AMSR2. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, 38(8), 13-16.
[10]. Skofronick-Jackson, G., Petersen, W. A., Berg, W., Kidd, C., Stocker, E. F., Kirschbaum, D. B., ... & Wilheit, T. (2017). The Global Precipitation Measurement (GPM) mission for science and society. Bulletin of the American Meteorological Society, 98(8), 1679-1695.
[11]. Pradhan, R., Aygun, R. S., Maskey, M., Ramachandran, R., & Cecil, D. J. (2018). Tropical cyclone intensity estimation using a deep convolutional neural network. IEEE Transactions on Image Processing, 27(2), 692-702.
[12]. Chen, R., Wang, X., Zhang, W., Zhu, X., Li, A., & Yang, C. (2019). A hybrid CNN-LSTM model for typhoon formation forecasting. Geoinformatica, 23(3), 375-396.
[13]. Girshick, R. (2015). Fast r-cnn. In Proceedings of the IEEE international conference on computer vision (pp. 1440-1448).
[14]. Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788).
[15]. Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., & Zagoruyko, S. (2020). End-to-end object detection with transformers. In European conference on computer vision (pp. 213-229). Springer.
[16]. Zhang, L., Liu, Y., & Liu, Y. (2019). Improved U-Net for building extraction from high resolution remote sensing imagery. In 2019 IEEE International Geoscience and Remote Sensing Symposium (pp. 3187-3190). IEEE.
[17]. Cho, K., Van Merriënboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., & Bengio, Y. (2014). Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv: 1406.1078.
[18]. Xingjian, S. H. I., Chen, Z., Wang, H., Yeung, D. Y., Wong, W. K., & Woo, W. C. (2015). Convolutional LSTM network: A machine learning approach for precipitation nowcasting. Advances in Neural Information Processing Systems, 28, 802-810.
[19]. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: a survey. Philosophical Transactions of the Royal Society A, 379(2194), 20200209.
[20]. Atrey, P. K., Hossain, M. A., El Saddik, A., & Kankanhalli, M. S. (2010). Multimodal fusion for multimedia analysis: a survey. Multimedia Systems, 16(6), 345-379.
[21]. Wang, Y., Yao, Q., Kwok, J. T., & Ni, L. M. (2020). Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys, 53(3), 1-34.
[22]. Zhang, W., Jiang, L., Chen, P., & Zhang, L. (2021). A survey on deep learning based typhoon trajectory and intensity prediction. Neurocomputing, 458, 301-313.
[23]. Jing, L., & Tian, Y. (2020). Self-supervised visual feature learning with deep neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(11), 4037-4058.