1. Introduction
Satellite remote sensing technology refers to the use of remote sensing sensors carried by satellites to observe and measure the Earth's surface and to obtain information about the Earth's surface, including topography, geomorphology, vegetation, climate, hydrology and other kinds of information [1,2]. Satellite remote sensing technology has become an indispensable and important means in the fields of national resource management, environmental protection, and disaster monitoring.
Satellite remote sensing image classification refers to the classification of pixel points in satellite remote sensing images according to the category to which they belong, and common classifications include land, water bodies, vegetation, etc [3,4]. Traditional remote sensing image classification methods are mainly based on feature extraction and machine learning algorithms, such as support vector machine (SVM) and decision tree (DT). These methods require manual feature extraction and are less efficient for complex scenes and large-scale data processing.
In recent years, deep learning algorithms have been widely used in satellite remote sensing image classification. Deep learning algorithm is a machine learning algorithm based on neural network structure, which has the advantages of automatic feature extraction and efficient processing of large-scale data [5]. Deep learning algorithms are mainly classified into two categories in satellite remote sensing image classification: methods based on convolutional neural networks (CNN) and methods based on recurrent neural networks (RNN) [6].
CNN-based methods are currently the most widely used deep learning algorithms in satellite remote sensing image classification, and their main idea is to automatically extract image features using multi-layer convolution and pooling operations. For example, feature extraction is performed by feeding satellite remote sensing images into a pre-trained deep convolutional neural network and classification is performed by fully connected layers [7,8]. In addition, there are some improved CNN-based methods, such as null convolution and residual networks, which can better handle complex scenes and large-scale data.
RNN-based methods, on the other hand, are mainly applied to time-series remote sensing data classification, such as meteorological prediction, vegetation growth monitoring and other fields [9]. The main idea is to use recurrent neural network structure to model time-series data and classify them through the output layer [10]. For example, meteorological data are modelled using Long Short-Term Memory (LSTM) networks and classified through fully connected layers.
Deep learning algorithms have great potential in satellite remote sensing image classification, this paper selects remote sensing image public dataset to classify remote sensing images based on a variety of deep learning algorithms, and compares the classification effect of various deep learning algorithms classifiers, to provide a certain foundation for subsequent research, and hope that it will be more widely used in future research.
2. Source of data sets
The dataset used in this paper is from the satellite image classification dataset-RSI-CB256, which is a new large-scale benchmark dataset containing multiple instances for RS image scene classification. The dataset consists of four types of images, namely, cloudy, desert, greenland and water for four scenes, which includes a total of 5631 images, and the dataset is divided into a training set and a validation set according to a ratio of 9:1. Examples of the four types of images (cloudy, desert, green field and water respectively) are shown in Figure 1.




Figure 1. Selected data sets.
(Photo credit: Original)
3. Method
3.1. AlexNet
AlexNet is a deep convolutional neural network that consists of 8 layers of neural networks, 5 of which are convolutional, 2 are fully connected and 1 softmax output layer. The model uses ReLU activation functions and Dropout regularisation to reduce overfitting.AlexNet also uses data enhancement techniques including random cropping, level flipping and colour transformations.
The innovation of AlexNet is the introduction of GPU-accelerated computation and the use of two GPUs for parallel computation. This approach greatly improves training speed and model accuracy.AlexNet also proposes a local response normalisation technique to enhance the contrast of the feature maps.
3.2. VggNet
VggNet is a deep convolutional neural network consisting of 16-19 layers of convolutional neural networks.VggNet is characterised by the use of small size 3x3 convolutional kernels instead of larger size convolutional kernels, which reduces the number of parameters and improves the model's ability to generalise.VggNet also employs multiple pooling layers to reduce the dimensionality of the feature maps, and adds fully-connected layers to achieve classification.
VggNet performed well in the ImageNet competition and has also become one of the classic models in deep learning.VggNet's structure is simple and straightforward, easy to understand and implement, and is therefore widely used in areas such as image classification, target detection, and semantic segmentation.
3.3. GoogleNet
GoogleNet is a deep convolutional neural network consisting of 22 convolutional layers.GoogleNet uses an Inception architecture, whereby convolutional kernels of different sizes are applied in parallel to the input data and stitched together to form a higher dimensional feature map.GoogleNet also uses a 1x1 convolutional kernel to reduce the dimensionality of the feature map, and a global average pooling layer instead of a fully connected layer, thus reducing the number of model parameters.
GoogleNet has achieved excellent results in the ImageNet competition and has proposed several innovative ideas and techniques. The most important of these is the Inception architecture, which drastically reduces the number of model parameters and the amount of computation, and improves model accuracy.
3.4. MobileNet
MobileNet is a lightweight convolutional neural network that drastically reduces model size and computation while maintaining high accuracy.Instead of the traditional convolution operation, MobileNet employs Depthwise Separable Convolution, which decomposes the standard convolution into two steps: depthwise convolution and point-by-point convolution. MobileNet also uses linear bottlenecks and residual concatenation to further improve model accuracy.
The innovation of MobileNet is the introduction of depth-separable convolution, which can significantly reduce the number of model parameters and the amount of computation, so that the model can be run in real time on mobile devices.MobileNet has been widely used in the fields of face recognition, image classification and target detection.
4. Result
The AlexNet, VggNet, GoogleNet and MobileNet models were imported respectively, 90% of the data were randomly selected for training and the remaining 10% for testing, the loss and accuracy changes of the validation set during training were recorded, and the accuracy of the best round of epoch for training was recorded, the batch size is set to 16, the learning rate is set to 0.0002, and the equipment used for the experiment is a 2080 graphics card with 32G RAM.
The changes in the loss and accuracy of the validation set during the training of the AlexNet model are shown in Fig. 2, the changes in the loss and accuracy of the validation set during the training of the VggNet model are shown in Fig. 3, the changes in the loss and accuracy of the validation set during the training of the GoogleNet model are shown in Fig. 4, and the changes in the loss and accuracy of the validation set during the training of the MobileNet model are shown in Fig. 4. The variation of loss and accuracy of validation set during training of MobileNet model is shown in Fig. 5.


Figure 2. Loss and accuracy of the validation set during AlexNet model training.
(Photo credit: Original)


Figure 3. Loss and accuracy of the validation set during training of VggNet models.
(Photo credit: Original)


Figure 4. Loss and accuracy of the validation set during GoogleNet model training.
(Photo credit: Original)


Figure 5. Loss and accuracy of the validation set during MobileNet model training.
(Photo credit: Original)
As can be seen from the change curves of the loss and accuracy of the validation set during the training process of AlexNet, VggNet, GoogleNet and MobileNet models, the loss of the validation set gradually decreases and tends to converge during the training process, and the accuracy of the validation set gradually increases and gradually stabilises.
The epoch with the best training accuracy for each model and the corresponding accuracy are counted, and the results are shown in Table 1 and Figure 6.
Table 1. Indicators for model evaluation.
Model |
Best epoch |
Accuracy(%) |
AlexNet |
15 |
93.3 |
VggNet |
19 |
95.4 |
Googlenet |
19 |
97.9 |
Mobilenet |
20 |
99.8 |

Figure 6. Indicators for model evaluation.
(Photo credit: Original)
From the prediction results of each model, it can be seen that the highest accuracy is the Mobilenet model with 99.8%, followed by the Googlenet model with 97.9%. In addition, the AlexNet and VggNet models also achieved an accuracy of 93.3% and 95.4%, and all four models were able to classify remote sensing images well.
5. Conclusion
In this paper, four deep learning algorithms (AlexNet, VggNet, GoogleNet and MobileNet) are used to classify the remote sensing image dataset obtained by using satellite remote sensing technology and their classification effects are compared. During the training process, the LOSS of the validation set gradually decreases and tends to converge, and the ACCURACY of the validation set gradually increases and stabilises. After comparing the prediction results, we found that the MobileNet model has the highest accuracy of 99.8%, followed by the Googlenet model with an accuracy of 97.9%.The AlexNet and VggNet models are also able to classify remote sensing images well, with accuracies of 93.3% and 95.4%, respectively. Therefore, all four deep learning algorithms can be effectively applied in remote sensing image classification tasks.
References
[1]. Huaxiang S .MBC-Net: long-range enhanced feature fusion for classifying remote sensing images[J].International Journal of Intelligent Computing and Cybernetics,2024,17(1):181-209.
[2]. Long H ,Chen T ,Chen H , et al.Principal space approximation ensemble discriminative marginalized least-squares regression for hyperspectral image classification[J].Engineering Applications of Artificial Intelligence,2024,133(PA):108031-.
[3]. Qiu C ,Zhang X ,Tong X , et al.Few-shot remote sensing image scene classification: Recent advances, new baselines, and future trends[J].ISPRS Journal of Photogrammetry and Remote Sensing,2024,209368-382.
[4]. Meyer F D M ,Gonçalves A J ,Bio F M A .Using Remote Sensing Multispectral Imagery for Invasive Species Quantification: The Effect of Image Resolution on Area and Biomass Estimation[J].Remote Sensing,2024,16(4):
[5]. Xujian Q ,Lei X ,Anxun H , et al.Simplified Multi-head Mechanism for Few-Shot Remote Sensing Image Classification[J].Neural Processing Letters,2024,56(1):
[6]. Zhao X ,Zhang M ,Tao R , et al.Fractional Fourier Image Transformer for Multimodal Remote Sensing Data Classification.[J].IEEE transactions on neural networks and learning systems,2024,35(2):2314-2326.
[7]. Feng S ,Gao M ,Jin X , et al.Fine-grained damage detection of cement concrete pavement based on UAV remote sensing image segmentation and stitching[J].Measurement,2024,226113844-.
[8]. Li H ,Li L ,Zhao L , et al.ResU-Former: Advancing Remote Sensing Image Segmentation with Swin Residual Transformer for Precise Global–Local Feature Recognition and Visual–Semantic Space Learning[J].Electronics,2024,13(2):
[9]. Hongning Q ,Zili L .Building classification extraction from remote sensing images combining hyperpixel and maximum interclass variance[J].International Journal of Image and Data Fusion,2024,15(1):86-103.
[10]. Wang W ,Yi X.A monitoring method of surface vegetation distribution in the Yellow River Basin based on remote sensing image segmentation[J].International Journal of Environmental Technology and Management,2024,27(1-2):37-48.
Cite this article
Ren,S. (2024). Satellite remote sensing image classification based on multiple deep learning algorithms. Applied and Computational Engineering,57,218-223.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 6th International Conference on Computing and Data Science
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).
References
[1]. Huaxiang S .MBC-Net: long-range enhanced feature fusion for classifying remote sensing images[J].International Journal of Intelligent Computing and Cybernetics,2024,17(1):181-209.
[2]. Long H ,Chen T ,Chen H , et al.Principal space approximation ensemble discriminative marginalized least-squares regression for hyperspectral image classification[J].Engineering Applications of Artificial Intelligence,2024,133(PA):108031-.
[3]. Qiu C ,Zhang X ,Tong X , et al.Few-shot remote sensing image scene classification: Recent advances, new baselines, and future trends[J].ISPRS Journal of Photogrammetry and Remote Sensing,2024,209368-382.
[4]. Meyer F D M ,Gonçalves A J ,Bio F M A .Using Remote Sensing Multispectral Imagery for Invasive Species Quantification: The Effect of Image Resolution on Area and Biomass Estimation[J].Remote Sensing,2024,16(4):
[5]. Xujian Q ,Lei X ,Anxun H , et al.Simplified Multi-head Mechanism for Few-Shot Remote Sensing Image Classification[J].Neural Processing Letters,2024,56(1):
[6]. Zhao X ,Zhang M ,Tao R , et al.Fractional Fourier Image Transformer for Multimodal Remote Sensing Data Classification.[J].IEEE transactions on neural networks and learning systems,2024,35(2):2314-2326.
[7]. Feng S ,Gao M ,Jin X , et al.Fine-grained damage detection of cement concrete pavement based on UAV remote sensing image segmentation and stitching[J].Measurement,2024,226113844-.
[8]. Li H ,Li L ,Zhao L , et al.ResU-Former: Advancing Remote Sensing Image Segmentation with Swin Residual Transformer for Precise Global–Local Feature Recognition and Visual–Semantic Space Learning[J].Electronics,2024,13(2):
[9]. Hongning Q ,Zili L .Building classification extraction from remote sensing images combining hyperpixel and maximum interclass variance[J].International Journal of Image and Data Fusion,2024,15(1):86-103.
[10]. Wang W ,Yi X.A monitoring method of surface vegetation distribution in the Yellow River Basin based on remote sensing image segmentation[J].International Journal of Environmental Technology and Management,2024,27(1-2):37-48.