Remote Sensing Image Segmentation Methods Based on Deep Learning Models
DOI:
https://doi.org/10.71222/15ra5x93Keywords:
remote sensing, semantic segmentation, U-Net, SegNet, DeepLabv3+, self-supervised learning, model fusion, transformer, high-resolution imagery, domain adaptationAbstract
Remote sensing image segmentation plays a pivotal role in Earth observation tasks by transforming raw satellite and aerial imagery into meaningful semantic regions. This process underpins numerous applications, such as urban planning, precision agriculture, disaster response, and ecological monitoring. With the advent of deep learning, segmentation accuracy has improved significantly due to the capacity of neural networks to learn complex spatial and semantic representations. This paper presents a comprehensive comparative study of three representative deep learning models — U-Net, SegNet, and DeepLabv3+ — applied to the ISPRS Potsdam dataset. We analyze performance across various dimensions, including segmentation accuracy, efficiency, robustness to noise, parameter complexity, and category-specific behaviors. Furthermore, we propose a hybrid model architecture that fuses U-Net’s spatial detail preservation with DeepLab’s contextual aggregation capabilities. To address label scarcity and enhance generalization, we incorporate self-supervised pretraining and transfer learning strategies. We also provide preliminary benchmarking with Transformer-based models. The findings contribute to the body of knowledge guiding the design and deployment of segmentation models in real-world remote sensing scenarios.
References
1. Y. Liu, W. Chen, and Y. Xu, "A lightweight transformer network for remote sensing image segmentation," Remote Sens., vol. 15, no. 8, 2023, doi: 10.3390/rs15082023.
2. Z. Li, X. Huang, and R. Wang, "An improved DeepLabv3+ model for semantic segmentation of high-resolution remote sensing images," Sensors, vol. 23, no. 5, p. 2456, 2023, doi: 10.3390/s23052456.
3. H. Zhang and Y. Sun, "Self-supervised contrastive learning for remote sensing semantic segmentation," ISPRS J. Photogramm. Remote Sens., vol. 198, pp. 72–85, 2023, doi: 10.1016/j.isprsjprs.2023.01.010.
4. K. Yamazaki, T. Hanyu, and M. Tran, "AerialFormer: Multi-resolution Transformer for aerial image segmentation," arXiv preprint arXiv:2306.06842, 2023,doi: 10.48550/arXiv.2306.06842.
5. L. Chen, W. Lu, and J. Zhao, "A cross-scale attention-guided CNN-Transformer hybrid network for remote sensing image segmentation," IEEE Geosci. Remote Sens. Lett., vol. 19, pp. 1–5, 2022, doi: 10.1109/LGRS.2022.3143046.
6. F. Gao, C. Zhang, and M. Li, "A semi-supervised learning approach for land cover classification from remote sensing imagery," Int. J. Appl. Earth Obs. Geoinf., vol. 122, p. 103336, 2023, doi: 10.1016/j.jag.2023.103336.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Hongyun Mao, John Lazaro (Author)

This work is licensed under a Creative Commons Attribution 4.0 International License.