Abstract
Semantic segmentation of remote sensing images has important application value in fields such as farmland anomaly detection and urban planning. However, the low-level features extracted by deep neural network models retain rich spatial detail information while introducing redundancy and noise. The significant differences in the semantic level and spatial distribution of high-level and low-level features pose challenges to their effective fusion. To this end, we propose a Multi-Feature Enhancement Fusion Network that improves local feature expression and global semantic modelling ability by fusing edge information and semantic information. The Edge Enhancement Module used traditional edge detection operators to enhance the details of edge features. The Multi-Feature Fusion Module effectively integrates semantic and edge features to enhance the ability to express fine-grained information. The Local-Global Feature Enhancement Module hierarchically establishes local details and global context information, and the Multi-Level Fusion segmentation head integrates the features of different levels to utilise both shallow spatial details and deep semantic information fully. Following this, our extensive experiments on three publicly available datasets demonstrate that the proposed model outperforms state-of-the-art methods. The code will be published on: https://github.com/zwsbh/MFEF.
Similar content being viewed by others
Data availability
The datasets analyzed during this study are available in the following public domains:https://github.com/SHI-Labs/Agriculture-Vision and https://www.isprs.org/resources/datasets/benchmarks/UrbanSemLab/2d-sem-label-vaihingen.aspx.
References
Li, R. et al. Multiattention network for semantic segmentation of fine-resolution remote sensing images. IEEE Transactions on Geosci. Remote. Sens. 60, 1–13 (2021).
Yuan, X., Shi, J. & Gu, L. A review of deep learning methods for semantic segmentation of remote sensing imagery. Expert. Syst. with Appl. 169, 114417 (2021).
Pan, L. et al. M 3-cr: Multi-scale multi-branch mamba for sar-assisted optical image thick cloud removal. IEEE Transactions on Geosci. Remote. Sens. (2025).
Dong, R. et al. High-resolution land cover mapping through learning with noise correction. IEEE Transactions on Geosci. Remote. Sens. 60, 1–13 (2021).
Li, Z., Zhu, Q., Yang, J., Lv, J. & Guan, Q. A cross-domain object-semantic matching framework for imbalanced high spatial resolution imagery water-body extraction. IEEE Transactions on Geosci. Remote. Sens. 62, 1–15 (2024).
Chiu, M. T. et al. Agriculture-vision: A large aerial image database for agricultural pattern analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2828–2838 (2020).
He, S. et al. A distinctive eocene asian monsoon and modern biodiversity resulted from the rise of eastern tibet. Sci. Bull. 67, 2245–2258 (2022).
Zhang, X., Yuan, G., Hua, Z. & Li, J. Tsmga: temporal-spatial multi-scale graph attention network for remote sensing change detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. (2025).
Ji, H., Xie, F., Pan, L., Zheng, Y. & Shi, Z. Huntnet: Homomorphic unified nexus topology for camouflaged object detection. IEEE Transactions on Image Processing (2025).
Ma, X. et al. Sam-assisted remote sensing imagery semantic segmentation with object and boundary constraints. IEEE Transactions on Geosci. Remote. Sens. (2024).
Zhang, X., Dong, K., Cheng, D., Hua, Z. & Li, J. Stwanet: Spatio-temporal wavelet attention aggregation network for remote sensing change detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. (2025).
Pan, L., Zhang, X., Xie, F., Zhang, H. & Zheng, Y. Sgiqa: semantic-guided no-reference image quality assessment. IEEE Transactions on Broadcast. (2024).
Liu, X., Jiao, L., Li, L., Tang, X. & Guo, Y. Deep multi-level fusion network for multi-source image pixel-wise classification. Knowledge-Based Syst. 221, 106921 (2021).
Wu, H., Zhang, M., Huang, P. & Tang, W. Cmlformer: Cnn and multiscale local-context transformer network for remote sensing images semantic segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 17, 7233–7241 (2024).
Long, J., Shelhamer, E. & Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, 3431–3440 (2015).
Gao, Y. et al. Semantic segmentation of remote sensing images based on multiscale features and global information modeling. Expert. Syst. with Appl. 249, 123616 (2024).
Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K. & Yuille, A. L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE transactions on pattern analysis and machine intelligence 40, 834–848 (2017).
Zhao, H., Shi, J., Qi, X., Wang, X. & Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2881–2890 (2017).
Ronneberger, O., Fischer, P. & Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, 234–241 (Springer, 2015).
Li, R. et al. Abcnet: Attentive bilateral contextual network for efficient semantic segmentation of fine-resolution remotely sensed imagery. ISPRS journal of photogrammetry and remote sensing 181, 84–98 (2021).
Yu, B., Yang, L. & Chen, F. Semantic segmentation for high spatial resolution remote sensing images based on convolution neural network and pyramid pooling module. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 11, 3252–3261 (2018).
Zhou, Y., Xia, H., Yu, D., Cheng, J. & Li, J. Outlier detection method based on high-density iteration. Inf. Sci. 662, 120286 (2024).
Vaswani, A. et al. Attention is all you need. Adv. neural information processing systems 30 (2017).
He, X. et al. Swin transformer embedding unet for remote sensing image semantic segmentation. IEEE transactions on geoscience and remote sensing 60, 1–15 (2022).
Wu, H., Huang, P., Zhang, M., Tang, W. & Yu, X. Cmtfnet: Cnn and multiscale transformer fusion network for remote-sensing image semantic segmentation. IEEE Transactions on Geosci. Remote. Sens. 61, 1–12 (2023).
Dosovitskiy, A. et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020).
Liu, Z. et al. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, 10012–10022 (2021).
Xie, E. et al. Segformer: Simple and efficient design for semantic segmentation with transformers. Advances in neural information processing systems 34, 12077–12090 (2021).
Strudel, R., Garcia, R., Laptev, I. & Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, 7262–7272 (2021).
Chen, J. et al. Transunet: Transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021).
Zhang, R., Zhang, Q. & Zhang, G. Lsrformer: Efficient transformer supply convolutional neural networks with global information for aerial image segmentation. IEEE Transactions on Geosci. Remote. Sens. 62, 1–13 (2024).
Zhang, X., Wang, Z., Li, J. & Hua, Z. Mvafg: Multiview fusion and advanced feature guidance change detection network for remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 17, 11050–11068 (2024).
Gu, A. & Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752 (2023).
Liu, M. et al. Cm-unet: Hybrid cnn-mamba unet for remote sensing image semantic segmentation. arXiv preprint arXiv:2405.10530 (2024).
Ma, X., Zhang, X. & Pun, M.-O. Rs 3 mamba: Visual state space model for remote sensing image semantic segmentation. IEEE Remote. Sens. Lett. 21, 1–5 (2024).
Wang, Z., Zheng, J.-Q., Zhang, Y., Cui, G. & Li, L. Mamba-unet: Unet-like pure visual mamba for medical image segmentation. arXiv preprint arXiv:2402.05079 (2024).
Liu, J. et al. Swin-umamba: Mamba-based unet with imagenet-based pretraining. In International conference on medical image computing and computer-assisted intervention, 615–625 (Springer, 2024).
Li, L., Yi, J., Fan, H. & Lin, H. A lightweight semantic segmentation network based on self-attention mechanism and state space model for efficient urban scene segmentation. IEEE Transactions on Geosci. Remote. Sens. (2025).
Chen, K. et al. Rsmamba: Remote sensing image classification with state space model. IEEE Geosci. Remote. Sens. Lett. 21, 1–5 (2024).
Dong, X. et al. Cswin transformer: A general vision transformer backbone with cross-shaped windows. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 12124–12134 (2022).
Woo, S., Park, J., Lee, J.-Y. & Kweon, I. S. Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV), 3–19 (2018).
Wang, L. et al. Unetformer: A unet-like transformer for efficient semantic segmentation of remote sensing urban scene imagery. ISPRS J. Photogramm. Remote. Sens. 190, 196–214 (2022).
Wang, L. et al. A novel transformer based semantic segmentation scheme for fine-resolution remote sensing images. IEEE Geosci. Remote. Sens. Lett. 19, 1–5 (2022).
Yang, Y., Yuan, G. & Li, J. Sffnet: A wavelet-based spatial and frequency domain fusion network for remote sensing segmentation. IEEE Transactions on Geosci. Remote. Sens. (2024).
Li, R., Wang, L., Zhang, C., Duan, C. & Zheng, S. A2-fpn for semantic segmentation of fine-resolution remotely sensed images. Int. journal of remote sensing 43, 1131–1155 (2022).
Li, R., Zheng, S., Duan, C., Su, J. & Zhang, C. Multistage attention resu-net for semantic segmentation of fine-resolution remote sensing images. IEEE Geosci. Remote. Sens. Lett. 19, 1–5 (2021).
Xiao, P. et al. Mf-mamba: Multi-scale convolution and mamba fusion model for semantic segmentation of remote sensing imagery. IEEE Transactions on Geosci. Remote. Sens. (2025).
Funding
This work is a research achievement supported by the National Key R&D Program of China Major Project (No. 2022ZD0115800) and the National Natural Science Foundation of China (No. 62262065).
Author information
Authors and Affiliations
Contributions
Zhang is responsible for manuscript drafting, review, editing, as well as model design and implementation. Yang is responsible for funding acquisition and supervision. Yin is responsible for manuscript review and editing. Chen is responsible for formal analysis and data curation. Wang, Zhao, are responsible for software design. All authors reviewed the manuscript.
Corresponding authors
Ethics declarations
Competing Interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Zhang, W., Yang, W., Yin, Y. et al. Multi-feature enhancement fusion network for remote sensing image semantic segmentation. Sci Rep (2026). https://doi.org/10.1038/s41598-026-35723-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-026-35723-y


