Abstract
Complete and accurate road network information is an important basis in the detection of EHV transmission lines, and regular updates of road distribution near transmission lines are necessary and meaningful. However, no relevant research has been found for this application area, and coupled with the fact that roads themselves are significantly challenging, extracting roads with good connectivity and integrity in remote sensing images remains a problem to be solved. Therefore, in this paper, we develop a new end-to-end road extraction network, Multiple Attention Networks (MANet). Specifically, by fusing convolutional and self-attentive approaches, we focus on global contextual features to obtain an effective feature map. In addition, the Strip Multi-scale Channel Attention (SMCA) module is specifically designed for the line features of roads, focusing on extracting row and column features, while the Edge-aware Module (EAM) is used to extract connected and complete roads, aided by edge information. Meanwhile, in order to enhance the practicality of the study, a Mengxi Transmission Line Road Dataset was constructed independently following the processing process of remote sensing images in industrial production. By conducting relevant quantitative and qualitative experiments on this dataset and the publicly available CHN6-CUG dataset, it is fully verified that the method in this paper is superior to other advanced methods and can still extract roads with strong connectivity in complex backgrounds, which has good potential and outstanding advantages in practical applications.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Hoeser, T., Kuenzer, C.J.R.S.: Object Detection and Image Segmentation with Deep Learning on Earth Observation Data: A Review-Part i: Evolution and Recent Trends. 12, 1667 (2020)
Mnih, V.: Machine Learning for Aerial Image Labeling. University of Toronto (Canada) (2013)
Wang, J., Song, J., Chen, M., Yang, Z.J.I.J.o.R.S.: Road Network Extraction: A Neural-Dynamic Framework Based on Deep Learning and a Finite State Machine 36, 3144-3169 (2015)
Long, J., Shelhamer, E., Darrell, T.: Fully Convolutional Networks for Semantic Segmentation (2017)
Zhong, Z., Li, J., Cui, W., Han, J.: Fully convolutional networks for building and road extraction: preliminary results. In: 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS) (2016)
Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5–9, 2015, Proceedings, Part III 18, pp. 234–241. Springer (2015)
Cheng, G., Wang, Y., Xu, S., Wang, H., Xiang, S., Pan, C.J.I.T.o.G., Sensing, R.: Automaticroad Detection and Centerline Extraction via Cascaded End-to-end Convolutional Neural Network 55, 3322–3337 (2017)
Zhou, L., Zhang, C., Wu, M.: D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 182–186 (2018)
Shi, Q., Liu, X., Li, X.J.I.a.: Road Detection from Remote Sensing Images by Generative Adversarial Networks 6, 25486–25494 (2017)
Vaswani, A., et al.: Attention is All You Need 30 (2017)
Jiang, X., et al.: Geoinformation: RoadFormer: Pyramidal Deformable Vision Transformers for Road Network Extraction with Remote Sensing Images 113, 102987 (2022)
Zhang, Z., Miao, C., Liu, C., Tian, Q., Zhou, Y.J.M.: HA-RoadFormer: Hybrid Attention Transformer with Multi-Branch for Large-Scale High-Resolution Dense Road Segmentation 10, 1915 (2022)
Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the European conference on computer vision (ECCV), pp. 801–818 (2018)
He, T., Zhang, Z., Zhang, H., Zhang, Z., Xie, J., Li, M.: Bag of tricks for image classification with convolutional neural networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 558–567 (2019)
Pan, X., et al.: On the integration of self-attention and convolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 815–825 (2022)
Sun, Y., Chen, G., Zhou, T., Zhang, Y., Liu, N.J.a.p.a.: Context-Aware Cross-Level Fusion Network for Camouflaged Object Detection (2021)
Huan, H., Sheng, Y., Zhang, Y., Liu, Y.J.R.S.: Strip Attention Networks for Road Extraction 14, 4516 (2022)
Zhu, Q., et al.: A Global Context-Aware and Batch-Independent Network for Road Extraction from VHR Satellite Imagery 175, 353–365 (2021)
MMSegmentation contributors. MMSegmentation: Openmmlab Semantic Segmentation Toolbox and Benchmark (2020). https://github.com/openmmlab/mmsegmentation. Accessed 11 Aug 2020
Bottou, L.: Large-scale machine learning with stochastic gradient descent. In: Proceedings of COMPSTAT’2010: 19th International Conference on Computational StatisticsParis France, August 22–27, 2010 Keynote, Invited and Contributed Papers, pp. 177–186. Springer (2010)
Abdollahi, A., Pradhan, B., Shukla, N., Chakraborty, S., Alamri, A.J.R.S.: Deep Learning Approaches Applied to Remote Sensing Datasets for Road Extraction: A State-of-the-Art Review 12, 1444 (2020)
Xiao, T., Liu, Y., Zhou, B., Jiang, Y., Sun, J.: Unified perceptual parsing for scene understanding. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 418–434 (2018)
Liu, Z., et al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10012–10022 (2021)
Fu, J., et al.: Dual attention network for scene segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3146–3154 (2019)
Dosovitskiy, A., et al.: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. (2020)
Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132–7141 (2018)
Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: ECA-Net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11534–11542 (2020)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19 (2018)
Acknowledgments
The authors gratefully acknowledge the financial supports by the National Natural Science Foundation of China under Grant No. 62077032, as well as the Inner Mongolia Science and Technology Plan Project under project No. 2021GG0159.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ren, Y., Bai, X., Han, Y., Hu, X. (2023). MANet: An End-To-End Multiple Attention Network for Extracting Roads Around EHV Transmission Lines from High-Resolution Remote Sensing Images. In: Yang, X., et al. Advanced Data Mining and Applications. ADMA 2023. Lecture Notes in Computer Science(), vol 14176. Springer, Cham. https://doi.org/10.1007/978-3-031-46661-8_37
Download citation
DOI: https://doi.org/10.1007/978-3-031-46661-8_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-46660-1
Online ISBN: 978-3-031-46661-8
eBook Packages: Computer ScienceComputer Science (R0)