Frequency-aware robust multidimensional information fusion framework for remote sensing image segmentation

被引:8
|
作者
Fan, Junyu [1 ]
Li, Jinjiang [2 ]
Liu, Yepeng [2 ]
Zhang, Fan [2 ]
机构
[1] Shandong Technol & Business Univ, Inst Network Technol ICT, Sch Informat & Elect Engn, Yantai 264005, Peoples R China
[2] Shandong Technol & Business Univ, Sch Comp Sci & Technol, Yantai 264005, Peoples R China
基金
中国国家自然科学基金;
关键词
Remote sensing; Semantic segmentation; Frequency information; Transformer; SEMANTIC SEGMENTATION; NETWORK;
D O I
10.1016/j.engappai.2023.107638
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Urban scene image segmentation is an important research area in high-resolution remote sensing image processing. However, due to its complex three-dimensional structure, interference factors such as occlusion, shadow, intra-class inconsistency, and inter-class indistinction affect segmentation performance. Many methods have combined local and global information using CNNs and Transformers to achieve high performance in remote sensing image segmentation tasks. However, these methods are not stable when dealing with these interference factors. Recent studies have found that semantic segmentation is highly sensitive to frequency information, so we introduced frequency information to make the model learn more comprehensively about different categories of targets from multiple dimensions. By modeling the target with local features, global information, and frequency information, the target features can be learned in multiple dimensions to reduce the impact of interference factors on the model and improve its robustness. In this paper, we consider frequency information in addition to combining CNNs and Transformers for modeling and propose a Multidimensional Information Fusion Network (MIFNet) for high-resolution remote sensing image segmentation of urban scenes. Specifically, we design an information fusion Transformer module that can adaptively associate local features, global semantic information, and frequency information and a relevant semantic aggregation module for aggregating features at different scales to construct the decoder. By aggregating image features at different depths, the specific representation of the target and the correlation between targets can be modeled in multiple dimensions, allowing the network to better recognize and understand the features of each class of targets to resist various interference factors that affect segmentation performance. We conducted extensive ablation experiments and comparative experiments on the ISPRS Vaihingen and ISPRS Potsdam benchmarks to verify our proposed method. In a large number of experiments, our method achieved the best results, with 84.53% and 87.3% mIoU scores on the Vaihingen and Potsdam datasets, respectively, proving the superiority of our method. The source code will be available at https://github.com/JunyuFan/MIFNet.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Elevation Information-Guided Multimodal Fusion Robust Framework for Remote Sensing Image Segmentation
    Fan, Junyu
    Li, Jinjiang
    Hua, Zhen
    Zhang, Fan
    Zhang, Caiming
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [2] Frequency-Aware Integrity Learning Network for Semantic Segmentation of Remote Sensing Images
    Yang, Penghan
    Zhou, Wujie
    Liu, Yuanyuan
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2025, 18 : 3398 - 3409
  • [3] Frequency-Aware Feature Fusion for Dense Image Prediction
    Chen, Linwei
    Fu, Ying
    Gu, Lin
    Yan, Chenggang
    Harada, Tatsuya
    Huang, Gao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 10763 - 10780
  • [4] Hybrid Attention Fusion Embedded in Transformer for Remote Sensing Image Semantic Segmentation
    Chen, Yan
    Dong, Quan
    Wang, Xiaofeng
    Zhang, Qianchuan
    Kang, Menglei
    Jiang, Wenxiang
    Wang, Mengyuan
    Xu, Lixiang
    Zhang, Chen
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 4421 - 4435
  • [5] Boundary-Aware Multiscale Learning Perception for Remote Sensing Image Segmentation
    You, Chao
    Jiao, Licheng
    Liu, Xu
    Li, Lingling
    Liu, Fang
    Ma, Wenping
    Yang, Shuyuan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [6] STAIR FUSION NETWORK FOR REMOTE SENSING IMAGE SEMANTIC SEGMENTATION
    Hua, Wenyi
    Liu, Jia
    Liu, Fang
    Zhang, Wenhua
    An, Jiaqi
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 5499 - 5502
  • [7] Prior-Guided Fuzzy-Aware Multibranch Network for Remote Sensing Image Segmentation
    Wei, Guangyi
    Xu, Jindong
    Chong, Qianpeng
    Huang, Jianjun
    Xing, Haihua
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21
  • [8] FCLFusion: A frequency-aware and collaborative learning for infrared and visible image fusion
    Wang, Chengchao
    Pu, Yuanyuan
    Zhao, Zhengpeng
    Nie, Rencan
    Cao, Jinde
    Xu, Dan
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 137
  • [9] CFFormer: A Cross-Fusion Transformer Framework for the Semantic Segmentation of Multisource Remote Sensing Images
    Zhao, Jinqi
    Zhang, Ming
    Zhou, Zhonghuai
    Wang, Zixuan
    Lang, Fengkai
    Shi, Hongtao
    Zheng, Nanshan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2025, 63
  • [10] CNN and Transformer Fusion for Remote Sensing Image Semantic Segmentation
    Chen, Xin
    Li, Dongfen
    Liu, Mingzhe
    Jia, Jiaru
    REMOTE SENSING, 2023, 15 (18)