共 44 条
Channel2DTransformer: A Multi-level Features Self-attention Fusion Module for Semantic Segmentation
被引:0
作者:
Liu, Weitao
[1
]
Wu, Junjun
[1
]
机构:
[1] Foshan Univ, Guangdong Prov Key Lab Ind Intelligent Inspection, Foshan, Peoples R China
基金:
中国国家自然科学基金;
国家重点研发计划;
关键词:
Semantic segmentation;
Channel2DTransformer;
Self-attention;
Deep learning;
D O I:
10.1007/s44196-024-00630-5
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Semantic segmentation is a crucial technology for intelligent vehicles, enabling scene understanding in complex driving environments. However, complex real-world scenarios often contain diverse multi-scale objects, which bring challenges to the accurate semantic segmentation. To address this challenge, we propose a multi-level features self-attention fusion module called Channel2DTransformer. The module utilizes self-attention mechanisms to dynamically fuse multi-level features by computing self-attention weights between their channels, resulting in a consistent and comprehensive representation of scene features. We perform the module on the Cityscapes and NYUDepthV2 datasets, which contain a large number of multi-scale objects. The experimental results validate the positive contributions of the module in enhancing the semantic segmentation accuracy of multi-scale objects and improving the performance of semantic segmentation in complex scenes.
引用
收藏
页数:11
相关论文
共 44 条