论文标题
混合多重注意网络,用于空中图像中语义分割
Hybrid Multiple Attention Network for Semantic Segmentation in Aerial Images
论文作者
论文摘要
非常高分辨率(VHR)的语义分割是空中图像是遥感图像理解中最具挑战性的任务之一。当前的大多数方法都是基于深度卷积神经网络(DCNNS)。但是,与本地接收场的标准卷积在建模全球依赖性时失败。先前的研究表明,基于注意力的方法可以捕获远程依赖性,并进一步重建特征图以更好地表示。然而,仅受到空间和渠道注意的观点的限制,以及自我发挥机制的巨大计算复杂性,不可能模拟复杂光谱的遥感数据的每个像素对之间的有效语义相互依赖性。在这项工作中,我们提出了一个名为Hybrid多重注意网络(HManet)的新型基于注意力的框架,以更加有效的方式从空间,渠道和类别的角度自适应地捕获全局相关性。具体而言,嵌入了类通道注意的类增强注意力(CAA)模块可用于计算基于类别的相关性并重新校准类级信息。此外,我们引入了一个简单而有效的区域洗牌(RSA)模块,以降低特征冗余,并通过区域表示来提高自我发挥机制的效率。关于ISPRS Vaihingen和Potsdam基准的广泛实验结果证明了我们的HManet对其他最先进方法的有效性和效率。
Semantic segmentation in very high resolution (VHR) aerial images is one of the most challenging tasks in remote sensing image understanding. Most of the current approaches are based on deep convolutional neural networks (DCNNs). However, standard convolution with local receptive fields fails in modeling global dependencies. Prior researches have indicated that attention-based methods can capture long-range dependencies and further reconstruct the feature maps for better representation. Nevertheless, limited by the mere perspective of spacial and channel attention and huge computation complexity of self-attention mechanism, it is unlikely to model the effective semantic interdependencies between each pixel-pair of remote sensing data of complex spectra. In this work, we propose a novel attention-based framework named Hybrid Multiple Attention Network (HMANet) to adaptively capture global correlations from the perspective of space, channel and category in a more effective and efficient manner. Concretely, a class augmented attention (CAA) module embedded with a class channel attention (CCA) module can be used to compute category-based correlation and recalibrate the class-level information. Additionally, we introduce a simple yet effective region shuffle attention (RSA) module to reduce feature redundant and improve the efficiency of self-attention mechanism via region-wise representations. Extensive experimental results on the ISPRS Vaihingen and Potsdam benchmark demonstrate the effectiveness and efficiency of our HMANet over other state-of-the-art methods.