location: Current position: Home >> Scientific Research >> Paper Publications

Deep gated attention networks for large-scale street-level scene segmentation

Hits:

Indexed by:期刊论文

Date of Publication:2019-04-01

Journal:PATTERN RECOGNITION

Included Journals:SCIE、Scopus

Volume:88

Page Number:702-714

ISSN No.:0031-3203

Key Words:Scene segmentation; Fully convolutional network; Spatial gated attention; Street-level image understanding

Abstract:Street-level scene segmentation aims to label each pixel of street-view images into specific semantic categories. It has been attracting growing interest due to various real-world applications, especially in the area of autonomous driving. However, this pixel-wise labeling task is very challenging under the complex street-level scenes and large-scale object categories. Motivated by the scene layout of street-view images, in this work we propose a novel Spatial Gated Attention (SGA) module, which automatically highlights the attentive regions for pixel-wise labeling, resulting in effective street-level scene segmentation. The proposed module takes as input the multi-scale feature maps based on a Fully Convolutional Network (FCN) backbone, and produces the corresponding attention mask for each feature map. The learned attention masks can neatly highlight the regions of interest while suppress background clutter. Furthermore, we propose an efficient multi-scale feature interaction mechanism which is able to adaptively aggregate the hierarchical features. Based on the proposed mechanism, the features of different levels are adaptively re-weighted according to the local spatial structure and the surrounding contextual information. Consequently, the proposed modules are able to boost standard FCN architectures and result in an enhanced pixel-wise segmentation for street-level scene images. Extensive experiments on three public available street-level benchmarks demonstrate that the proposed Gated Attention Network (GANet) approach achieves consistently superior performance and outperforms the very recent state-of-the-art methods. (C) 2018 Elsevier Ltd. All rights reserved.

Pre One:Device-Free Tracking via Joint Velocity and AOA Estimation With Commodity WiFi

Next One:Hyperspectral Image Classification Based on Deep Deconvolution Network With Skip Architecture