Title: Multi-level spatial attention network for image data segmentation
Authors: Jun Guo; Zhixiong Jiang; Dingchao Jiang
Addresses: School of Software, Quanzhou University of Information Engineering, Quanzhou, Fujian, China ' School of Business, Shanghai Dianji University, Shanghai, China ' School of Electronic and Information Engineering, Xi'an Jiaotong University, Xi'an, Shaanxi, China
Abstract: Deep learning models for semantic image segmentation are limited in their hierarchical architectures to extract features, which results in losing contextual and spatial information. In this paper, a new attention-based network, MSANet, which applies an encoder-decoder structure, is proposed for image data segmentation to aggregate contextual features from different levels and reconstruct spatial characteristics efficiently. To model long-range spatial dependencies among features, the multi-level spatial attention module (MSAM) is presented to process multi-level features in the encoder network and capture global contextual information. In this way, our model learns multi-level spatial dependencies between features by the MSAM and hierarchical representations of the input image by the stacked convolutional layers, which means the model is more capable of producing accurate segmentation results. The proposed network is evaluated on the PASCAL VOC 2012 and Cityscapes datasets. Results show that our model achieves excellent performance compared with U-net, FCNs, and DeepLabv3.
Keywords: deep learning; semantic segmentation; big data.
International Journal of Embedded Systems, 2021 Vol.14 No.3, pp.289 - 299
Received: 05 Dec 2020
Accepted: 11 Jan 2021
Published online: 12 Jul 2021 *