TY - JOUR
T1 - See more than once
T2 - Kernel-sharing atrous convolution for semantic segmentation
AU - Huang, Ye
AU - Wang, Qingqing
AU - Jia, Wenjing
AU - Lu, Yue
AU - Li, Yuxin
AU - He, Xiangjian
N1 - Publisher Copyright:
© 2021 Elsevier B.V.
PY - 2021/7/5
Y1 - 2021/7/5
N2 - The state-of-the-art semantic segmentation solutions usually leverage different receptive fields via multiple parallel branches to handle objects of different sizes. However, employing separate kernels for individual branches may degrade the generalization of the network to objects with different scales, and the computational cost increases with the increase of the number of branches. To tackle this problem, we propose a novel network structure, namely Kernel-Sharing Atrous Convolution (KSAC), where branches with different receptive fields share the same kernel, i.e., let a single kernel ‘see’ the input feature maps more than once with different receptive fields. Experiments conducted on the benchmark PASCAL VOC 2012 dataset show that our proposed sharing strategy can not only boost the network's generalization and representation abilities but also reduce the computational cost significantly. Specifically, on the validation set, when compared with DeepLabv3+, about 2.7G FLOPs and 12.7G FLOPs are saved for output stride = 16 and 8 respectively. In addition, different from the widely used ASPP structure, our proposed KSAC is able to further improve the mIOU by taking benefit of wider context with larger atrous rates. Finally, our KSAC achieves mIOUs of 88.1%, 45.47% and 80.7% on the PASCAL VOC 2012 test set (Everingham et al., 2009), ADE20K dataset (Zhou et al., 2017) and Cityscapes datasets (Marius et al., 2016), respectively. Our full code will be released on Github: https://github.com/edwardyehuang/iSeg.
AB - The state-of-the-art semantic segmentation solutions usually leverage different receptive fields via multiple parallel branches to handle objects of different sizes. However, employing separate kernels for individual branches may degrade the generalization of the network to objects with different scales, and the computational cost increases with the increase of the number of branches. To tackle this problem, we propose a novel network structure, namely Kernel-Sharing Atrous Convolution (KSAC), where branches with different receptive fields share the same kernel, i.e., let a single kernel ‘see’ the input feature maps more than once with different receptive fields. Experiments conducted on the benchmark PASCAL VOC 2012 dataset show that our proposed sharing strategy can not only boost the network's generalization and representation abilities but also reduce the computational cost significantly. Specifically, on the validation set, when compared with DeepLabv3+, about 2.7G FLOPs and 12.7G FLOPs are saved for output stride = 16 and 8 respectively. In addition, different from the widely used ASPP structure, our proposed KSAC is able to further improve the mIOU by taking benefit of wider context with larger atrous rates. Finally, our KSAC achieves mIOUs of 88.1%, 45.47% and 80.7% on the PASCAL VOC 2012 test set (Everingham et al., 2009), ADE20K dataset (Zhou et al., 2017) and Cityscapes datasets (Marius et al., 2016), respectively. Our full code will be released on Github: https://github.com/edwardyehuang/iSeg.
KW - Atrous convolution
KW - Feature augmentation
KW - Kernel sharing
KW - Semantic segmentation
UR - https://www.scopus.com/pages/publications/85103128695
U2 - 10.1016/j.neucom.2021.02.091
DO - 10.1016/j.neucom.2021.02.091
M3 - 文章
AN - SCOPUS:85103128695
SN - 0925-2312
VL - 443
SP - 26
EP - 34
JO - Neurocomputing
JF - Neurocomputing
ER -