[1]TAN Ruijun,ZHAO Zhicheng,XIE Xinlin.Double-residual semantic segmentation network and traffic scenic application[J].CAAI Transactions on Intelligent Systems,2022,17(4):780-787.[doi:10.11992/tis.202106020]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
17
Number of periods:
2022 4
Page number:
780-787
Column:
学术论文—人工智能基础
Public date:
2022-07-05
- Title:
-
Double-residual semantic segmentation network and traffic scenic application
- Author(s):
-
TAN Ruijun; ZHAO Zhicheng; XIE Xinlin
-
School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan 030024, China
-
- Keywords:
-
double-residual network; detail feature extraction; jump feature fusion; up sampling; network branch training; semantic image segmentation; CamVid data set; Cityscapes data set
- CLC:
-
TP183
- DOI:
-
10.11992/tis.202106020
- Abstract:
-
An image semantic segmentation network based on the double-residual network, named DRsenet, is proposed to address the depth problem of feature extraction network in semantic image segmentation and resolution reduction of feature maps by a down-sampling pooling layer. Firstly, a double-residual network is proposed to extract the detailed features of each target in the training set and improve the perception ability of the network for some detailed targets. Secondly, the jump feature fusion starts from Layer 1, and the up-sampling operation is continued by the 2×deconvolution method to fuse the low-level and the high-level features to reduce the impact of partial detail information loss on the segmentation accuracy. Finally, the network branch training method is adopted to train the outline features of each target on the image, followed by training the detailed features of each target. The results indicate that the network’s PMIOU improves from 49.72% to 59.44% on CamVid and from 44.35% to 47.77% on Cityscapes when compared to the full convolution network. The network can produce image segmentation results with higher accuracy and more complete edge segmentation.