Abstract

Street scene images contain various objects of different scales. The segmentation model with single scale and feature extraction and fusion can not get good segmentation and prediction results. Therefore, a semantic segmentation model based on multi-scale feature fusion and attention mechanism is proposed. Firstly, the asymmetric structure of atrous spatial pyramid pooling (ASPP) is used to optimize the extraction of different levels and scales of street scene image. Secondly, the attention mechanism is introduced into the feature maps of different scales, so that the network can focus on the salient features of each level. Finally, all the feature images are adjusted to the same size for fusion, and the key feature information of each scale object in the street scene is fully extracted to segment it effectively. The experimental results on the dataset Cityscapes show that the semantic segmentation network model based on multi-scale and attention mechanism can further improve the segmentation accuracy and optimize the segmentation results.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call