Chih-Yang LinYi-Cheng ChiuHui-Fuang NgTimothy K. ShihKuan-Hung Lin2025-01-092025-01-092020-05-2110.3390/s20102907https://dspace-cris.utar.edu.my/handle/123456789/11038Semantic segmentation of street view images is an important step in scene understanding for autonomous vehicle systems. Recent works have made significant progress in pixel-level labeling using Fully Convolutional Network (FCN) framework and local multi-scale context information. Rich global context information is also essential in the segmentation process. However, a systematic way to utilize both global and local contextual information in a single network has not been fully investigated. In this paper, we propose a global-and-local network architecture (GLNet) which incorporates global spatial information and dense local multi-scale context information to model the relationship between objects in a scene, thus reducing segmentation errors. A channel attention module is designed to further refine the segmentation results using low-level features from the feature map. Experimental results demonstrate that our proposed GLNet achieves 80.8% test accuracy on the Cityscapes test dataset, comparing favorably with existing state-of-the-art methods.Global-and-Local Context Network for Semantic Segmentation of Street View Imagesjournal-article