scholarly journals BRRNet: A Fully Convolutional Neural Network for Automatic Building Extraction From High-Resolution Remote Sensing Images

2020 ◽  
Vol 12 (6) ◽  
pp. 1050 ◽  
Author(s):  
Zhenfeng Shao ◽  
Penghao Tang ◽  
Zhongyuan Wang ◽  
Nayyer Saleem ◽  
Sarath Yam ◽  
...  

Building extraction from high-resolution remote sensing images is of great significance in urban planning, population statistics, and economic forecast. However, automatic building extraction from high-resolution remote sensing images remains challenging. On the one hand, the extraction results of buildings are partially missing and incomplete due to the variation of hue and texture within a building, especially when the building size is large. On the other hand, the building footprint extraction of buildings with complex shapes is often inaccurate. To this end, we propose a new deep learning network, termed Building Residual Refine Network (BRRNet), for accurate and complete building extraction. BRRNet consists of such two parts as the prediction module and the residual refinement module. The prediction module based on an encoder–decoder structure introduces atrous convolution of different dilation rates to extract more global features, by gradually increasing the receptive field during feature extraction. When the prediction module outputs the preliminary building extraction results of the input image, the residual refinement module takes the output of the prediction module as an input. It further refines the residual between the result of the prediction module and the real result, thus improving the accuracy of building extraction. In addition, we use Dice loss as the loss function during training, which effectively alleviates the problem of data imbalance and further improves the accuracy of building extraction. The experimental results on Massachusetts Building Dataset show that our method outperforms other five state-of-the-art methods in terms of the integrity of buildings and the accuracy of complex building footprints.

2020 ◽  
Vol 86 (4) ◽  
pp. 235-245 ◽  
Author(s):  
Ka Zhang ◽  
Hui Chen ◽  
Wen Xiao ◽  
Yehua Sheng ◽  
Dong Su ◽  
...  

This article proposes a new building extraction method from high-resolution remote sensing images, based on GrabCut, which can automatically select foreground and background samples under the constraints of building elevation contour lines. First the image is rotated according to the direction of pixel displacement calculated by the rational function Model. Second, the Canny operator, combined with morphology and the Hough transform, is used to extract the building's elevation contour lines. Third, seed points and interesting points of the building are selected under the constraint of the contour line and the geodesic distance. Then foreground and background samples are obtained according to these points. Fourth, GrabCut and geometric features are used to carry out image segmentation and extract buildings. Finally, WorldView satellite images are used to verify the proposed method. Experimental results show that the average accuracy can reach 86.34%, which is 15.12% higher than other building extraction methods.


Sensors ◽  
2020 ◽  
Vol 20 (24) ◽  
pp. 7241
Author(s):  
Dengji Zhou ◽  
Guizhou Wang ◽  
Guojin He ◽  
Tengfei Long ◽  
Ranyu Yin ◽  
...  

Building extraction from high spatial resolution remote sensing images is a hot spot in the field of remote sensing applications and computer vision. This paper presents a semantic segmentation model, which is a supervised method, named Pyramid Self-Attention Network (PISANet). Its structure is simple, because it contains only two parts: one is the backbone of the network, which is used to learn the local features (short distance context information around the pixel) of buildings from the image; the other part is the pyramid self-attention module, which is used to obtain the global features (long distance context information with other pixels in the image) and the comprehensive features (includes color, texture, geometric and high-level semantic feature) of the building. The network is an end-to-end approach. In the training stage, the input is the remote sensing image and corresponding label, and the output is probability map (the probability that each pixel is or is not building). In the prediction stage, the input is the remote sensing image, and the output is the extraction result of the building. The complexity of the network structure was reduced so that it is easy to implement. The proposed PISANet was tested on two datasets. The result shows that the overall accuracy reached 94.50 and 96.15%, the intersection-over-union reached 77.45 and 87.97%, and F1 index reached 87.27 and 93.55%, respectively. In experiments on different datasets, PISANet obtained high overall accuracy, low error rate and improved integrity of individual buildings.


Author(s):  
Z. Ding ◽  
X. Q. Wang ◽  
Y. L. Li ◽  
S. S. Zhang

Building extraction from high resolution remote sensing images is a hot research topic in the field of photogrammetry and remote sensing. However, the diversity and complexity of buildings make building extraction methods still face challenges in terms of accuracy, efficiency, and so on. In this study, a new building extraction framework based on MBI and combined with image segmentation techniques, spectral constraint, shadow constraint, and shape constraint is proposed. In order to verify the proposed method, worldview-2, GF-2, GF-1 remote sensing images covered Xiamen Software Park were used for building extraction experiments. Experimental results indicate that the proposed method improve the original MBI significantly, and the correct rate is over 86 %. Furthermore, the proposed framework reduces the false alarms by 42 % on average compared to the performance of the original MBI.


2019 ◽  
Vol 11 (23) ◽  
pp. 2813 ◽  
Author(s):  
Wenchao Kang ◽  
Yuming Xiang ◽  
Feng Wang ◽  
Hongjian You

Automatic building extraction from high-resolution remote sensing images has many practical applications, such as urban planning and supervision. However, fine details and various scales of building structures in high-resolution images bring new challenges to building extraction. An increasing number of neural network-based models have been proposed to handle these issues, while they are not efficient enough, and still suffer from the error ground truth labels. To this end, we propose an efficient end-to-end model, EU-Net, in this paper. We first design the dense spatial pyramid pooling (DSPP) to extract dense and multi-scale features simultaneously, which facilitate the extraction of buildings at all scales. Then, the focal loss is used in reverse to suppress the impact of the error labels in ground truth, making the training stage more stable. To assess the universality of the proposed model, we tested it on three public aerial remote sensing datasets: WHU aerial imagery dataset, Massachusetts buildings dataset, and Inria aerial image labeling dataset. Experimental results show that the proposed EU-Net is superior to the state-of-the-art models of all three datasets and increases the prediction efficiency by two to four times.


Sign in / Sign up

Export Citation Format

Share Document