Segmentation of Shadowed Buildings in Dense Urban Areas from Aerial Photographs
Abstract
:1. Introduction
2. Study Area
3. Segmentation Algorithm
- Set the number of DN intervals for quantization Ndisc, the associated interval widths Δdi (i = 1, ⋯, Ndisc), and the number of offsets Noff. The offset width Δoffi is defined as Δoffi = Δdi/Noff, and Noff quantized images are generated at a given value of Δdi by applying the different offsets. For example, with Δdi = 40 and Noff = 5, Δoffi is 40/5 = 8, and the offsets are {0, 8, 16, 24, 32}. With offset = 0, DNs are quantized into the intervals [0, 39], [40, 79], [80, 119], [120, 159], [160, 199], [200, 239], and [240, 255], and all pixels having a DN within an interval are assigned the same quantum value.
- Taking each quantized image in turn, regions are extracted and labeled by examining both the four neighboring pixels surrounding a given pixel and all other connecting pixels having the same quantum value. Large regions are removed, and then small regions are merged with neighboring larger regions, if such larger regions exist; otherwise, the small regions are removed. Finally, the edges of any remaining regions are extracted.
- All edges of the Noff quantized images at a given value of Δdi are merged, and the number of edge detections within each pixel is counted.
- A pixel whose edge count is greater than or equal to a threshold Tcount1 is preserved as an edge. Moreover, a pixel whose edge count is smaller than Tcount1, but greater than or equal to Tcount2, is added to an edge group if the pixel is connected to preserved edge pixels. Finally, a non-edge pixel is changed into an edge if linear alignments of edges pixels are found either side of it.
- Segmented regions are generated using the edges found in each quantization. To perform segmentation, a “rectangular index” is calculated as follows (see Figure 3).
- By using the 2D coordinates of the edges in a region, a main axis and sub-axis are determined, where the sub-axis is orthogonal to the main axis.
- The region is then projected onto the main axis, and the maximum, V1,max, and minimum, V1,min, coordinate values along the main axis are obtained. In the same manner, the maximum, V2,max, and minimum, V2,min, coordinate values along the sub-axis are obtained. A rectangular area is calculated by using the formula Srect = (V1, max − V1, min + 1) * (V2, max − V2, min + 1).
- The rectangular index idx is defined as the ratio between the actual area of the region Sactual and Srect,
- If idx is lower than a given threshold, the region is removed because a strong likelihood exists that the region does not correspond to a building.
- Regions obtained in the Ndisc images are sorted according to their rectangular index.
- Regions with high rectangular index are selected as buildings, as long as no part of the region overlaps with regions already selected. The unselected regions are next considered, and a region is examined if both its overlap area with previously selected regions and the ratio between this area and the region’s total area are less than or equal to given thresholds. If idx for the portion of the region without overlap is greater than or equal to a further threshold, that portion is added to the group of regions nominated as buildings. Finally, any holes in the buildings are filled.
- The local scores in Figure 4(a,b) are greater than or equal to Tcount3.
- The total score of all (7 × 7 pixels) components is greater than or equal to Tcount4.
4. Results
5. Discussions
5.1. Effect of Quantization and Edge Completion
5.2. Rectangular Index
5.3. Optimization of Parameters
5.4. Computation Time
5.5. Applications
6. Conclusions
Acknowledgments
References
- Elberink, S.O. Target Graph Matching for Building Reconstruction. Proceedings of Laserscanning’09, Paris, France, 1–2 September 2009. In International Archives of the Photogramm., Remote Sensing and Spatial Information Sciences; 2009; Volume XXXVIII, 3/W8, pp. 49–54.
- Steuer, H. Height Snakes: 3D Building Reconstruction from Aerial Image and Laser Scanner Data. Proceedings of 2011 Urban Remote Sensing Joint Event, Munich, Germany, 11–13 April 2011; pp. 113–116.
- Huber, M.; Schickler, W.; Hinz, S.; Baumgartner, A. Fusion of LIDAR Data and Aerial Imagery for Automatic Reconstruction of Building Surfaces. Proceedings of 2nd GRSS/ISPRS Joint Workshop on “Data Fusion and Remote Sensing over Urban Areas”, Berlin, Germany, 22–23 May 2003; pp. 82–86.
- Sohn, G.; Dowman, I. Data fusion of high-resolution satellite imagery and LiDAR data for automatic building extraction. ISPRS J. Photogramm. 2007, 62, 43–63. [Google Scholar]
- Rottensteiner, F.; Trinder, J.; Clode, S.; Kubik, K. Building detection by fusion of airborne laser scanner data and multi-spectral images: Performance evaluation and sensitivity analysis. ISPRS J. Photogramm. 2007, 62, 135–149. [Google Scholar]
- Awrangjeb, M.; Ravanbakhsh, M.; Fraser, C.S. Automatic detection of residential buildings using LIDAR data and multispectral imagery. ISPRS J. Photogramm. 2010, 65, 457–467. [Google Scholar]
- Tso, B.; Mather, P.M. Classification Methods for Remotely Sensed Data, 2nd ed.; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar]
- Soille, P.; Pesaresi, M. Advances in mathematical morphology applied to geoscience and remote sensing. IEEE Trans. Geosci. Remote Sens 2002, 40, 2042–2055. [Google Scholar]
- Benediktsson, J.A.; Pesaresi, M.; Amason, K. Classification and feature extraction for remote sensing images from urban areas based on morphological transformations. IEEE Trans. Geosci. Remote Sens 2003, 41, 1940–1949. [Google Scholar]
- Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens 2005, 43, 480–491. [Google Scholar]
- Bellens, R.; Gautama, S.; Martinez-Fonte, L.; Philips, W.; Chan, J.C.-W.; Canters, F. Improved classification of VHR images of urban areas using directional morphological profiles. IEEE Trans. Geosci. Remote Sens 2008, 46, 2803–2813. [Google Scholar]
- Tuia, D.; Pacifici, F.; Kanevski, M.; Emery, W.J. Classification of Very High Spatial Resolution imagery using mathematical morphology and Support Vector Machines. IEEE Trans. Geosci. Remote Sens 2009, 47, 3866–3879. [Google Scholar]
- Novack, T.; Esch, T.; Kux, H.; Stilla, U. Machine learning comparison between WorldView-2 and QuickBird-2-simulated imagery regarding object-based urban land cover classification. Remote Sens 2011, 3, 2263–2282. [Google Scholar]
- Moskal, L.M.; Styers, D.M.; Halabisky, M. Monitoring urban tree cover using object-based image analysis and public domain remotely sensed data. Remote Sens 2011, 3, 2243–2262. [Google Scholar]
- Tsai, Y.H.; Stow, D.; Weeks, J. Comparison of object-based image analysis approaches to mapping new buildings in Accra, Ghana using multi-temporal QuickBird satellite imagery. Remote Sens 2011, 3, 2707–2726. [Google Scholar]
- Canny, J. A. Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell 1986, PAMI-8, 679–698. [Google Scholar]
- Wan, T.; Canagarajah, N.; Achim, A. Segmentation of noisy colour images using cauchy distribution in the complex wavelet domain. IET Image Process 2011, 5, 159–170. [Google Scholar]
- Zhong, J.; Ning, R. Image denoising based on wavelets and multifractals for singularity detection. IEEE Trans. Image Process 2005, 14, 1435–1447. [Google Scholar]
- Rezaee, M.R.; van der Zwet, P.M.J.; Lelieveldt, B.P.E.; van der Geest, R.J.; Reiber, J.H.C. A multiresolution image segmentation technique based on pyramidal segmentation and fuzzy clustering. IEEE Trans. Image Proces 2000, 9, 1238–1248. [Google Scholar]
- Cheng, H.; Bouman, C.A. Mutliscale Bayesian segmentation using a trainable context model. IEEE Trans. Image Processing 2001, 10, 511–525. [Google Scholar]
- Esch, T.; Thiel, M.; Bock, M.; Roth, A.; Dech, S. Improvement of image segmentation accuracy based on multiscale optimization procedure. IEEE Geosci. Remote Sens. Lett 2008, 5, 463–467. [Google Scholar]
- Berlemont, S.; Olivo-Marin, J.-C. Combining local filtering and multiscale analysis for edge, ridge, and curvilinear objects detection. IEEE Trans. Image Process 2010, 19, 74–84. [Google Scholar]
- Bouman, C.A.; Shapiro, M. A mutliscale random field model for Bayesian image segmentation. IEEE Trans. Image Process 1994, 3, 162–177. [Google Scholar]
- Paget, R.; Longstaff, I.D. Texture synthesis via a noncausal nonparametric multiscale Markov random field. IEEE Trans. Image Process 1998, 7, 925–931. [Google Scholar]
- Liang, K.-H.; Tjahjadi, T. Adaptive scale fixing for multiscale texture segmentation. IEEE Trans. Image Process 2006, 15, 249–256. [Google Scholar]
- Tonazzini, A.; Bedini, L.; Salerno, E. A Markov model for blind image separation by a mean-field EM algorithm. IEEE Trans. Image Process 2006, 15, 473–482. [Google Scholar]
- Ding, J.; Ma, R.; Chen, S. A scale-based connected coherence tree algorithm for image segmentation. IEEE Trans. Image Process 2008, 17, 204–216. [Google Scholar]
- Chien, S.Y.; Ma, S.Y.; Chen, L.G. Efficient moving object segmentation algorithm using background registration technique. IEEE T. Circ. Syst. Vid 2002, 12, 577–586. [Google Scholar]
- Tsai, V.J.D. A comparative study on shadow compensation of color aerial images in invariant color models. IEEE Trans. Geosci. Remote Sens 2006, 44, 1661–1671. [Google Scholar]
- Ma, H.; Qin, Q.; Shen, X. Shadow Segmentation and Compensation in High Resolution Satellite Images. Proceedings of 2008 IEEE International Geosci. Remote Sensing Symposium, Boston, MA, USA, 7–11 July 2008; p. II-1036-II-1039.
- Deng, Y.; Manjunath, B.S. Unsupervised segmentation of color-texture regions in images and video. IEEE Trans. Pattern Anal. Mach. Intell 2001, 23, 800–810. [Google Scholar]
- Advanced Image Analysis for GIS from ITT. ENVI Feature Extraction Module (ENVI FX)ENVI Feature Extraction Module (ENVI FX). Available online: http://www.ittvis.com/language/en-us/productsservices/envi/enviex.aspx (accessed on 5 June 2011).
- Jin, X. Segmentation-based image processing system. US Patent 20,090,123,070; filed 14 November 2007, and issued 14 May 2009,
- Susaki, J. Fusion of airborne LiDAR data and aerial photographs for automatic modeling of buildings in dense urban areas. IEEE Trans. Geosci. Remote Sens 2012. in preparation. [Google Scholar]
Process | Parameters | Value Used |
---|---|---|
Quantization and edge detection | Number of quantizations Ndisc | 3 |
Quantization interval widths Δdi (i = 1, ..., Ndisc) | 40, 30, 20 | |
Number of offsets Noff | 5 | |
Edge count T count1 and T count2 | 5 and 3 | |
Minimum and maximum areas | 50 and 30,000 pixels | |
Minimum score for edge completion using filters shown in Figure 4, Tcount3 and T count4 | 2 and 8 | |
Segmentation and calculation of rectangular index | Minimum rectangular index | 0.45 |
Minimum and maximum distance between edges for rectangular index calculation, dedge_min, dedge_max | 5 and 20 pixels | |
Minimum valid length of rectangle | 8 pixels | |
Selection regions according to rectangular index | Maximum ratio of overlapping area to original area for selecting areas overlapping with previously selected areas | 0.2 |
Minimum area for selecting areas overlapping with previously selected areas (same for Step (1)) | 50 pixels |
Share and Cite
Susaki, J. Segmentation of Shadowed Buildings in Dense Urban Areas from Aerial Photographs. Remote Sens. 2012, 4, 911-933. https://doi.org/10.3390/rs4040911
Susaki J. Segmentation of Shadowed Buildings in Dense Urban Areas from Aerial Photographs. Remote Sensing. 2012; 4(4):911-933. https://doi.org/10.3390/rs4040911
Chicago/Turabian StyleSusaki, Junichi. 2012. "Segmentation of Shadowed Buildings in Dense Urban Areas from Aerial Photographs" Remote Sensing 4, no. 4: 911-933. https://doi.org/10.3390/rs4040911
APA StyleSusaki, J. (2012). Segmentation of Shadowed Buildings in Dense Urban Areas from Aerial Photographs. Remote Sensing, 4(4), 911-933. https://doi.org/10.3390/rs4040911