Next Article in Journal
Gross Floor Area Estimation from Monocular Optical Image Using the NoS R-CNN
Next Article in Special Issue
Quality Control for the BPG Lossy Compression of Three-Channel Remote Sensing Images
Previous Article in Journal
Filling Temporal Gaps within and between GRACE and GRACE-FO Terrestrial Water Storage Records: An Innovative Approach
Previous Article in Special Issue
Multi-Epoch and Multi-Imagery (MEMI) Photogrammetric Workflow for Enhanced Change Detection Using Time-Lapse Cameras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Building Height Extraction from GF-7 Satellite Images Based on Roof Contour Constrained Stereo Matching

1
State Key Laboratory of Information Engineering in Surveying Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
2
School of Remote Sensing & Information Engineering, Wuhan University, Wuhan 430079, China
3
School of Computer Science, China University of Geosciences, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(7), 1566; https://doi.org/10.3390/rs14071566
Submission received: 13 February 2022 / Revised: 20 March 2022 / Accepted: 22 March 2022 / Published: 24 March 2022

Abstract

:
Building height is one of the basic geographic information for planning and analysis in urban construction. It is still very challenging to estimate the accurate height of complex buildings from satellite images, especially for buildings with podium. This paper proposes a solution for building height estimation from GF-7 satellite images by using a roof contour constrained stereo matching algorithm and DSM (Digital Surface Model) based bottom elevation estimation. First, an object-oriented roof matching algorithm is proposed based on building contour to extract accurate building roof elevation from GF-7 stereo image, and DSM generated from the GF-7 stereo images is then used to obtain building bottom elevation. Second, roof contour constrained stereo matching is conducted between backward and forward image blocks, in which the difference of standard deviation maps is used for the similarity measure. To deal with the multi-height problem of podium buildings, the gray difference image is adopted to segment podium buildings, and re-matching is conducted to find out their actual heights. Third, the building height is obtained through the elevation difference between the building top and bottom, in which the evaluation of the building bottom is calculated according to the elevation histogram statistics of the building buffer in DSM. Finally, two GF-7 stereo satellite images, collected in Yingde, Guangzhou, and Xi’an, Shanxi, are used for performance evaluation. Besides, the aerial LiDAR point cloud is used for absolute accuracy evaluation. The results demonstrate that compared with other methods, our solution obviously improves the accuracy of height estimation of high-rise buildings. The MAE (Mean Absolute Error) of the estimated building heights in Yingde is 2.31 m, and the MAE of the estimated elevation of building top and bottom is approximately 1.57 m and 1.91 m, respectively. Then the RMSE (Root Mean Square Error) of building top and bottom is 2.01 m and 2.57 m. As for the Xi’an dataset with 7 buildings with podium out of 40 buildings, the MAE of the estimated building height is 1.69 m and the RMSE is 2.34 m. The proposed method can be an effective solution for building height extraction from GF-7 satellite images.

1. Introduction

With the acceleration of China’s urbanization process, the concept of “urban fine management” has been proposed [1]. Urban fine management is designed to solve the problems of blind expansion [2,3], population assessment [4,5,6], urban climate [7,8], destruction of natural environment and cultural heritage [9,10], urban 3D reconstruction [11,12,13,14], etc. As the fundamental element of urban cities, buildings are one of the most important research targets, and their height information plays a critical role in urban exemplary management. Thus, effective solutions are necessarily required in practice.
In contrast to shadow-based methods [15,16,17,18,19,20,21,22], other research aims to use dense matching technology to obtain DSM (Digital Surface Model) of building areas [23,24,25,26], in which building height can be calculated from the roof and bottom elevation of buildings [27,28]. The methods based on DSM have become the most commonly used solution, and the building height accuracy depends on the accuracy of the roof and bottom elevation. In general, there are two kinds of DSM extraction methods. One is the traditional dense matching methods, such as SGM (semi-global matching) [29]. The other is the deep learning-based methods [30,31,32,33,34]. The SGM algorithm combines the advantages of the local algorithm and the global algorithm, and avoids the disadvantages of them [29,35]. It has been adopted by most commercial satellite image processing software. For example, Tao et al. [36] used PCI and INPHO to explore DSM reconstruction of GF-7 images, and the accuracy of flat area can reach 0.655 m. However, the experimental data that was selected from Xinjiang with only a few low buildings, which cannot conclude for urban areas with dense high-rise buildings. Wang et al. [28] adopted SGM to generate point clouds from GF-7 stereo images for building height extraction, in which building footprints are extracted on the backward image by using a proposed multi-stage attention U-Net. The root mean square error (RMSE) between the extracted building height and the reference building height is 5.41 m, and the mean absolute error (MAE) is 3.39 m. The results demonstrate the promising potential of GF-7 RS images in stereo applications. However, they did not discuss the big forward angle of the GF-7 satellite camera, which leads to the loss of high-rise buildings when applying the pyramid SGM methods to estimate DSM.
On the other side, deep learning-based methods have been designed to extract elevation or height information from aerial or satellite images [37,38,39,40]. Karatsiolis S et al. [39] proposed a task-centered deep learning (DL) model, which combines the structural characteristics of the U-Net and residual network, and learns the mapping from single aerial image to standardized digital ground model. The model was trained on aerial images whose corresponding DSM and Digital Terrain Models (DTM) were available, and they were then used to infer the nDSM of images with no elevation information. To promote the deep learning based elevation and height information extraction from satellite images, IEEE released the worldview satellite images and corresponding DSM, DTM, and nDSM data [37]. Cao et al. [38] used ZY-3 data combined with A-map (a map service provider of China) building height data and proposed a multi-view, multi-spectral, and multi-objective neural network (called M3Net) to extract large-scale building footprints and heights on the backward image. He verified the applicability of the proposed method in various cities. The RMSE in the test site of Shenzhen is 6.43 m by M3Net. Although deep learning-based methods have shown exciting performance, we still can not use deep learning-based methods in the operational system, due to the requirement of a large number of sample data with ground truth for different satellites, different locations and different views. At present, there are few sample data databases for satellite image stereo matching. Due to the different design of different satellites in image spatial resolutions and stereo intersection angles, the sample database for one satellite cannot be used directly for other satellites [36].
Considering the above-mentioned issues, this paper proposes an object-oriented building roof matching method to improve the accuracy of building roof elevation of GF-7 images. Different from Wang et al. [28], who focused on the extraction of building footprints and directly used SGM for extracting building height, we focus on the extraction of roof elevation, in which DSM provides only bottom elevation information in the proposed method. In this paper, we do not discuss the extraction of building contour, but use the self-marked contours as input. In general, DSM and building contour are our input.
Our method consists of three steps: (1) The first step is image pre-processing, including the calculation of the forward image search range and feature map preparation. The image search range is determined by the minimum and maximum possible elevation of building roofs, which can be estimated with a DSM and the maximum possible height of the highest building. With the search range, the contour of the building on the backward image is projected to the forward image to obtain epipolar image blocks. (2) The second step is roof contour constrained stereo matching and podium segmentation, in which the backward image block is matched against the forward image with the similarity metre of feature map difference. Since the podium contains multiple elevation planes, the podium is segmented with image gray difference and rematched to obtain the elevation of multiple planes. (3) The third step is bottom elevation estimation and building height calculation. The bottom elevation is calculated according to the elevation histogram statistics of building buffer by using DSM. The top elevation is calculated using a space intersection algorithm. The building height is then obtained through the difference between the top and bottom elevations.
This paper is organized as follows. Our methodology is presented in Section 2. The experiment data and results are reported in Section 3. Finally, conclusions are drawn in Section 4.

2. Methodology

This paper proposes a roof contour constrained stereo matching method to extract the accurate height of high-rise buildings from GF-7 satellite images. The overall workflow of the proposed solution is illustrated in Figure 1. The input of our method consists of GF-7 backward and forward images, building roof contours on the backward image, and the corresponding DSM data. The method includes three major steps, as described as follows:
(1)
Epipolar image block and feature map generation from stereo images. The first step is to prepare the epipolar image block for roof contour matching, in which the parallax search range is determined by using the DSM data and a supposed maximum range of building heights. After image resampling, the epipolar image blocks are transformed into feature map blocks by computing regional standard deviation to extract the structural characteristics.
(2)
Roof contour constrained stereo matching considering podium structure. The backward image block is first matched against the forward image block with the similarity measure of image feature difference. To deal with the multi-heights problem arising from podium buildings, the gray difference image is used to detect the mismatched contours, which can include podium. The image is then rematched iteratively until no podium structure can be detected. With the matched parallax, the roof elevation is calculated through the space intersection using the RPCs of satellite stereo images and the 2D roof contours on the backward image is transformed into 3D roof contours in the geographic coordinate system.
(3)
Bottom elevation estimation and building height calculation. Bottom elevation is estimated using histogram analysis of the DSM in the buffer of a building contour with a buffer size of 20 m. Building height is then calculated by the subtraction between bottom elevation and roof elevation.

2.1. Epipolar Image Block and Feature Map Generation for Stereo Images

This step aims to prepare epipolar image blocks for roof contour constrained stereo matching. The source block range on the backward image is determined by the building contour itself. The destination block range on the forward image depends on the building roof elevation, which needs to be determined by image matching. So we use DSM for estimating the bottom elevation and plus the maximum possible height for roof elevation. Then the roof contour is transformed from the backward image to the forward image for epipolar resampling. At last, convolution processing is used to extract the standard deviation feature of the epipolar image blocks for decreasing the impact of bright change between the forward image and the backward image.

2.1.1. Calculation of Parallex Search Range

The principle of parallex search range is illustrated in Figure 2. The input building roof contour is on the 2D backward image, and the DSM is in the 3D object space. An iteration algorithm is used to obtain building roof elevation from the DSM data, in which it starts from an initial elevation Z 0 , such as the mean elevation of the study area. Then, the algorithm executes iteratively as follows:
(1)
With the initial evaluation Z 0 , we can get point P 0 in the 3D object space by intersecting imaging ray through the roof center with the horizontal plane created by the elevation of Z 0 .
(2)
After obtaining point P 0 , we use the horizontal coordinates X and Y of point P 0 to interpolate a new elevation value of Z 1 from the DSM data.
(3)
If the elevation difference between Z 1 and Z 0 is greater than a given threshold, then update Z 0 with Z 1 .
(4)
Steps 1 to 3 are iteratively executed until reaching the terminal condition.
By using the above-mentioned algorithm, the minimum elevation can be obtained, which is termed as Z m i n . Considering the maximum building height h m a x in the test area, the elevation search range can be calculated as Z m i n + h m a x . With the elevation search range ( Z m i n , Z m a x ), the parallex search range (begin, end) on forward image can be calculated by projecting the points of roof contours on the backward image to the forward image, whose envelope consists of the entire parallex search range on the forward image.

2.1.2. Epipolar Image Block and Feature Map Generation

By using the determined parallex search range, the GF-7 image is resampled by the epipolar transform to generate epipolar image blocks. After resampling the forward and backward images, their y parallax is eliminated. The two-dimensional image matching is simplified as one-dimensional matching in order to improve the matching efficiency [23,41].
The brightness change between the forward image and the backward image can greatly affect the image matching. To decrease the impact of brightness change, we conduct the image matching on the standard deviation maps instead of the original epipolar image block themselves. As shown in Figure 3, we use a convolution of 5 × 5 window to generate standard deviation maps. The larger the standard deviation value, the richer the image pixel information. The standard deviation is calculated with Equation (1)
S ( i , j ) = a = n n b = n n ( x ( i + a , j + b ) x ¯ ) 2 ( 2 n + 1 ) 2
where S ( i , j ) is the standard deviation value at point ( i , j ) ; n is the radius of the window; x ¯ is the mean value of the window; x ( i + a , j + b ) is the pixel in the window.
For a visual interpretation, Figure 4a shows a roof image of a building, and Figure 4b shows its standard deviation map. Bright areas include the internal structure of a building and dark areas are regarded as the background.

2.2. Roof Contour Constrained Stereo Matching Considering Podium Structure

2.2.1. Roof Contour Constrained Stereo Matching

Figure 5 illustrates the epipolar image blocks and their corresponding standard deviation maps. Figure 5a is the backward image block of a building; Figure 5b is the forward image block to be matched. Due to the imaging view difference, these two image blocks have dramatic bright difference. Figure 5c,d respectively shows the standard deviation map of Figure 5a,b, whose bright difference is greatly reduced.
Based on the standard deviation maps, we use the gray difference to measure the similarity between the backward image and the forward image. The gray difference similarity is measured as the absolute gray difference between the template and the searched image, as shown in Equation (2), where g S i + p j is the searched pixel value with x parallax p; w, h is the template window size; g T i + j is the pixel value of the template window, respectively. The parallax p with the smallest gray difference sum is regarded as the matched result.
S ( p ) = i = 1 w j = 1 h g S i + p j g T i + j
In order to exclude the interruption of pixels outside the building roof contours, the gray difference is modified to be a masked gray difference, as shown in Equation (3), in which only the roof pixels are taken into account.
S M p = i , j M g S i + p j g T i + j
where M is the mask matrix of template block; (i,j) is the pixel in the mask set.
As illustrated in Figure 6, a simple building with only one roof height plane can be matched in one pass, in which the mask is generated with the roof contour indicating the inner part of a building roof, and the histogram of the gray difference image has only one obvious peak.

2.2.2. Segmentation and Rematching of Building with Podium

In the previous section, we suppose that the building roof has only one height. That is correct for most buildings. In the real world, buildings, however, have different shapes and structures, and many buildings have several height planes. Such as, the podium building refers to an auxiliary building that is general at the bottom of the main body of a multi-story, high-rise, or super high-rise building. Its floor area is greater than the standard floor area of the main body. As shown in Figure 7, the pink rectangle and blue rectangle belong to one building. However, they have two different height. For simplicity, we call the building part below the main body as podium building. Apparently, we can not estimate the two heights in one pass matching processing.
Fortunately, the gray difference map of the matched blocks can provide more clues to refine the initial matching result. The gray difference map data represents the average brightness difference of backward and forward parallax. As shown in Figure 6, the gray difference histogram distribution should be in the range of (0–100) values if the matching result is completely correct. However, there are two peaks in the histogram as shown in Figure 8a, one is in the range of (0–100), and the other is in the range of (200–250). This indicates that there are two different heights in the building, and only one height is matched correctly. To solve this problem, we match the different height of a building with podium using the masked gray difference in an iterative manner. As shown in Figure 8a,b, by using two different masks, the bottom height of the podium is estimated in the first pass, and the top height of the podium is estimated in the second pass.
Figure 9 shows the principle of mask refinement and contour segmentation of podium building. First, the gray difference image (Figure 9a) is binarized to generate the new mask (Figure 9b). Second, an a connectivity analysis is conducted to erase small regions whose pixel numbers is smaller than a given threshold. The region with maximum number of pixels is selected as the area to be matched (Figure 9c). Third, the morphological filter is carried out on the selected area (Figure 9c) to get a refined mask (Figure 9d). Finally, one or several sub roof contours (Figure 9e) can be generated using the rectangles covering the sub regions of the refined mask, in which the covering rectangles are parallel to the original roof contour.
Based on above idea, the process of matching and segmentation of a building with podium is outlined as follows, as illustrated in Figure 10:
(1)
Set the mask as the whole roof using the blue contour polygon.
(2)
Search the x-parallax with the minimum masked gray difference in the standard deviation map;
(3)
Compute the gray difference map between matched image blocks;
(4)
Compute the histogram of the gray difference map in the mask set;
(5)
If there are more than one peaks in the histogram, the gray difference image is binarized and refined as a new mask for next matching.
(6)
Repeat steps 2 to 5 until 90% pixel is matched.

2.3. Bottom Elevation Estimation and Building Height Calculation

Three steps are used to estimate the building bottom evaluation and building height. First, combined with the matched x-parallax, the ground point of the building contour center can be computed using the intersection of the two homogeneous imaging rays of the stereo image pair [42]. Then using the RPCs of the backward image, the contour of the building roof on the backward image can be transformed into the object space as the ground footprint, which is on the horizontal plane defined by the elevation of the building center. Second, as shown in Figure 11, the bottom elevation is estimated by using elevation histogram analysis of DSM values, which is created in a 20 m buffer around the building footprint. Then the elevation at the minimum peak of the histogram is chosen as the bottom elevation. Third, the building height is obtained by subtracting the bottom elevation from the roof elevation.

3. Experiment and Results

3.1. Data and Study Area

3.1.1. Study Area

There are two datasets used in this study for performance evaluation, as shown in Figure 12. Dataset one locates in Yingde City, Guangdong Province, between 113.31–113.50°E, 24.25–24.41°N. This study area contains more than 8000 buildings with their footprints in shapefile format, whose height ranges from 10 m to 100 m. By using aerial LiDAR data, we estimated the roof elevation, bottom elevation, and building height as ground truth. With the GF-7 images, we used ENVI and INPHO to generate DSM, which can be used to estimate building height directly or as the input of our roof contour constrained matching method. Then, our result and DSM based result were compared with the LiDAR based ground truth. In addition, the shadow based result was also compared.
Dataset two locates in Xi’an high tech Industrial Development Zone, Shanxi Province, between 108.85–108.89°E, 34.18–34.24°N. In this area, we focus on high-rise buildings and podium buildings. In this test site, there are no building footprints and LiDAR data to generate ground truth. Instead, we took the building heights measured from GF-7 stereo images as reference values. We selected 40 buildings, which include 33 single buildings and 7 podium buildings. The height distribution of this dataset ranges from 20 m to 350 m, and the average building height is about 95.55 m, among which the Xi’an Guorui financial center with a height of 348.384 m is the highest building.

3.1.2. GF-7 Stereo Images

GF-7 satellite stereo images are used to extract building height. The parameters of used GF-7 satellite stereo images are presented in Table 1. GF-7 satellite is equipped with payloads, including a dual-line array camera and a laser altimeter. The three-line array camera can effectively obtain panchromatic stereo images with a width of 20 km and a resolution better than 0.8 m. In addition, infrared multispectral images with a resolution of 2.6 m can also be obtained. The infrared multi-spectral camera is aligned with the backward panchromatic camera for pansharpening. The backward direction is nearly nadir, designed for balance between less occlusion and bigger stereo intersection angle. Thus, the backward image is better for the extraction of building roof contours.

3.1.3. LiDAR Data

The LiDAR data used in this paper was acquired with Leica CityMapper-2 [43], which is an airborne hybrid sensor of oblique camera and LiDAR. The flight altitude is about 1450 m and the point cloud density in Yingde is 8.1 points/m 2 . As an example, a small area of LiDAR point cloud in Yingde is shown in Figure 13.

3.1.4. Building Roof Contours and Ground Truth

Our method supposes that building roof contours on the backward image are available for matching against the forward image. For the experiment purpose, building roof contours are generated from footprint data or digitized directly on the backward image. In the practical processing of city scale data, the building contour can be generated by semantic segmentation of backward image.
In the Yingde study area, roof contours are based on the building footprint, which were provided by Hubei Sunrise Photogrammetric Company. Based on the histogram analysis of the point cloud in the buffer of a building footprint, the roof elevation and bottom elevation of a building can be estimated, then the building footprint with roof elevation can be projection onto the backward image as the input roof contour for our workflow. Partial roof contours are shown in Figure 14a.
In the Xi’an study area, we directly digitized 40 buildings’ roof contours on the backward image with the ArcMap software, as shown in Figure 14b.
The ground truth of building height was calculated according to the roof elevation and bottom elevation. As mentioned ahead, the roof elevation and bottom elevation in the Yingde study area were estimated with LiDAR data. In the Xi’an study area, the roof elevation and bottom elevation were measured in a self-developed stereo model viewer software.

3.1.5. DSM Generation

The input DSM in the workflow(shown in Figure 1) is generated by a self-developed dense matching software based on the pyramid semi-global algorithm, which is silimar to SURE [44].
The DSM used in the Section 3.3.2 is generated by commercial software ENVI and INPHO, which are industry leading software.
As shown in Figure 15, five control points were selected from the LiDAR point cloud and the positioning accurarcy of GF-7 stereo image pair was imporved with RPC modification [45] before DSM generation.
For the ENVI, the parameters are set as follows: the minimum overlap is 55; the matching threshold is 15; the edge threshold is 5; the quality threshold is 60; the terrain type is Flat.
For INPHO generating DSM, the parameters are set as follows: the terrain type is Flat; the feature density is Dense; the point cloud density is 3 pixels; the Parallax threshold is 20 pixels. Three block of DSM results generated from ENVI and INPHO are shown in Figure 16.

3.2. Evaluation Metrics

Height accuracy is evaluated by comparing the estimated building height and the reference building height. Three metrics are used for performance evaluation, including the mean absolute error (MAE), root mean square error (RMSE), and max absolute error (maxAE), which are formulated as Equations (4)–(6):
MAE = 1 N i = 1 N h i h i ^
RMSE = 1 N i = 1 N h i h i ^ 2
m a x AE = m a x { | h i h i ^ | , i = 1 , 2 , , N }
where, h i is the estimated building, h i ^ is the reference building height.

3.3. Experimental Results of Yingde Dataset

In Yingde study area, there are a total number of 8653 buildings, whose height is less than 100 m. For accuracy evaluation, the ground truth of building height is generated from aerial LiDAR data. Table 2 presents the accuracy statistic of building height estimation. The results show that the MAE of the roof and bottom elevation estimation are 1.57 m and 1.91 m, respectively, and the accuracy of building height estimation is 2.31 m. The RMSE of building height is 3.01 m. Besides, Table 3 shows the number and accuracy statistic of building height at varying elevation ranges. The results show that the overall building roof elevation estimation is relatively stable, and the bottom elevation estimation accuracy fluctuates wildly since the DSM is easily affected by the occlusion of trees and buildings.

3.3.1. Comparison with the Shadow-Based Method

The proposed solution is compared with the shadow-based method in this section. Two sub datasets that are extracted from Yingde dateset, are used for evaluation. The first sub dataset is a low-rise building dataset with the height ranging between 0–30 m, and there are 299 buildings in the low-rise building dataset; the second sub dataset is a high-rise building dataset with the height larger than 30 m, and there are 170 buildings in the high-rise building dataset. According to Xie [20], when the ground height is horizontal, the shadow length of a building is directly proportional to the height of the building. Based on this observation, we measure the shadow length of buildings manually to avoid the deficiencies of automatic extraction methods. Because the inclined angles of forward camera and backward camera are different in GF-7 satellite, we estimate the building height on both the forward image and the backward image for extensive comparison.
The statistic results of shadow-based method and the proposed solution are shown in Table 4 and Table 5. Table 4 shows the comparison of accuracy statistics of building height. According to the experimental results, we can conclude that: (1) the MAE of the shadow estimation method on the high-rise building dataset is less than that of the low-rise building; (2) In low-rise building dataset, due to the occluded image in the shadow area, 10% buildings in the forward and 36% buildings in the backward participate in the metric calculation. The average accuracy of the proposed solution in terms of MAE, RMSE and maxAE is significantly higher than that of the shadow-based method; (3) In high-rise building dataset, 45% the buildings in forward and 59% the buildings in backward participate in the calculation. Similarly, the accuracy of the proposed solution is significantly higher than that of the shadow-based method for the three metrics.
In addition, Table 5 shows the comparison of number statistic of building height at varying elevation ranges, in which the term occlusion indicates the number of buildings that occluded. we can see that for the low-rise building dataset, 268 (89%) buildings have shadow occlusions on the forward image, and 190 (64%) buildings have shadow occlusions on the backward image; for the high-rise building dataset, 93 (55%) buildings have shadow occlusions on the forward image, and 69 (40%) buildings have shadow occlusions on the backward image. The height of all the occluded buildings cannot be estimated. On the contrary, the proposed method can successfully estimate the height of all buildings.
For a visual interpretation, Figure 17 illustrates some shadow occlusion cases: 1 Sheltered by trees; 2 blocked by nearby buildings. As shown in Figure 17a,c, the shadows at the edge of the house are blocked by trees, which belongs to case 1; On the backward image in Figure 17b, the shadow is not obscured due to the small inclination of the camera; however, in Figure 17d, the shadow on the forward image is occluded by the buildings in the backward and forward rows. It can be seen that the shadow-based method has obvious limitations when applied to GF-7 images. In a word, a majority of buildings cannot be extracted by using the shadow-based method due to the serious occlusion that occur in GF-7 satellite images, especially for the forward image.

3.3.2. Comparison with DSM Based Method

In this section we compare the proposed solution with DSM Based Method. We used ENVI (version 5.6) and INPHO (version 12.1) to generate DSM in Yingde. After generating the DSM, we overlaped the building contour vector onto the DSM for building height extraction. The roof elevation and the bottom elevation are estimated with histogram analysis, which has been discussed in the Section 2.3. The peaks of maximum elevation and minimum elevation are selected as of the building roof and the bottom. Shown in Figure 16, DSM generated by INPHO is relatively smooth, however, it loses some high-rise heights. On the contrary, the DSM generated by ENVI can retain high-rise buildings better, however, it can contain some holes.
Table 6 lists the accuracy statistic of building roof elevation from the DSM based methods, and the distribution of absolute error (AE) of ENVI and INPHO is presented in Figure 18. The results show that ENVI achieves better accuracy than INPHO for building roof height estimation in the two datasets. For both ENVI and INPHO, the estimation accuracy of low-rise buildings is better than that of high-rise buildings. As was expected, our method achieves the highest precision. In a conclusion, INHPO and ENVI have limitations in roof estimation of high-rise buildings when applied to GF-7 images, and the performance of our method is very stable for both high and low buildings. Figure 19 shows the 3D reconstruction models of the proposed solution in this dataset.

3.4. Experimental Results of Xi’an Dataset

In the dataset of Xi’an, there are a total number of 40 buildings, in which 7 buildings are with podium. Similar to the procedure presented in Section 3.3.2, the proposed method is compared with DSM based results of ENVI and INPHO. Table 7 shows the accuracy statistic for this dataset. Figure 20 shows the correlation the estimated building heights and their reference values, in which the dots nearer the dashed line have the higher estimation precision. From the results listed in Table 7, we can conclude that our method achieves the best performance, and obviously improves the accuracy of building height estimation; its MAE and RMSE of building height estimation is 1.69 m and 2.34 m, respectively. By further observation from Figure 20, we can see that within the range of 0–100 m, ENVI can achieve comparable accuracy when compared with the proposed method. However, with the increase of building height, its performance dramatically decrease. It can be seen from the green dots, which are nearly on the dashed line, our method archives consistently high precision within the entire building height range.

Building Height Estimation of Buildings with Podium

In the Xi’an dataset, the building heights of the 7 podium buildings are evaluated, in which more than one roof elevations are extracted using our iterative roof matching method. For performance evaluation, the height of each podium building is divided into three parts, and their heights are measured manually from GF-7 stereo image as the reference values.
Table 8 shows the statistical results of building height estimation, which lists the value of the first estimated building height by using the term Whole and height for each part after the podium segmentation by using the terms Part1, Part2 and Part3. Besides, the empty values in Table 8 represent that the part of the second podium is completely covered by other parts, and the ground truth building height cannot be estimated, as shown in Figure 21. The results show that for each part of the building with podium, the proposed method by using podium segmentation can find several elevation planes in a building contour, and can also accurately estimate their height. By the further analysis, we conclude that the main error of height estimation comes from the matching error caused by partial or entire occlusion of building roofs. Figure 22 shows the 3D reconstruction models of the proposed solution in this dataset.

4. Conclusions

This study proposes a method for extracting building height information from GF-7 satellite stereo images. First, an object-oriented roof matching algorithm is proposed based on building contour to extract accurate building roof elevation from GF-7 stereo image, and DSM generated by business software is then used to obtain building bottom elevation. Second, to cope with buildings with multiple level height plane, a mask gray difference metric is proposed to search multiple elevation planes and segment the building contour. Finally, by using ground truth data from LiDAR point clouds or manually measured building height, the proposed solution is extensively evaluated and compared with shadow-based method and DSM based method. The experimental results demonstrate that the proposed solution can achieve the best performance and could be an useful solution for accurate and automatic building height information extraction from GF-7 satellite stereo images.
Compared with the DSM and shadow-based method, our algorithm skillfully solves the problem of height estimation of high-rise buildings. Compared with deep learning, this algorithm does not need training data set. For the first time, we focused on the problem of multiple elevation planes within the height of a building.

Author Contributions

Conceptualization, C.Z., W.J. and Y.C.; Methodology, C.Z., W.J. and Y.C.; Software, C.Z. and W.J.; Visualization, C.Z.; Validation, C.Z. and Z.Z.; Formal analysis, C.Z.; Investigation, C.Z.; Resources, W.J.; Data curation, C.Z. and Z.Z.; Writing—original draft preparation, C.Z.; Writing—review and editing, C.Z., W.J. and S.J.; Supervision, W.J.; Project administration, C.Z. and W.J.; Funding acquisition, W.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Fund Project: High-Resolution Remote Sensing Application Demonstration System for Urban Fine Management. Grant number is 06-Y30F04-9001-20/22.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, A.J.; Xu, G.Y.; Shi, Y.C. Automatic 3D building modeling based on urban aerial stereo pairs. J. Surv. Mapp. 2002, 1, 54–59. [Google Scholar]
  2. Xia, M.; Jia, K.; Zhao, W.; Liu, S.; Wei, X.; Wang, B. Spatio-temporal changes of ecological vulnerability across the Qinghai-Tibetan Plateau. Ecol. Indic. 2021, 123, 107274. [Google Scholar]
  3. Qin, R. Change detection on LOD 2 building models with very high resolution spaceborne stereo imagery. ISPRS J. Photogramm. Remote Sens. 2014, 96, 179–192. [Google Scholar]
  4. Tomás, L.; Fonseca, L.; Almeida, C.; Leonardi, F.; Pereira, M. Urban population estimation based on residential buildings volume using IKONOS-2 images and lidar data. Int. J. Remote Sens. 2016, 37, 1–28. [Google Scholar]
  5. Xie, Y.; Weng, A.; Weng, Q. Population estimation of urban residential communities using remotely sensed morphologic data. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1111–1115. [Google Scholar]
  6. Xu, M.; Cao, C.; Jia, P. Mapping fine-scale urban spatial population distribution based on high-resolution stereo pair images, points of interest, and land cover data. Remote Sens. 2020, 12, 608. [Google Scholar]
  7. Berger, C.; Rosentreter, J.; Voltersen, M.; Baumgart, C.; Schmullius, C.; Hese, S. Spatio-temporal analysis of the relationship between 2D/3D urban site characteristics and land surface temperature. Remote Sens. Environ. 2017, 193, 225–243. [Google Scholar]
  8. Venter, Z.S.; Brousse, O.; Esau, I.; Meier, F. Hyperlocal mapping of urban air temperature using remote sensing and crowdsourced weather data. Remote Sens. Environ. 2020, 242, 111791. [Google Scholar]
  9. Hang, J.; Li, Y.; Sandberg, M.; Buccolieri, R.; Di Sabatino, S. The influence of building height variability on pollutant dispersion and pedestrian ventilation in idealized high-rise urban areas. Build. Environ. 2012, 56, 346–360. [Google Scholar]
  10. Güneralp, B.; Zhou, Y.; Ürge-Vorsatz, D.; Gupta, M.; Yu, S.; Patel, P.L.; Fragkias, M.; Li, X.; Seto, K.C. Global scenarios of urban density and its impacts on building energy use through 2050. Proc. Natl. Acad. Sci. USA 2017, 114, 8945–8950. [Google Scholar]
  11. Facciolo, G.; De Franchis, C.; Meinhardt-Llopis, E. Automatic 3D reconstruction from multi-date satellite images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 57–66. [Google Scholar]
  12. Bullinger, S.; Bodensteiner, C.; Arens, M. 3D Surface Reconstruction From Multi-Date Satellite Images. arXiv 2021, arXiv:2102.02502. [Google Scholar]
  13. Qin, R.; Song, S.; Ling, X.; Elhashash, M. 3D reconstruction through fusion of cross-view images. In Recent Advances in Image Restoration with Applications to Real World Problems; IntechOpen: London, UK, 2020; p. 123. [Google Scholar]
  14. Gui, S.; Qin, R. Automated LoD-2 model reconstruction from very-high-resolution satellite-derived digital surface model and orthophoto. ISPRS J. Photogramm. Remote Sens. 2021, 181, 1–19. [Google Scholar]
  15. Zhao, Y.D.; Wang, X.X.; Bai, Y. High resolution remote sensing image building extraction in combination with morphological reconstruction. Bull. Surv. Mapp. 2012, s1, 253–254. [Google Scholar]
  16. Tian, Y.J. Application prospect of high score satellite in surveying and mapping. Sci. Technol. Innov. 2020, 22, 160–161. [Google Scholar]
  17. Zhao, Z.M.; Zhou, X.C.; Fu, Q.K.; Wang, X.Q. Building height information extraction method based on ZY-3 image. Land Resour. Remote Sens. 2015, 3, 19–24. [Google Scholar]
  18. Shi, Y.F.; Wang, X.Q.; Sun, Z.H.; Chen, Y.Z.; Fu, Q.K. Urban building height estimation based on shadow ZY-3 satellite data. J. Earth Inf. Sci. 2015, 2, 236–243. [Google Scholar]
  19. Tian, F.; Chen, D.H.; Huang, X.L.; Li, H.; Yao, G.H.; Zhao, J.P.; Ding, F.; Xing, F. Building height estimation of GF2 image based on morphological shadow index. Remote Sens. Technol. Appl. 2017, 5, 844–850. [Google Scholar]
  20. Xie, Y.K. Research on Building Height Information Extraction Method of Multiple Scenes Based on High Resolution Image Shadow. Master’s Thesis, Southwest Jiaotong University, Chengdu, China, 2018. [Google Scholar]
  21. Chen, C. Height inversion of urban buildings based on Google Earth remote sensing images. Bull. Surv. Mapp. 2020, 1, 98–101. [Google Scholar]
  22. Wang, X.D.; Duan, F.Z.; Qu, X.Y.; Li, D.; Yu, P.F. UAV data building extraction based on object-oriented and SVM. Land Resour. Remote Sens. 2017, 1, 97–103. [Google Scholar]
  23. Wu, J.; Yao, Z.X.; Cheng, M.M. Dense matching of oblique aerial images fused with sift and SGM. J. Remote Sens. 2015, 3, 431–442. [Google Scholar]
  24. Li, Y.; Zhu, L.; Gong, P.; Shimamura, H. A refined marker controlled watershed for building extraction from DSM and imagery. Int. J. Remote Sens. 2010, 31, 1441–1452. [Google Scholar]
  25. Shaker, I.F.; Abd-Elrahman, A.; Abdel-Gawad, A.K.; Sherief, M.A. Building Extraction from High Resolution Space Images in High Density Residential Areas in the Great Cairo Region. Remote Sens. 2021, 3, 781–791. [Google Scholar]
  26. Alobeid, A.; Jacobsen, K.; Heipke, C. Building height estimationin urban areas from very high Resolution satellite stereo images. Proc. ISPRS 2009, 5, 2–5. [Google Scholar]
  27. Li, X.H. Research on urban building height extraction method based on DSM. Nat. Resour. North China 2020, 5, 97–98. [Google Scholar]
  28. Wang, J.Y.; Hu, X.L.; Meng, Q.Y.; Zhang, L.L.; Wang, C.Y.; Liu, X.C.; Zhao, M.F. Developing a Method to Extract Building 3D Information from GF-7 Data. Remote Sens. 2021, 22, 4532. [Google Scholar]
  29. Hirschmuller, H. Accurate and efficient stereo processing by semi-global matching and mutual information. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 807–814. [Google Scholar]
  30. Liu, C.J.; Krylov, V.A.; Kane, P.; Kavanagh, G.; Dahyot, R. IM2ELEVATION: Building Height Estimation from Single-View Aerial Imagery. Remote Sens. 2020, 12, 2072–4292. [Google Scholar]
  31. Tao, R.S.; Xiang, Y.M.; You, H.J. An Edge-Sense Bidirectional Pyramid Network for Stereo Matching of VHR Remote Sensing Images. Remote Sens. 2020, 12, 4025. [Google Scholar]
  32. Pang, J.H.; Sun, W.X.; Ren, J.S.J.; Yang, C.X.; Yan, Q. Cascade Residual Learning: A Two-stage Convolutional Neural Network for Stereo Matching. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 878–886. [Google Scholar]
  33. Liu, J.; Ji, S.P. A Novel Recurrent Encoder-Decoder Structure for Large-Scale Multi-view Stereo Reconstruction from An Open Aerial Dataset. In Proceedings of the 2020 IEEE CVF Conference Oncomputer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 6049–6058. [Google Scholar]
  34. Yao, Y.; Luo, Z.X.; Li, S.W.; Shen, T.W.; Fang, T.; Quan, L. Recurrent mvsnet for high-resolution multiview stereo depth inference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5525–5534. [Google Scholar]
  35. Yang, X.B.; Lv, J.G.; Jiang, S.; Zhang, D.L. Improved semiglobal matching generation method for high resolution remote sensing image DSM. J. Surv. Mapp. 2018, 10, 1372–1384. [Google Scholar]
  36. Tao, L.H.; Ma, J.; Li, N.N.; Li, C.N. Application of PCI GXL and INPHO software in GF7 image production. Geospat. Inf. 2021, 8, 90–96. [Google Scholar]
  37. Bosch, M.; Foster, K.; Christie, G. Semantic Stereo for Incidental Satellite Images. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision, Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1524–1532. [Google Scholar]
  38. Cao, Y.X.; Huang, X. A deep learning method for building height estimation using high-resolution multi-view imagery over urban areas: A case study of 42 Chinese cities. Remote Sens. Environ. 2021, 12, 19–24. [Google Scholar]
  39. Karatsiolis, S.; Kamilaris, A.; Cole, I. IMG2nDSM: Height Estimation from Single Airborne RGB Images with Deep Learning. Remote Sens. 2021, 13, 2417. [Google Scholar]
  40. Liu, J.; Ji, S.P. Dense matching of aerial remote sensing images based on deep learning. J. Surv. Mapp. 2019, 9, 1141–1150. [Google Scholar]
  41. Pan, H.B.; Zhang, G.; Chen, T. A general method of generating satellite epipolar images based on rpc model. In Proceedings of the 2011 IEEE International Geoscience And Remote Sensing Symposium (IGARSS), Vancouver, BC, Canada, 24–29 July 2011; pp. 3015–3018. [Google Scholar]
  42. Zhang, J.Q. Forward intersection method and accuracy analysis of multi baseline photogrammetry. J. Wuhan Univ. (Inf. Sci. Ed.) 2007, 10, 847–857. [Google Scholar]
  43. New Leica City Mapper Configuration for Challenging Digital Twin Creation. GIM Int.— Worldw. Mag. Geomat. 2021, 35, 7.
  44. Rothermel, M.; Wenzel, K.; Fritsh, D.; Haala, N. SURE: Photogrammetric Surface Reconstruction from Imagery. In Proceedings of the LC3D Workshop, Berlin, Germany, 4–5 December 2012. [Google Scholar]
  45. Fraser, C.; Hanley, H.; Yamakawa, T. Three-Dimensional Geopositioning Accuracy of Ikonos Imagery. Photogramm. Rec. 2002, 17, 465–479. [Google Scholar]
Figure 1. The overall workflow of the building height extraction solution.
Figure 1. The overall workflow of the building height extraction solution.
Remotesensing 14 01566 g001
Figure 2. The principle of parallex search range calculation.
Figure 2. The principle of parallex search range calculation.
Remotesensing 14 01566 g002
Figure 3. Illustration of standard deviation image extraction using convolution processing.
Figure 3. Illustration of standard deviation image extraction using convolution processing.
Remotesensing 14 01566 g003
Figure 4. Illustration of the image feature map. (a) building roof image; (b) extracted feature map.
Figure 4. Illustration of the image feature map. (a) building roof image; (b) extracted feature map.
Remotesensing 14 01566 g004
Figure 5. Epipolar image blocks and their standard deviation maps. (a) backward image block; (b) forward image block; (c,d) standard deviation maps of image blocks (a,b).
Figure 5. Epipolar image blocks and their standard deviation maps. (a) backward image block; (b) forward image block; (c,d) standard deviation maps of image blocks (a,b).
Remotesensing 14 01566 g005
Figure 6. A single building matching based on masked gray difference.
Figure 6. A single building matching based on masked gray difference.
Remotesensing 14 01566 g006
Figure 7. Backward and forward view of a podium building. (a) podium on backward image; (b) podium on forward podium.
Figure 7. Backward and forward view of a podium building. (a) podium on backward image; (b) podium on forward podium.
Remotesensing 14 01566 g007
Figure 8. Roof matching based on masked gray difference. (a) first pass matching for a podium building; (b) second pass matching the podium building in (a). The blue part of the masks represent the area need to be matched; the pink part of the mask represents the matched area.
Figure 8. Roof matching based on masked gray difference. (a) first pass matching for a podium building; (b) second pass matching the podium building in (a). The blue part of the masks represent the area need to be matched; the pink part of the mask represents the matched area.
Remotesensing 14 01566 g008
Figure 9. Mask refinement and contour segmentation of podium building. (a) gray difference map after last pass matching; (b) binarization of gray difference map; (c) new mask after connectivity analysis; (d) refined mask after morphological filtering (e) polygon of sub roof contour.
Figure 9. Mask refinement and contour segmentation of podium building. (a) gray difference map after last pass matching; (b) binarization of gray difference map; (c) new mask after connectivity analysis; (d) refined mask after morphological filtering (e) polygon of sub roof contour.
Remotesensing 14 01566 g009
Figure 10. Podium Segmentation workflow.
Figure 10. Podium Segmentation workflow.
Remotesensing 14 01566 g010
Figure 11. Bottom elevation extraction. (a) the buffer of DSM; (b) the elevation histogram.
Figure 11. Bottom elevation extraction. (a) the buffer of DSM; (b) the elevation histogram.
Remotesensing 14 01566 g011
Figure 12. GF-7 images and building samples of the study areas.
Figure 12. GF-7 images and building samples of the study areas.
Remotesensing 14 01566 g012
Figure 13. A small area of LiDAR point cloud in Yinde.
Figure 13. A small area of LiDAR point cloud in Yinde.
Remotesensing 14 01566 g013
Figure 14. Self-annotated building roof contours on the GF-7 backward image. (a) the building roof contours in Yingde; (b) the countours in Xi’an.
Figure 14. Self-annotated building roof contours on the GF-7 backward image. (a) the building roof contours in Yingde; (b) the countours in Xi’an.
Remotesensing 14 01566 g014
Figure 15. Ground control points selected for the Yingde stereo image pair.
Figure 15. Ground control points selected for the Yingde stereo image pair.
Remotesensing 14 01566 g015
Figure 16. Epipolar image blocks and the DSM generated by ENVI and INPHO.
Figure 16. Epipolar image blocks and the DSM generated by ENVI and INPHO.
Remotesensing 14 01566 g016
Figure 17. The illustration of shadow occlusions. (a,c) building shadows occluded by trees; (b,d) building shadows occluded by buildings.
Figure 17. The illustration of shadow occlusions. (a,c) building shadows occluded by trees; (b,d) building shadows occluded by buildings.
Remotesensing 14 01566 g017
Figure 18. Error distribution of DSM based roof elevation.
Figure 18. Error distribution of DSM based roof elevation.
Remotesensing 14 01566 g018
Figure 19. The 3D reconstruction models of the proposed solution in Yingde.
Figure 19. The 3D reconstruction models of the proposed solution in Yingde.
Remotesensing 14 01566 g019
Figure 20. The distribution of estimated and ground truth building elevation.
Figure 20. The distribution of estimated and ground truth building elevation.
Remotesensing 14 01566 g020
Figure 21. The illustration of the second Podium structure. The red box indicates the coverd part on backward image.
Figure 21. The illustration of the second Podium structure. The red box indicates the coverd part on backward image.
Remotesensing 14 01566 g021
Figure 22. The 3D reconstruction models of the proposed solution in Xi’an. The red box indicates the podium buildings.
Figure 22. The 3D reconstruction models of the proposed solution in Xi’an. The red box indicates the podium buildings.
Remotesensing 14 01566 g022
Table 1. Characteristics of the GF-7 satellite stereo images.
Table 1. Characteristics of the GF-7 satellite stereo images.
ParameterValue
Forward camera inclination26°
Backward camera inclination
Panchromatic resolutionBackward 0.65 m, Forward 0.8 m
Multispectral resolutionBackward 2.6 m
Width of Windows≥20 km
Table 2. Accuracy statistic of building height estimation.
Table 2. Accuracy statistic of building height estimation.
IndexRoof ElevationBottom ElevationBuilding Height
MAE (m)1.571.912.31
RMSE (m)2.012.573.01
maxAE (m)12.674.999.97
Table 3. Number and accuracy statistic at varying elevation ranges.
Table 3. Number and accuracy statistic at varying elevation ranges.
Height RangeBuilding NumbersMAE
Roof ElevationBottom ElevationBuilding Height
0–10 m57741.561.662.05
10–20 m24661.592.322.76
20–30 m3051.482.733.07
30–40 m481.563.293.75
40–50 m211.294.394.85
50–60 m171.123.203.65
60–70 m82.271.692.71
70–80 m81.241.471.98
80–90 m60.763.672.90
90–100 m11.994.316.31
Table 4. The comparison of accuracy statistic of building height.
Table 4. The comparison of accuracy statistic of building height.
Para.Building < 30 mBuilding > 30 m
ForwardBackwardOursForwardBackwardOurs
MAE (m)2.633.021.343.723.091.43
RMSE (m)3.4814.931.774.606.971.90
maxAE (m)8.1113.504.759.598.594.63
Table 5. The comparison of building number statistic at varying height ranges.
Table 5. The comparison of building number statistic at varying height ranges.
AEBuilding < 30 mBuilding > 30 m
ForwardBackwardOursForwardBackwardOurs
<2 (m)16532292538127
2–4 (m)82265252330
4–6 (m)314571213
>6 (m)420020280
occlusion268190093690
Table 6. Accuracy statistics of building roof elevation.
Table 6. Accuracy statistics of building roof elevation.
Para.Building < 30 mBuilding > 30 m
ENVIINPHOOursENVIINPHOOurs
MAE (m)5.126.531.3415.632.651.43
RMSE (m)6.138.641.7728.1938.701.90
maxAE (m)26.6735.954.7582.8789.864.63
Table 7. Accuracy statistics of experimental results in the Xi’an.
Table 7. Accuracy statistics of experimental results in the Xi’an.
IndexBuilding Height
ENVIINPHOOurs
MAE (m)41.4655.201.69
RMSE (m)77.86100.642.34
maxAE (m)326.38345.387.47
Table 8. Accuracy statistics of Podium building results in the study area.
Table 8. Accuracy statistics of Podium building results in the study area.
IndexWholePart1Part2Part3
Ours (m)Truth (m)Ours (m)Truth (m)Ours (m)Truth (m)Ours (m)
1106.19108106.191618.86
274.71106105.18--
360.497570.657570.65109.72
417.855249.332518.86
530.045550.351613.79
64.658786.90127.70
74.659390.6524.65
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, C.; Cui, Y.; Zhu, Z.; Jiang, S.; Jiang, W. Building Height Extraction from GF-7 Satellite Images Based on Roof Contour Constrained Stereo Matching. Remote Sens. 2022, 14, 1566. https://doi.org/10.3390/rs14071566

AMA Style

Zhang C, Cui Y, Zhu Z, Jiang S, Jiang W. Building Height Extraction from GF-7 Satellite Images Based on Roof Contour Constrained Stereo Matching. Remote Sensing. 2022; 14(7):1566. https://doi.org/10.3390/rs14071566

Chicago/Turabian Style

Zhang, Chenni, Yunfan Cui, Zeyao Zhu, San Jiang, and Wanshou Jiang. 2022. "Building Height Extraction from GF-7 Satellite Images Based on Roof Contour Constrained Stereo Matching" Remote Sensing 14, no. 7: 1566. https://doi.org/10.3390/rs14071566

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop