Next Article in Journal
A New Method for Analysing the Pressure Response Delay in a Pneumatic Brake System Caused by the Influence of Transmission Pipes
Next Article in Special Issue
From Point Clouds to Building Information Models: 3D Semi-Automatic Reconstruction of Indoors of Existing Buildings
Previous Article in Journal
Ultralow Friction of ZrO2 Ball Sliding against DLC Films under Various Environments
Previous Article in Special Issue
Axial Fan Blade Vibration Assessment under Inlet Cross-Flow Conditions Using Laser Scanning Vibrometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LiDARgrammetry: A New Method for Generating Synthetic Stereoscopic Products from Digital Elevation Models

by
Ricardo Rodríguez-Cielos
1,
José Luis Galán-García
2,*,
Yolanda Padilla-Domínguez
2,
Pedro Rodríguez-Cielos
2,
Ana Belén Bello-Patricio
3 and
José Antonio López-Medina
1
1
Universidad Politécnica de Madrid, University of Madrid, 28040 Madrid, Spain
2
Escuela de Ingenierías Industriales, University of Málaga, 29071 Málaga, Spain
3
Universidad Complutense de Madrid, University of Madrid, 28040 Madrid, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2017, 7(9), 906; https://doi.org/10.3390/app7090906
Submission received: 16 May 2017 / Revised: 10 August 2017 / Accepted: 4 September 2017 / Published: 12 September 2017
(This article belongs to the Special Issue Laser Scanning)

Abstract

:
There are currently several new technologies being used to generate digital elevation models that do not use photogrammetric techniques. For example, LiDAR (Laser Imaging Detection and Ranging) and RADAR (RAdio Detection And Ranging) can generate 3D points and reflectivity information of the surface without using a photogrammetric approach. In the case of LiDAR, the intensity level indicates the amount of energy that the object reflects after a laser pulse is transmitted. This energy mainly depends on the material and the wavelength used by LiDAR. This intensity level can be used to generate a synthetic image colored by this attribute (intensity level), which can be viewed as a RGB (red, green and blue) picture. This work presents the outline of an innovative method, designed by the authors, to generate synthetic pictures from point clouds to use in classical photogrammetric software (digital restitution or stereoscopic vision). This is conducted using available additional information (for example, the intensity level of LiDAR). This allows mapping operators to view the LiDAR as if it were stereo-imagery, so they can manually digitize points, 3D lines, break lines, polygons and so on.

1. Introduction

At present, there are a number of diverse techniques for generating digital elevation models (such as terrain or surface models) that do not require images to be directly captured. Some of these technologies that are widely used include LiDAR, echo sounder and laser scanner [1].
All these techniques can generate dense point clouds with defined spatial attributes. However, these point clouds cannot be displayed by classical methods of stereoscopic vision and, thus, they are incompatible with the software used in classic photogrammetric stations. This software usually allows us to visualize images with stereoscopic vision (using 3D glasses) to digitalize elements, such as points, lines or polygons. These tools are necessary for the process of cartographic production and they are not available in software for computing point clouds (with no 3D vision available). Therefore, when the images are not available, we cannot use photogrammetric software to conduct this process and a synthetic stereoscopic pair (photogrammetric model) must be generated from the available point cloud.
The adjective ‘synthetic’ is used to describe this stereoscopic pair for the following reason. As no radiometric information is available in the visible range, the generated image will have a specific characteristic provided by the sensor used (LiDAR intensity level, terrain reflectivity, etc.). This is achieved assigning a “false color” to the images obtained in the process in order to ensure correct 3D visualization [2].
LiDARgrammetry is a term coined by GeoCue for using the intensity information as the source data for generating stereo-models of the project area directly from the LiDAR data itself rather than from ancillary imagery. By creating appropriate stereo-pairs from the LiDAR intensity images, it is possible to generate stereo-models that can be directly exploited using existing photogrammetric workstations and the traditional photogrammetric workflow in areas, such as feature extraction or break line compilation [3].
LiDARgrammetry concerns the production of inferred stereo-pairs from LiDAR intensity images, which aims to stereo-digitize the spatial data in digital photogrammetric stations. The production of inferred stereo-pairs is based on the principle of stereo ortho-mates and the extraction of a derivative intensity image, in which an artificial x-parallax is introduced. Furthermore, other techniques have also developed in order to fully utilize the 3D nature of LiDAR data. LiDARgrammetry aims to quantify the derivative spatial data quality and the impact of the reduced photo-interpretative ability of LiDAR data with comparisons to typical photogrammetric stereo-models [4]. However, this relatively new approach has not been yet assessed properly.
The LiDARgrammetry technique is based on the inverse algorithm of the photogrammetric technique, which is namely the 3D dataset (point cloud) that is used to generate 2D information (synthetic images). Therefore, the resolution and definition of the images depend on the density (points per square meters) and radiometric characteristics (intensity level) of the LiDAR dataset as well as the resampling algorithm used. The images are generated as if they had been registered at a given position using the “base–height” ratio related to the dataset’s parameters, which are used as a base reference.
The objective of the work presented here is to design and develop a suitable methodology for generating and visualizing 3D images from point clouds. Within this context, the “radargrammetry” technique [5] can be considered, which is similar to the photogrammetric one as it allows us to display point clouds using photogrammetric stations via the generation of synthetic images.
The purpose of this methodology is to create an alternative (cheap and easy) method for photogrammetry in cases where it is not possible to use photogrammetry (for example, in adverse weather conditions).
Restitution is often applied in a considerable number of work environments, including engineering, architecture, archaeology and cartography. The different types of software for restitution allow the operator to identify some points, lines and polygons to create vector layers using 3D glasses. In general, the software used for processing point clouds does not allow this functionality. Thus, it is very useful to obtain a pair of synthetic pictures using the methodology that we propose in this article as these can be used to create these new layers. Digital restitution provides shapes and sizes for engineering and architecture as well as different-scaled maps of cities, continents or celestial objects. The present study used these techniques to develop a specific software, in order to generate synthetic images and obtain some break lines of interest, using digital restitution DIGI 3D software from a LiDAR flight (SHAKE, I+D project financed in public announcements) [6].

2. Methodology to Generate a Synthetic Image

This paper describes a methodology that will allow us to extract photogrammetric images from any point cloud, which can then be used as input information. Figure 1 shows a block diagram of the methodology designed by the authors for generating synthetic images.
This methodology must be carried out in a series of phases.

2.1. Treatment of Initial Data

Initially, the starting data is a cloud of LiDAR points. The aspect of the point cloud can be seen in Figure 2, which is shown as the areas with a greater density of points (dark zones) when compared to other areas where there is a smaller density of points. The bottom part of Figure 2 shows a dark zone with a double density of points, which corresponds with the overlap zone for LiDAR flight (right zoom of Figure 2).
Usually, in this type of point cloud, the density is calculated in points per square meter, which is an average density for the whole cloud in practice.
The attributes that define a LiDAR point cloud are described below (this development is similar for any other type of point cloud):
  • Planimetric coordinates ( X , Y ) in a given reference system;
  • Height ( Z ) in a given reference system;
  • Intensity level or another attribute as reflectivity.
The point cloud could be irregular and, from it, two digital elevation models can be extracted:
A digital surface model (or ground, if this is the case) using the planimetric coordinates ( X ,   Y ) and the height ( Z ) . The mathematical expression is:
N Z = { ( X 1 , Y 1 , Z 1 ) , ( X 2 , Y 2 , Z 2 ) , , ( X n , Y n , Z n ) } .
A digital elevation model, using the planimetric coordinates ( X ,   Y ) and the intensity level ( I ) . The mathematical expression is:
N I = { ( X 1 , Y 1 , I 1 ) , ( X 2 , Y 2 , I 2 ) , , ( X n , Y n , I n ) } .
In both expressions, n is the number of elements in a point cloud.
The main target of this procedural step is to obtain a good density and a well-distributed point cloud in order to be similar to a photogram where all pixels have a regular distribution and the distance between two contiguous pixels is always the same. Therefore, all white spaces are filled with new points that have an intensity level similar to their neighbors.
In Figure 4, we define a reference system R t with a center O t and one generic terrain point P with coordinates ( X t , Y t , Z t ) referring to R t . We created two pictures ( I i ,   I j ) of the same place (containing P ) with two different points of view O i :   ( X t O i , Y t O i , Z t O i ) and O j :   ( X t O j , Y t O j , Z t O j ) , which refer to R t . The reflection of the rays in P are depicted as the orange lines in Figure 4 ( R y i and R y j ). These lines intersect I i   a n d   I j in P i and P j with coordinates ( X t i , Y t i , Z t i ) and ( X t j , Y t j , Z t j ) , respectively, over the same reference system R t . For all points P , the coplanarity condition [7] must be fulfilled ( O i , O j , P i , P j and P are all in the same plane, Π ) or Equation (3a) must be satisfied.

2.2. Delaunay Triangulation Applied to Digital Elevation Models

One of the most common problems, which will be resolved later on, is the need of LiDAR points in areas where this data is non-existent.
To solve this problem, the density of the LiDAR point cloud is increased by using a Delaunay triangulation [8] using the following process:
  • Align a point cloud (denoted as cloud N Z or cloud N I ) over a continuous mathematical surface consisting of flat and triangular elemental surfaces. Among the possibilities that need to be triangulated, there is an ideal solution for the problem we are addressing. This possibility consists of considering the option that ensures all the generated triangles will be as regular as possible and the length of their sides will be at its minimum. This type of triangulation is called “Delaunay triangulation”. The classic Delanuay triangulation can then be characterized as an optimal triangulation that minimizes the interpolation error for isotropic function among all the triangulations with a given set of vertices. For a more general function, a function dependent Delanuay triangulation is then defined to be an optimal triangulation that minimizes the interpolation error for this function and its construction can be obtained by a simple lifting and projection procedure. The Delanuay triangulation of a finite point cloud can be defined by the empty sphere property: no vertices in the point cloud are inside the circumsphere of any simplex in the triangulation. The triangulation formed this way will be unique and the result is a set of triangles with each one defined by three vertices within the points in the cloud [9].
  • Obviously, each of these triangles defines a plane in the space that must meet the requirements of Equation (3b), where ( X 1 ,   Y 1 ,   Z 1 ) , ( X 2 ,   Y 2 ,   Z 2 ) and ( X 3 ,   Y 3 ,   Z 3 ) denote the coordinates of the vertices of triangles over the reference system R t . Given any (X, Y) inside this triangle, we can calculate its Z (or its I) by interpolation, which enforces the equation of the specified plane. Using this method, we will have increased the density of the initial point cloud and the process can be repeated based on the new points obtained and triangulating once again (see Figure 3):
    | X X t Y Y t Z Z t X t O i Y t O i Z t O i X t O j Y t O j Z t O j | = 0   ,
    | X X 1 Y Y 1 Z Z 1 X 2 Y 2 Z 2 X 3 Y 3 Z 3 | = 0 .

2.3. Equalization of the Digital Images

This project uses grayscale images (false color images can also be displayed). There are two concepts associated with grayscale images, which are the brightness and the contrast.
In order for a digital image to be optimally visualized, it has to be contrasted properly and with a suitable brightness [10]. The synthetic digital images used in this paper are encoded with 8 bits. In other words, the digital values of the pixels must be in the range of 0–255 in order for these images to be displayed on a computer. As the images will be generated by using a specific characteristic of the point cloud (the intensity level in the case of LiDAR), this characteristic can have some values that do not correspond with the levels of 0–255, although they will have to be converted accordingly. This process is known as histogram equalization, which involves each pixel of the image being mapped to a different digital value depending only on its digital value.
The definition of a histogram [11] is given as a graphical representation of a frequency variable presented in the form of bars. In this case, the variable will be the frequency of the pixels according to their intensity within the area defined by the DSM (Digital Surface Model). The contrast of an image provides us with the scatter measurement for the histogram. In other words, a low-contrast image will have a concentrated histogram in the same area. Therefore, the objective will be to equalize the histogram in such a way as to achieve a suitable brightness and contrast, depending on the variable to be represented. To achieve this objective, the mean values are used together with the standard deviation of the variable to be represented (intensity in the case of LiDAR) to find the optimum equalization function.
Given a N I point cloud and any i point of this cloud, we have:
I i n f = I ¯ 1.5 · σ I I s u p = I ¯ + 1.5 · σ I ,
V D i = 255 · [ I i I i n f ] 3 · σ I   I f   I i n f I i I s u p V D i = 0   I f   I i I i n f V D i = 255   I f   I i I s u p ,
where I ¯ and σ I are the average and the standard deviation of the intensity values in the point cloud; I i is the value of intensity at point i in the cloud; and V D i is the equalized digital value for this same point i .
The selection of this function is not random. Considering that the frequency of repetition of the represented variable follows a normal standard distribution, we guarantee that more than 80% of the values of the variables are to be found within the interval I i n f I i I s u p by taking these indicators.

3. Methodology Development

After the above-explained process, the methodology is further developed following the steps listed below in sequential order:

3.1. Filling the Point Cloud, Calculation of Statistics and Magnitudes

The load of the point cloud will produce a set of n points denoted by N that can be mathematically expressed as:
N = { ( X 0 , Y 0 , Z 0 , I 0 ) , ( X 1 , Y 1 , Z 1 , I 1 ) , , ( X t , Y t , Z t , I t ) ,   , ( X n 1 , Y n 1 , Z n 1 , I n 1 ) } ,
where any point P t   N has coordinates ( X t ,   Y t ,   Z t ) with respect to a system of reference called   R t and with the center in O t . To construct the synthetic images (the intensity level in the case of LiDAR), the attribute I t   will be used.
Once the point cloud has been read, it is used to calculate the following statistics and magnitudes:
X m i n = M i n ( X 0 , X 1 , , X n 1 )   X m a x = M a x ( X 0 , X 1 , , X n 1 ) ,
Y m i n = M i n ( Y 0 , Y 1 , , Y n 1 )   Y m a x = M a x ( Y 0 , Y 1 , , Y n 1 ) ,
Z ¯ = i = 0 i = n 1 Z i n σ z = i = 0 i = n 1 ( Z i Z ¯ ) 2 n Z m i n = M i n ( Z 0 , Z 1 , , Z n 1 )   Z m a x = M a x ( Z 0 , Z 1 , , Z n 1 ) ,
I ¯ = i = 0 i = n 1 I i n σ I = i = 0 i = n 1 ( I i I ¯ ) 2 n I m i n = M i n ( I 0 , I 1 , , I n 1 )   I m a x = M a x ( I 0 , I 1 , , I n 1 ) ,
W t = ( X m a x X m i n ) H t = ( Y m a x Y m i n ) δ = n W t · H t S = W t · H t .

3.2. Calculation of the Initial Parameters

In this second step, the initial parameters needed to generate the synthetic models are calculated:
(a)
The GSD (Ground Sample Distance) is the resolution of the pixel on the ground. The problem is due to the infinite solutions, as, according to the values that we wish to give to the parameters, different images will be obtained. However, only one will be the optimal solution and this is what needs to be found. Given the density of the point cloud δ , calculated according to Equation (11), the best GSD (the GSD with the least magnitude) that can be found is given by the equation:
G S D = 1 δ .
In this way, we have at least one point per pixel. There are other possibilities, but this one is very simple, easy to use and easy to implement.
There are other parameters, which can be fixed according to the data that we already have. Moreover, they can be fixed freely (we have developed the methodology using a model of two frames or images). These parameters include:
(b)
The resolution of pixel r p ;
(c)
The focus of the cameras f o and f 1 ;
(d)
The width and height of the sensor W m and H m ;
(e)
The width and height of the image W and H ; and
(f)
The flight height h .
Once one or more of them have been fixed, the rest are calculated using Equations (13)–(15):
W m f i = W · G S D h   G S D = W m · h W · f i ,
G S D = r p · h f i ,
r p = W m W = H m H .
Parameters bf can be fixed by the operators. It is possible to assign any value to these parameters, but normally all restitution software has tolerances for the size of the digital files, digital camera work with a standard focal length, sensors for cameras have typical dimensions and there is a typical flight height. Therefore, we recommend that typical values are used for all these parameters. Using our own past experience, we first fixed the width and height for pictures to have a standard magnitude (not above 10,000 pixels for width or height), before continuing with focal length and flight length (standard values). Another important advantage of this methodology is that we can work with near 100% overlap. Therefore, most of the pixels in the synthetic images are useful for restitution (Figures 6 and 7). It is very difficult to simulate this situation in real photogrammetry. If one camera is used for all frames, it must be moved from one place to another and if two or more cameras are used at once, the same place must be targeted. Therefore, it is not easy to obtain near 100% overlap in photogrammetry (see Figure 4 where the frames are convergent not orthogonal). In this procedure, we can set the target.
It is possible to introduce (or fix according to all previous magnitudes) the values of the named parameters of the external orientation (Figure 4) for each of the two photographs from which the images I 0 and I 1 originate. These values are:
(a)
Coordinates of the projection centers O 0 and O 1 in the systems R t , ( X t O o ,   Y t O o ,   Z t O o ) and ( X t O 1 ,   Y t O 1 ,   Z t O 1 ) , respectively. A value (in the case that it is not fixed according to our needs) can be fixed according to the coating that is wished for and is called R e (in %) [7,12]:
X t O o = X m i n + X m a x X m i n 2 G S D · W · 100 R e 200 X t O 1 = X m i n + X m a x X m i n 2 + G S D · W · 100 R e 200 Y t O o = Y t O 1 = Y m i n + Y m a x Y m i n 2 Z t O o = Z t O 1 = Z ¯ + h .
These values are not random. In the best case, the photographs are taken from the center of the area that is being photographed and the flight height h in order to obtain an overlap on the given ground to determine a given perspective between two photographs (separation between the centers of projection) taken orthogonally from the ground.
(b)
Rotations of the cameras ω 0 ,   φ 0 , κ 0 and ω 1 ,   φ 1 , κ 1 that the photographs, respectively I 0 and I 1 , have been taken with. In the best case, they would all be 0, but an appropriate one can be fixed. It is necessary to establish this value in the case that an overlap near 100% is wanted; thus, this creates convergent photographs instead of orthogonal ones.

3.3. Reading the Point Cloud, Calculating the Pixel Coordinates and Generating Synthetic Pixel

This step is carried out for each point that belongs to the point cloud. The process is laid out in the following sections.

3.3.1. Calculation of the Digital Value of the Pixels to be Represented in Each of the Images I 0 and I 1

Given a point P t N whose attribute I is I t , the digital value V D t will be in each of the two images given by Equation (17) without having to consider that V D i = V D t and I i = I t :
V D i = 255 · [ I i I i n f ] 3 · σ I   I f   I i n f I i I s u p V D i = 0   I f   I i I i n f V D i = 255   I f   I i I s u p .
At this point, we have to calculate the equations of the rays denoted by R y 0 and R y 1 . Ray R y 0 passes through points P t and O o , while ray R y 1 passes through points P t and O 1 . Using Equations (18) and (19), we obtain Equations (20) and (21), which are the equations of lines R y 0 and R y 1 , respectively:
( X X t ) X t X t O i = ( Y Y t ) Y t Y t O i = ( Z Z t ) Z t Z t O i = l ,
( X X t ) X t X t O j = ( Y Y t ) Y t Y t O j = ( Z Z t ) Z t Z t O j = k ,
( X X t ) X t X t O 0 = ( Y Y t ) Y t Y t O o = ( Z Z t ) Z t Z t O o = l ,
( X X t ) X t X t O 1 = ( Y Y t ) Y t Y t O 1 = ( Z Z t ) Z t Z t O 1 = k .
The calculation of the equations of the planes denoted by I o and I 1 is shown in Figure 5. In this figure, the plane I 0 passes through points p : ( 0 ,   0 ,   f 0 ) , p : ( 0 ,   1 ,   f 0 ) and p : ( 1 ,   0 , f 0 ) in the system of reference R 0 , while the plane I 1 passes through points q : ( 0 ,   0 , f 1 ) , q : ( 0 ,   1 , f 1 ) and q : ( 1 ,   0 , f 1 ) in the system of reference R 1 (we checked that points p ,     p ,     p and points q ,     q ,     q are not aligned).
Equation (3a) can be applied to planes I 0 and I 1 to analytically obtain the equations of the planes that refer to the systems of reference R 0 and R 1 , respectively. However, the equations of those planes referring to the system of reference R t are required. Looking at plane I 0 (a similar reasoning as for plane I 1 ) defined by points p ,     p ,     p in the system of reference R 0 and applying a spatial transformation of seven parameters (where: λ = 1 ), the rotations are ω 0 ,     φ 0 ,   κ 0   and the translation is ( X t O o ,     Y t O o ,   Z t O o ) (Figure 5).
Following this, Equation (22) is applied:
[ X t Y t Z t ] = λ · R ω , φ , κ · [ X Y Z ] +   [ X 0 Y 0 Z 0 ] ,
where R ω , φ , κ is the so-called rotation matrix [13]; λ is the scaling factor/custom scale factor; [ X 0 Y 0 Z 0 ] is a vector that shows the translation; and [ X t Y t Z t ] are the transformed coordinates of point P of the coordinates [ X Y Z ] in the initial system.
The result is:
[ X p t Y p t Z p t ] = R ω 0 , φ 0 , κ 0 · [ 0 0 f 0 ] +   [ X t O o Y t O o Z t O o ] [ X p t Y p t Z p t ] = R ω 0 , φ 0 , κ 0 · [ 0 1 f 0 ] +   [ X t O o Y t O o Z t O o ] [ X p t Y p t Z p t ] = R ω 0 , φ 0 , κ 0 · [ 1 0 f 0 ] +   [ X t O o Y t O o Z t O o ] ,
[ X q t Y q t Z q t ] = R ω 1 , φ 1 , κ 1 · [ 0 0 f 1 ] +   [ X t O 1 Y t O 1 Z t O 1 ] [ X q t Y q t Z q t ] = R ω 1 , φ 1 , κ 1 · [ 0 1 f 1 ] +   [ X t O 1 Y t O 1 Z t O 1 ] [ X q t Y q t Z q t ] = R ω 1 , φ 1 , κ 1 · [ 1 0 f 1 ] +   [ X t O 1 Y t O 1 Z t O 1 ] ,
where the equations of planes I 0 and I 1 are given by the analytical expressions (applying Equation (3)), referring to system R t :
| X X p t Y Y p t Z Z p t X p t Y p t Z p t X p t Y p t Z p t | = 0 ,
| X X q t Y Y q t Z Z q t X q t Y q t Z q t X q t Y q t Z q t | = 0 .

3.3.2. Calculating Coordinates of Points p 0 and p 1

The coordinates of points p 0 and p 1 may now be calculated (referring to systems R 0 and R 1 ), which are the image coordinates of point P t of the ground in each image (Figure 4). Once they are transformed to pixel coordinates, the exact position (in column and row) of point P t of the ground data is generated in each image. This is conducted by calculating the intersection of plane I 0 with the line R y 0 (we already have both references from the same system of coordinates, R t ) and ensuring that they also meet the requirements of Equations (27) and (28):
( X X t ) X t X t O 0 = ( Y Y t ) Y t Y t O o = ( Z Z t ) Z t Z t O o = l ,
( X X t ) X t X t O 1 = ( Y Y t ) Y t Y t O 1 = ( Z Z t ) Z t Z t O 1 = k ,
| X X p t Y Y p t Z Z p t X p t Y p t Z p t X p t Y p t Z p t | = 0 ,
[ X p 0 Y p 0 Z p 0 ] = R ω 0 , φ 0 , κ 0 T · { [ X t 0 Y t 0 Y t 0 ] [ X t O o Y t O o Z t O o ] } [ X p 1 Y p 1 Z p 1 ] = R ω 1 , φ 1 , κ 1 T · { [ X t 1 Y t 1 Y t 1 ] [ X t O 1 Y t O 1 Z O 1 ] } .
This intersection will be a point p 0 on plane I 0 with coordinates (referring to system R t ) p 0 : ( X t 0 ,   Y t 0 ,   Z t 0 ) . The same procedure will be followed to calculate the intersection of plane I 1 with the line R y 1 (point p 1 on plane I 1 ) with the resulting coordinates needing to meet the requirements of both Equations (29) and (30), which means that the coordinates of point p 1 , referring to system R t , are p 1 : ( X t 1 ,   Y t 1 ,   Z t 1 ) . To solve this system, it is advisable to leave just one equation with the parameter l or k, calculate it and then take the coordinates). Now, we have the intersections (points p 0 and p 1 ) of Figure 4.

3.3.3. Transformation of Coordinates

After this, the coordinates of these points have to be transformed to the systems of reference R 0 and R 1 . To do this, we use a 3D transformation of seven parameters [13] (three translations, three turns and one scale factor), where Equation (31) is applied to obtain the results after the transformation:
R w i , φ i , κ i T = R w i , φ i , κ i 1 R w j , φ j , κ j T = R w j , φ j , κ j 1 .

3.3.4. Verification of Results

Following this, various checks can be completed to see whether the calculations have been properly compiled. We first checked whether the coordinates Z p 0 and Z p 1 from Equation (9) (copied below as Equation (32) for a better reference) should coincide with f 0 and f 1 because these are the points in the planes I 0 and I 1 , respectively:
Z ¯ = i = 0 i = n 1 Z i n , σ z = i = 0 i = n 1 ( Z i Z ¯ ) 2 n , Z m i n = M i n ( Z 0 , Z 1 , , Z n 1 ) ,   Z m a x = M a x ( Z 0 , Z 1 , , Z n 1 ) .
However, the plane defined by points P t O 0 and O 1 , should contain points p 0 : ( X t 0 ,   Y t 0 ,   Z t 0 ) and p 1 : ( X t 1 ,   Y t 1 ,   Z t 1 ) . The intersection on the lines R y 0 and R y 1 should be P t .

3.3.5. Transformation of Image Coordinates into Pixel Coordinates

The last step to do is to transform the image coordinates ( X p 0 ,     Y p 0 ) and the image coordinates ( X p 1 ,   Y p 1 ) into pixel coordinates. This is conducted based on Equation (33):
x = C O L · r p W m 2 , y = H m 2 ROW · r p .
Following this, Equation (34) is obtained:
C O L I 0 = X p 0 + W m 2 r p , R O W I 0 = H m 2 Y p 0 , r p
C O L I 1 = X p 1 + W m 2 r p R O W I 1 = H m 2 Y p 1 r p .
Finally, the digital value V D t , which was previously calculated, is assigned to each of these pixels in both images.

3.4. Densifying Point Cloud

When Step 3 is applied to all points of the cloud, a photogrammetric model is generated. The generated images will be similar to pictures shown in Figure 6.
In Figure 6, we can see that there is an insufficient number points to visualize the images in the digital photogrammetric station using photogrammetric software. This obvious lack of quality is due to the fact that the point cloud used is not a regular mesh. In other words, the density of points per square meter is not constant throughout the cloud and the points are not well-distributed. There are points on the ground that correspond to the same pixel in the images despite having different intensity levels. In this case, we discard all except one and we consider this digital value as the mean value of intensity level.
To resolve this issue and to obtain a well-distributed and uniform cloud of points, we could use a “nearest neighbor” type of algorithm in both images. With this algorithm, we could fill the areas with no points and we could create new points using the intensity level of neighbors. In our case, this is impossible because we would lose 3D vision as it would become noise in the digital image. The reason for this can be easily understood by analyzing Figure 4. If we create a new point P 0 h in I o using the nearest neighbor algorithm, the next stage would be to find this point ( P 1 h ) in image I 1 and assign the same digital value to it. However, this is not possible because we do not know the position of P 0 h in the terrain, so that it is impossible for us to find the homologous point in I 1 and 3D vision requires the same point in both images ( P 0 h and P 1 h ) The reconstruction of ray R y 1 is not available and, therefore, point P 1 h cannot be obtained.
The solution for this issue is to densify the initial point cloud N by adding new points as follows:
-
First of all, we must find all these zones so we used the software to automatically find all pixels with no value in both synthetic images (left and right).
-
We selected a zone with no points in the left image called H 0 with k neighbor pixels in this image (once we complete the process with the left image, we apply the same process to the right image) with pixel coordinates ( C O L k ,   R O W k ) and intensity level I L k . Following this, we obtained the ground coordinates ( X k , Y k ,   Z k ) for all k points, which are previously calculated in the process of generating the synthetic image as there is one unique correspondence between all points in the cloud, all points in image I o and all points in image I 1 . Following this, we created a new point P h (in the center of all k points) with ground coordinates ( X h , Y h , Z h ) and intensity level I L h , using the mean intensity level of all k points for digital value of intensity. After this, we have the new point P h , and, then, we can obtain pixel coordinates in both images I 0 and I 1 . We repeated this process as many times as necessary to get a well-distributed point cloud. The point cloud N has a new point called P h and the synthetic images are generated again with the initial point cloud in addition to the set of interpolated points created. Obviously, the holes are filled in both images.
The result obtained after completing this process in the images is shown in Figure 7. As you can see, the images of Figure 7 appear to be denser compared to the images of Figure 6.

4. Quality Control

Quality is defined as the “property inherent in anything that allows it to be compared with anything else of the same species or nature” [14]. For this purpose, certain procedures and controls have to be provided to guarantee the integrity, accuracy and entirety of the data. In other words, it is necessary to verify that the desired quality has been reached [14].
In the case of the information provided by the photogrammetric method, these are the classic aspects [15] related to redundancy measures. The results of the photogrammetric triangulation provide the quantitative measurements for the precision of the results:
-
Variance of the components;
-
Variance–covariance of the coordinates of the calculated objects;
-
Comparison of values with nominal data;
-
Independent measurements to verify the precision via a point control analysis; and
-
Comparison of coordinates of the photogrammetric points with the independently obtained coordinates (i.e., GPS in the field).
With regards to the LiDAR information, the control is a secondary procedure to ensure and verify the quality of the registered data. It can be conducted in terms of two criteria:
  • “Causes”: study the behavior of the elements that define the mathematical model of adjustment, where we consider both “internal causes” (flight plan: external orientation and calibration) and “external causes” (flight conditions: direction of the flight paths, flight height, etc.) as well as ground type (altitude, type of vegetation, etc.).
  • “Effects”: the results given by the point cloud and where the following has to be taken into consideration: the “relative internal effects” (topographical control and planimetric control) and “relative external effects” (control points, difference between Digital Terrain Model and DSM, differing intensity levels and stereoscopic checking). In most cases, the quality control of the results is conducted starting with this second criterion, while the possible corrections of the results are completed using the elements of the equations chosen for the adjustment.
In our case of study, we performed three tests to estimate the quality of the results obtained (a test zone in the city of Alicante, Spain) with the software developed. First, a planimetric control was used (only to estimate accuracy for X and Y coordinates). In this case, we compared a picture obtained with classical photogrammetry PNOA (Plan Nacional de Ortofotografía Aérea) [16] with a synthetic picture generated for the same place using our methodology (we describe this procedure in planimetric analysis, Section 4.1). The PNOA picture has the best accuracy so this is used as reference. For the second test (to estimate accuracy for Z-coordinate, described in Section 4.2 as 3D analysis), we used a similar procedure to compare two digital models. The first digital model was obtained from classical photogrammetric pictures, while the second one was obtained from synthetic pictures (both in the same area, Alicante, Spain). Finally, we compared these two images with X-, Y- and Z-coordinates using a 3D transformation. In order to compare all these data and to measure the quality, we have used the software developed within the framework of a previous study [14].

4.1. Planimetric Analysis

In this case, we are interested in obtaining planimetric precisions. We compared two images (see Figure 8) with the PNOA picture being used as a reference (right side in Figure 8), while the another one was obtained using a synthetic procedure (left side in Figure 8, using a blue color ramp) for a test zone (2 × 2 km) in the city of Alicante ( X m i n = 720,000   m ;   X m a x = 722,000   m ;   Y m i n = 4,246,000   m ;   Y m a x = 4,248,000   m , UTM 30N datum WGS84). Note that X m i n , X m a x , Y m i n and Y m a x refer to test area, so the pictures shown in Figure 8 cover more territory. A good way to compare these two images is to use a 2D transformation [17] by selecting at least three of the same points in both images to obtain six parameters for affine transformation (Tx, Ty, Sx, Sy, α and β). In the example of Figure 8, the operator selected the same four points for both pictures and using the method of least squares [18], we obtained the following values for the six parameters (two translations, Tx and Ty; two scale factors Sx and Sy; one rotation α and one perpendicularity β):
T x = 0.128   m T y = 0.001   m S x = 0.999 S y = 1.000 α = 0.0185 g β = 0.0248 g .
There is a mean squared error of ± 0.506   m for the adjustment. This estimated error is below the GSD used in generating the synthetic images (1 m).

4.2. 3D Analysis

We will estimate the accuracy for the Z-coordinate in this section. In this case, we compared the calculated DEM obtained with the synthetic images with a reference one that has better precision [16] used as a reference. After that, both models were superimposed and compared in the same area of Alicante ( X m i n = 720,000   m ;   X m a x = 722,000   m ;   Y m i n = 4,246,000   m ;   Y m a x = 4,248,000   m , UTM 30N datum WGS84) to obtain the difference in a total of 843,247 points. This is seen in the histogram of Figure 9, which represents the difference in Z-coordinate between the two models (X-axis represents the difference in meters and Y-axis represents the number of times that this difference occurs). From this analysis, we obtained a mean difference of 52.779 ± 2.671   m between the two models. In a first analysis, this difference is huge, but it is necessary to remember that one digital model (PNOA used as reference) has orthometric heights and the other one (synthetic) has ellipsoidal heights. In Alicante, the geoid height above the ellipsoid is over 50 m [19].

4.3. Plani-Altimetric Analysis

The tool provided by the software allows the calculation of the spatial transformation [20] between two images with Z-coordinates. Using this transformation, we can evaluate the errors for the X-, Y- and Z-coordinates. The Helmert 3D transformation provides us with three translations (Tx, Ty and Tz), one scale factor (S) and three rotations (ω, φ and ϰ). In this case, we compared two raster files with the Z-coordinates, with one having better precision (right side in Figure 10) and the other obtained using the synthetic procedure (left side in Figure 10). A color ramp was used in both cases. The operator selects a minimum of three well-distributed points (the same for two images) to obtain these parameters:
T x = 0.000   m T y = 0.000   m T z = 50.318   m S = 1.004 ω = 0.0000 g φ = 0.0000 g = 0.0000 g .
There is a mean squared error of ± 0.666   m for the adjustment. Figure 10 shows that there is a displacement in the Z-coordinate (the geoid height above the ellipsoid for Alicante). We can see that the rotations are 0 and the scale factor is 1.004.

5. Conclusions

The main goal of this work is to provide an easy and cheap alternative technique to visualize point clouds using photogrammetric software and digital restitution stations. This will be extremely useful in cases where the use of photogrammetry may not be an option due to high costs, adverse weather and so on.

5.1. Comparison with a Classic Photogrammetric Flight

In classic photogrammetry, the hypsometric error σ h (which is what determines the overall accuracy where the Z-axis contains the direction of the optical axis) is expressed as:
σ h = σ p · h f · h b σ p = 2 2 · r p ,
where σ p is the planimetric error in the instrument; h is the flight height; f is the focus of the camera; r p is the pixel resolution (squared in our example) and b is the base area. In the equation, h f is defined as the flight scale E v .
The prototype has a GSD of 1 m, a focus of 50 mm, a pixel resolution of 50 microns, a flight height of 1 m and a base area of 165 m. With this data, E v is 1:20,000, σ p is 35 microns and σ h = ±4 m, which gives an equidistance in the curve level of 10 m. Therefore, a minimum scale for the cartography is obtained as 1:10,000, which reinforces the consistency of the model. A GSD of 1 m marks a cartography scale of 1:5000.

5.2. Uses of LiDARgrammetry

The mere act of carrying out a LiDAR flight or a Laser Scanner survey does not mean we have to take photogrammetric shots in order to carry out classic photogrammetric restitution. By applying the methodology developed in this approach, stereoscopic models that are returned using a modern photogrammetric station scan can be generated. This can be an added advantage in cases where it is impossible to obtain images (night flights, clouds, radar images, sea beds and so on).

5.3. Quality Control Tools

These techniques can also be used to see whether the cartography obtained with classic photogrammetry is sufficiently accurate. This evaluation involves first generating the synthetic photogrammetric models of an area surveyed with LiDAR and collecting the results obtained using traditional photogrammetric techniques in this area. Following this, a comparison will show the accuracy of the original cartography.

5.4. Help with Classification and Filtering

The generation of synthetic images can help with the editing of LiDAR point clouds, as it allows us to carry out a process prior to classification and filtering.

Acknowledgments

We thank the anonymous reviewers for their suggestions and comments that have improved the quality of the paper.

Author Contributions

All of the authors participated actively in the design and development of the different stages of the work presented. The paper was mainly written by R.R.-C., while J.L.G.-G. adapted the text to the journal format.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Song, J.H.; Han, S.H.; Yu, K.; Kim, H. Assessing the Possibility of Land-cover Classification Using Lidar Intensity Data. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences (IAPRS), Graz, Austria, 9–13 September 2002; Volume 34, p. 4. [Google Scholar]
  2. Rodríguez, R.; Álvarez, M.; Miranda, M.; Díez, A.; Rodríguez, P. Obtención de modelos urbanos tridimensionales. Inf. Constr. 2013, 65, 229–240. [Google Scholar] [CrossRef]
  3. GeoCue Group Support. Available online: http://support.geocue.com (accessed on 30 May 2014).
  4. Fragkos, P.; Ioannidis, C. Assessment of LiDARgrammetry for Spatial Data Extraction. In Proceedings of the SPIE 9688, Fourth International Conference on Remote Sensing and Geoinformation of the Environment (RSCy2016), Paphos, Cyprus, 4–8 April 2016. [Google Scholar] [CrossRef]
  5. Kirk, R.L.; Howington-Kraus, E. Radargrammetry on Three Planets: Mapping the Solar System’s Hidden Corners. AGU Fall Meeting Abstracts. Abstract #P33E-01. 2010. Available online: http://adsabs.harvard.edu/abs/2010AGUFM.P33E..01K (accessed on 20 January 2017).
  6. Masana, E. SHAKE: Registro de Terremotos Prehistóricos en el Sur de Iberia: Tecnologías Avanzadas en paleosismología Terrestre y Marina; CGL2011-30005-C02-01; Ministerio de Ciencia e Innovación: Barcelona, Spain, 2011–2015.
  7. Krauss, K. Photogrammetry. In Fundamentals and Standard Processes, 4th ed.; Dümmler Verlag: Bonn, Germany, 1993; Volume 1. [Google Scholar]
  8. Delaunay, B. Sur la sphère vide, Izvestia Akademii Nauk SSSR. Otdelenie Matematicheskikhi Estestvennykh Nauk 1934, 7, 793–800. [Google Scholar]
  9. Chen, L.; Xu, J. Optimal Delaunay Triangulations. J. Comput. Math. 2004, 22, 299–308. [Google Scholar]
  10. Lerma, J.L. Fotogrametría Moderna: Analógica y Digital; Universidad Politécnica de Valencia: Valencia, Spain, 2002; pp. 383–384. [Google Scholar]
  11. Dean, S.; Illowsky, B. Descriptive Statistics: Histogram. Retrieved from the Connexions. 2009. Available online: http://cnx.org/content/m16298 (accessed on 6 May 2017).
  12. Paul, R.; Bon, A. Elements of Photogrammetry; McGraw Hill: New York, NY, USA, 1974; pp. 1–2. [Google Scholar]
  13. Arfken, G. Mathematical Methods for Physicists, 3rd ed.; Academic Press: Orlando, FL, USA, 1985. [Google Scholar]
  14. Díez, A.; Arozarena, A.; Ormeño, S.; Aguirre, J.; Rodríguez, R.; Saenz, A. Integración y optimización de tecnologías y metodologías LiDAR y fotogramétricas para la producción cartográfica. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, ISPRS Congress, Beijing, China, 3–11 July 2008. [Google Scholar]
  15. Mikhail, E.M.; Bethel, J.S.; McGlone, J.C. Introduction to Modern Photogrammetry; John Wiley & Sons: New York, NY, USA, 2001. [Google Scholar]
  16. PNOA (Plan Nacional de Ortofotografía Aérea, España). 2011. Available online: http://pnoa.ign.es/ (accessed on 15 July 2017).
  17. Nomizu, K.; Sasaki, S. Affine Differential Geometry, New ed.; Cambridge University Press: Cambridge, UK, 1994; ISBN 978-0-521-44177-3. [Google Scholar]
  18. Aldrich, J. Doing Least Squares: Perspectives from Gauss and Yule. Int. Stat. Rev. 1998, 66, 61–81. [Google Scholar] [CrossRef]
  19. Pavlis, N.K.; Holmes, S.A.; Kenyon, S.C.; Factor, J.K. The development and evaluation of the Earth gravitational model 2008. J. Geophys. Res. 2012, 117, B04406. [Google Scholar] [CrossRef]
  20. Fischer, W.; Helmert, F.R. Dictionary of Scientific Biography; Scribners: New York, NY, USA, 1973; Volume 7, pp. 239–241. [Google Scholar]
Figure 1. Block diagram for the generation of synthetic images.
Figure 1. Block diagram for the generation of synthetic images.
Applsci 07 00906 g001
Figure 2. Visual aspect of a LiDAR point cloud using the intensity level to generate the image in grayscale. In the darkest areas, a greater density of points can be seen.
Figure 2. Visual aspect of a LiDAR point cloud using the intensity level to generate the image in grayscale. In the darkest areas, a greater density of points can be seen.
Applsci 07 00906 g002
Figure 3. LiDAR point cloud before and after the densification process is completed.
Figure 3. LiDAR point cloud before and after the densification process is completed.
Applsci 07 00906 g003
Figure 4. Spatial reference systems in the two frames.
Figure 4. Spatial reference systems in the two frames.
Applsci 07 00906 g004
Figure 5. Transformations to pass from system R t and the systems R i and R j .
Figure 5. Transformations to pass from system R t and the systems R i and R j .
Applsci 07 00906 g005
Figure 6. Synthetic model of intensity generated for a LiDAR point cloud.
Figure 6. Synthetic model of intensity generated for a LiDAR point cloud.
Applsci 07 00906 g006
Figure 7. Synthetic model of intensity generated and densified for a LiDAR point cloud.
Figure 7. Synthetic model of intensity generated and densified for a LiDAR point cloud.
Applsci 07 00906 g007
Figure 8. Screenshot of the planimetric control with the mean squared error obtained. The software allows you to move the images independently. See the coordinates of two images at the button in the first box. The cross is only used to target a point.
Figure 8. Screenshot of the planimetric control with the mean squared error obtained. The software allows you to move the images independently. See the coordinates of two images at the button in the first box. The cross is only used to target a point.
Applsci 07 00906 g008
Figure 9. Details of the hypsometric analysis carried out.
Figure 9. Details of the hypsometric analysis carried out.
Applsci 07 00906 g009
Figure 10. Screenshot of the hypsometric analysis with the mean squared error. The software allows you to move the images independently. See the coordinates of two images on the button in the first box. The cross is only used to target a point.
Figure 10. Screenshot of the hypsometric analysis with the mean squared error. The software allows you to move the images independently. See the coordinates of two images on the button in the first box. The cross is only used to target a point.
Applsci 07 00906 g010

Share and Cite

MDPI and ACS Style

Rodríguez-Cielos, R.; Galán-García, J.L.; Padilla-Domínguez, Y.; Rodríguez-Cielos, P.; Bello-Patricio, A.B.; López-Medina, J.A. LiDARgrammetry: A New Method for Generating Synthetic Stereoscopic Products from Digital Elevation Models. Appl. Sci. 2017, 7, 906. https://doi.org/10.3390/app7090906

AMA Style

Rodríguez-Cielos R, Galán-García JL, Padilla-Domínguez Y, Rodríguez-Cielos P, Bello-Patricio AB, López-Medina JA. LiDARgrammetry: A New Method for Generating Synthetic Stereoscopic Products from Digital Elevation Models. Applied Sciences. 2017; 7(9):906. https://doi.org/10.3390/app7090906

Chicago/Turabian Style

Rodríguez-Cielos, Ricardo, José Luis Galán-García, Yolanda Padilla-Domínguez, Pedro Rodríguez-Cielos, Ana Belén Bello-Patricio, and José Antonio López-Medina. 2017. "LiDARgrammetry: A New Method for Generating Synthetic Stereoscopic Products from Digital Elevation Models" Applied Sciences 7, no. 9: 906. https://doi.org/10.3390/app7090906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop