Next Article in Journal
An Integrated LSTM-Rule-Based Fusion Method for the Localization of Intelligent Vehicles in a Complex Environment
Previous Article in Journal
An Enhanced FGI-GSRx Software-Defined Receiver for the Execution of Long Datasets
Previous Article in Special Issue
Intelligent Path Planning with an Improved Sparrow Search Algorithm for Workshop UAV Inspection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel 3D Reconstruction Sensor Using a Diving Lamp and a Camera for Underwater Cave Exploration

Laboratory of Informatics, Robotics and MicroElectronics (LIRMM) (UMR 5506 CNRS—UM), University of Montpellier, 161 Rue Ada, CEDEX 5, 34392 Montpellier, France
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(12), 4024; https://doi.org/10.3390/s24124024
Submission received: 11 April 2024 / Revised: 23 May 2024 / Accepted: 29 May 2024 / Published: 20 June 2024

Abstract

:
Aquifer karstic structures, due to their complex nature, present significant challenges in accurately mapping their intricate features. Traditional methods often rely on invasive techniques or sophisticated equipment, limiting accessibility and feasibility. In this paper, a new approach is proposed for a non-invasive, low-cost 3D reconstruction using a camera that observes the light projection of a simple diving lamp. The method capitalizes on the principles of structured light, leveraging the projection of light contours onto the karstic surfaces. By capturing the resultant light patterns with a camera, three-dimensional representations of the structures are reconstructed. The simplicity and portability of the equipment required make this method highly versatile, enabling deployment in diverse underwater environments. This approach is validated through extensive field experiments conducted in various aquifer karstic settings. The results demonstrate the efficacy of this method in accurately delineating intricate karstic features with remarkable detail and resolution. Furthermore, the non-destructive nature of this technique minimizes disturbance to delicate aquatic ecosystems while providing valuable insights into the subterranean landscape. This innovative methodology not only offers a cost-effective and non-invasive means of mapping aquifer karstic structures but also opens avenues for comprehensive environmental monitoring and resource management. Its potential applications span hydrogeological studies, environmental conservation efforts, and sustainable water resource management practices in karstic terrains worldwide.

1. Introduction

Access to water has become one of the most pressing challenges facing humanity today. However, water is abundant on a global scale; the oceans are an inexhaustible reserve, provided that salt and water are separated. Unfortunately, desalination methods are still too energy-intensive to be used profitably on a large scale. Pending a possible scientific breakthrough, people are therefore relying essentially on freshwater resources, which account for 2.5% of the world’s water [1] (Figure 1). Only 1.2% of this freshwater is accessible at the surface, although it is now very polluted. Of the remainder, 68.7% is trapped in glaciers and ice caps and 30.1% is found underground. This groundwater, which represents 0.76% of the world’s resources, is extremely valuable. Groundwater is the result of a long and natural process of filtration through the various layers of soil, which means that its quality is generally considered to be very good [2,3]. This groundwater is stored in aquifers, most of which are karstic.
Aquifer karstic structures, due to their complex and porous nature, present significant challenges in accurately mapping their intricate features. Traditional methods often rely on invasive techniques or sophisticated equipment, limiting accessibility and feasibility. In this paper, a novel non-invasive approach is proposed utilizing structured light projection in conjunction with a camera and a diving light to effectively map aquifer karstic formations. Section 2 presents some related works in underwater mapping. Section 3 details the materials and methods used in this original approach based on the principles of structured light, leveraging the precise projection of light contours onto the karstic surfaces. It also outlines our process of capturing the resultant light patterns with a camera (reflex Nikon D700 with underwater housing), leading to reconstruct detailed three-dimensional representations of the subsurface structures. Section 4 presents the results to validate this approach via the simulation of an aquifer galley first, and finally through extensive outdoor experiments conducted in a particular equivalent aquifer karstic environment. The results demonstrate the efficacy of this method in accurately delineating intricate karstic features with remarkable detail and resolution. Furthermore, the non-destructive nature of this technique minimizes disturbance to delicate aquatic ecosystems while providing valuable insights into the subterranean landscape. This innovative methodology not only offers a cost-effective and non-invasive means of mapping aquifer karstic structures but also opens avenues for comprehensive environmental monitoring and resource management. Its potential applications span hydrogeological studies, environmental conservation efforts, and sustainable water resource management in karstic terrains worldwide.

2. Related Work

Underwater mapping is a valuable tool in a wide range of applications, including marine biology, geology, archaeology, and the offshore industry. Various methods and sensors are employed to obtain 3D reconstructions of the environment, with data typically collected by divers, ROVs ( Remotely Operated Vehicle), or AUVs ( Autonomous Underwater Vehicle). In most cases, this involves performing surveys with downward-facing sensors and reconstructing the seafloor surface [4,5]. The studies conducted by the authors of [6,7] provide comprehensive reviews of the various underwater 3D reconstruction methods. This synthesis of methods is illustrated in Figure 2.
Sensors used for 3D reconstruction can be classified into two main categories:
  • Active sensors: these sensors physically interact with the environment by emitting a signal or wave, such as sound waves in the case of sonars or light waves in the case of LiDAR systems.
  • Passive sensors: these sensors do not modify the environment; the most commonly used ones for 3D reconstruction are camera-based.
Time-of-flight methods are the most direct methods for 3D reconstruction. They work on the same principle as a rangefinder, namely emitting a signal or wave and retrieving its echo. By knowing the speed of the signal thanks to the nature of the wave and the propagation medium, it is possible to proportionally deduce the relative distance between the sensor and the observed element from the measurement of the time elapsed between the emission of the wave and the return of the echo.
Among time-of-flight methods, those based on acoustic waves are by far the most used because their propagation is favored by the medium. They offer an interesting coverage capacity in front of the size of the areas to be studied in underwater mapping.
In addition, and to allow the localization of 3D data during the measurement, most underwater navigation algorithms are based on acoustic positioning systems such as Doppler Velocity Log (DVL) and Ultra-short baseline (USBL) [8,9]. On the other hand, Multibeam sonars (MBS) are adapted to bathymetric mapping (measurement of the seabed relief) [10]; however, they are much less suitable for structured (non-flat) environments such as karst aquifers.
Gary et al. [11] and Kantor et al. [12] presented a 3D model of the Zacatón cenote (a water-filled cave at least 300 m deep) using sonar data collected by the DEPTHX (Deep Phreatic THermal eXplorer) vehicle. It is a NASA autonomous underwater vehicle measuring just over 2 m and weighing over a ton, prefiguring among the robots that could explore underwater environments on other planets. Its sensors consist of a DVL, a sonar, and an IMU (Inertial Measurement Unit); additionally, its movements are ensured by six motors allowing it to move in all directions.
Another method using acoustics is described by the authors of [13], which shows the results of an AUV performing limited penetration inside an underwater cave using a mechanically scanned imaging sonar (MSIS, namely mechanically scanned imaging sonar). The results, obtained only for localization, show that it is possible to use this type of technology in karst environments.
Though they are efficient for obtaining bathymetry and macro-characteristics of habitats, these sensors nevertheless provide little information on fine characteristics (micro-bathymetry), do not allow for the recovery of the texture of the environment, and are very expensive.
Still in the class of time-of-flight methods, those based on electromagnetic waves, such as LiDAR, offer access to a very dense reconstruction of the seabed. However, on the other hand, they are reserved for very local observations, because the propagation of these waves are limited by the medium as explained in the constraints section. They are therefore not adapted to the situation envisaged.
The other class of methods are triangulation methods which are favored in this presentation. These methods require at least two devices (or two different views) which will provide distance information from the same point, and thus form a triangle between the two devices and the point to be measured.
There are several ways to perform triangulation:
  • Multi-view triangulation involves taking multiple views of the same scene. By knowing the disparities of each viewpoint, it is possible to project the points of the scene onto different “image” planes, and thus triangulate the points of the scene to obtain a depth map. This technique is based on epipolar geometry (explained in [14]) with two possible approaches:
    The use of a binocular stereo system where the relative positions of the two cameras are known. It is possible to make the system active by using a structured light projector to add discriminant points or features related to the observed environment, and thus facilitate the matching of points between the two images for triangulation. In [15], an original stereovision system is presented carried by divers for pipeline measurements; however, its performance needs to be evaluated.
    The method known as Structure from Motion (SFM) involves taking a sequence of sequential images of an object or scene, typically from a single camera. To obtain a metric reconstruction, it is necessary to know the camera trajectory to resolve the scale ambiguity.
  • Structured light triangulation relies on an emitter-receiver system. The emitter is a light source (laser or not) that projects a pattern (distinctive patterns) onto a scene observed by the receiver—a camera. Using the positions of the camera and the projector, it is possible to triangulate the discriminant points brought by the light in the scene captured by the camera [16]. In [17], results obtained on small objects (spheres) with a measurement distance of about 1 m lead to radii estimation of less than 0.5 mm. In [18], surface measurements based on the fringe projection technique are presented. They describe an extended camera model which takes refraction effects into account. The results have a resolution of less than 150 μ m; the study was performed on small objects and needs to be conducted on bigger scenes.
Regarding structured light triangulation methods, they are very accurate, but are generally only used to reconstruct small areas or objects. This is because light is quickly absorbed in water; therefore, the sensor must be close to the target to best detect the projected light. Additionally, structured light sensors are not yet well-suited for mobility and are therefore mostly used statically for the moment.
Digital reflex cameras, on the other hand, have the advantage of being inexpensive, providing density as well as textural data. However, most camera-based methods operate in open areas with natural lighting (or with artificial lighting that completely illuminates the workspace) and generally try to reconstruct the ground or small objects.
For example, refs. [19,20] propose a 3D reconstruction method based on a stereo pair for archaeological objects in shallow water and in very poor conditions. Special filters are implemented to solve the noise problems caused by these poor conditions. This static 3D measurement is fused with a 3D map of the excavation area obtained using a scanning sonar. Brandou et al. [21] performed a dense reconstruction of submerged structures at very high depths (up to 6000 m) with a stereo system mounted on a controlled manipulator arm. The imposition of a known trajectory allows one to know the different positions of the cameras to calculate a precise 3D model of the scene.
Beall et al. [22] reconstructed coral reefs with a wide-baseline stereo system using high-resolution video and a dense reconstruction method.
The context of this paper is different from these works since our experiments are in total darkness in a structured and extended environment whose content appears as the diver or robot moves.
Zhou et al. [23] presented a description of the problems related to metrology in the underwater environment. The importance of calibrating the stereo pair underwater is emphasized, taking into account the refraction problems related to the medium. The detection of interest points is performed on sufficiently bright images allowing a robust matching then used for a 3D reconstruction. The comparison of the object measurements obtained is made with a reference metrology and gives good results; however, only in static cases.
The work of [24] is directly related to the present theme. The objective is to reconstruct a karstic underwater tunnel of Woodville Karst Plain in Florida using a new approach to stereovision generating a 3D point cloud of the cave. In their method, after calibration and rectification of the highly distorted images, they extract the contours of the artificial light projected into the cave which creates a cone of light illuminating part of the walls. These contours are then used as inputs in their stereovision algorithm. The displacement is estimated by visual odometry (ORB-SLAM ( Oriented FAST and Rotated BRIEF-Simultaneous Localization and Mapping) [25]) and allows a 3D reconstruction of a portion of the cave of about one hundred meters; however, it does so without specifying the comparison of the result with a known ground truth.
For an even more optimal mapping of this underwater gallery, ref. [26] developed a SLAM method where they fuse this stereo method with a sonar and an inertial measurement unit, thus improving the accuracy of their measurements. Reviews on SLAM methods in underwater environment can be found in refs. [27,28].
As stated by Joshi et al. [29] and Ende [30], the underwater environment is very specific and induces new challenges both in exploring and mapping tasks. The interested reader can refer to the work of Han, Hao and Qi [31] or Papp et al. [32] for an example on how light refraction can impact photogrametric approaches, or to Hou et al. [33] to see how the camera calibration is altered.
Of course, 3D reconstructions have not been preserved from the current trend of deep learning techniques. Ju et al. outlined a good state of such methods for calibrated approaches in ref. [34], but the work of Abdullah et al. [35] is more specific to the semantics of underwater scenes.
Wu et al. [36] used a SFM method coupled with multi-view stereo photography to perform measurements of silting areas at the outlet of water transport pipelines to the seabed. The accuracy of their method allows a robust estimation of these elements useful for monitoring these areas, but it also only uses static measurements.
Fan et al. [37] provided an overview of calibration and underwater image processing methods in structured light. Wang et al. [38] proposed a method of 3D reconstruction in underwater environments combining a SLAM approach and dense reconstruction using a stereo system embedded on the Aqua2 robot whose IMU allows for localization. The results are mainly on coral reefs and partially on the reconstruction of underwater cavities. The question of the accuracy of the measurements with respect to a ground truth remains. Braeuer-Burchardt et al. [39] proposed an application of 3D reconstruction for underwater archeology. Their system consists of two monochrome cameras and a sinusoidal light pattern projector. A combination of 3D reconstruction in structured light and SLAM-based visual odometry is proposed. Their design is mounted on an ROV whose IMU data complements that of visual odometry. The first results were obtained on objects in a controlled environment. A test was then carried out over a dozen meters to follow a pseudo pipeline, but the size of the system was too large for usable application in underground galleries.
Yu et al. [40] used Ariane’s thread to enable the movement and guidance of a BlueRov robot. This learning method gives good results for localization and assisted piloting. The next steps envisaged are autonomous navigation and then mapping of the cavities studied (but were already explored by divers).
Hu et al. [41] proposed a real-time monocular 3D reconstruction method for underwater environments. It relies on optical flow tracking of characteristic seabed points and uses Delaunay triangulation to complete the seabed estimation. The method assumes a nearly planar seabed but remains unclear about the achieved metric accuracy.

3. Materials and Methods

The complete development of the proposed method is detailed in Figure 3 and is based on three main blocks:
  • Zhang’s camera calibration method [42] is used to estimate its intrinsic parameters, including radial distortion;
  • light projector calibration consists of estimating the geometric parameters of the cone (vertex O P , direction d , half-angle of opening α );
  • three-dimensional local reconstruction is based on the camera/projector triangulation leading to a 3D point cloud expressed in C , the camera reference frame.
This triangulation is illustrated in Figure 4. It shows the system, with one camera and light projector. To simplify the figure, the light is projected onto a plane, but one can imagine that the procedure has to be applied to non-planar surfaces such as the walls of a gallery. Once the light contour has been detected in the image, the associated camera ray is calculated for each point on the contour. Figure 4 shows two points, X and Y , belonging to the contour detected in the image and the associated rays. In the configuration shown in Figure 4, the ray associated with X and Y , respectively, intersects the cone at two points X 1 and X 2 , and Y 1 and Y 2 , respectively. To calculate these intersections, one needs to know the angle of aperture α and the transformation between the references of the cone and the camera called T { P C } .
Also in Figure 4, the intersections which correspond to the 3D points of the contour ( X 1 and Y 2 ) are shown in green and those which should not be taken into account ( Y 1 ) are shown in red. The point X 2 is not shown, but one can imagine that it would be red and beyond the area of the plane in Figure 4. For a given ray, it is therefore necessary to identify whether it is the first intersection or the second intersection that belongs to the observed light contour.
This method of selecting intersections requires a study of the geometric relationships between the cone and the camera.

3.1. Cone Parameters

3.1.1. Cone of Revolution

The cone of revolution, referred to as a cone in the remainder of this document, is associated with its surface of revolution and not with its solid, as may be the case in certain applications. The terms used to define the relationship between a point and the cone are as follows:
  • a point belonging to the cone is a point which belongs to the surface of the cone;
  • a point inside the cone is a point which belongs to the solid bounded by the surface of the cone;
  • a point outside the cone is a point which does not belong to the solid bounded by the surface of the cone.
This surface is generated by the revolution of a line, called the generatrix, noted g and which passes through a vertex, in this case O P .
The revolution takes place around a fixed axis which also passes through the vertex and which turns out to be an axis of symmetry of the cone, the direction of which is defined by the unit vector d .
If a plane not passing through O P is orthogonal to the axis of symmetry, then its intersection with the cone is a circle.
Thus, the angle formed by g and the axis of symmetry is constant and corresponds to the half-angle of the cone opening α ] 0 ; π 2 [ .
The cone is represented in Figure 5 with the orthonormal base ( O P , x P , y P , z P ) , oriented such that z P = d , which defines the P reference frame associated with the cone.
The generatrix g has the unit direction vector w which is set by a new angle θ [ 0 ; 2 π [ (Figure 5).
In the cone frame P , one can simply express the vector w { P } (Equation (1)) which then depends only on the angles θ and α :
w { P } = sin ( α ) cos ( θ ) sin ( α ) sin ( θ ) cos ( α )
This vector can be used to express all the points on the surface parametrically.
Let X { P } = x X { P } y X { P } z X { P } T be a point X expressed in the reference frame P . Then, if X belongs to the cone, one can calculate (Equation (2)):
X { P } = l w { P } x X { P } = l sin ( α ) cos ( θ ) y X { P } = l sin ( α ) sin ( θ ) z X { P } = l cos ( α )
where l is a real parameter whose absolute value defines the distance between X { P } and the cone vertex. However, with l R , the parametric equation defines a “double cone”. However, since the cone models the light projector, the study is only interested in the upper part of the “double cone”, i.e., when the parameter l R + .

3.1.2. Orthogonal Distance

This parameterization of the cone will be useful to write the equation for the orthogonal distance between a point and the cone, an equation that will be useful for cone estimation. Let M be a point outside the cone and H be its orthogonal projection on the cone (Figure 5). The angle θ H is the angle that gives the generatrix of the cone closest to M . This generatrix, called g H , passes through H and has the unit direction vector w H whose expression in the P reference frame is:
w H { P } = sin ( α ) cos ( θ H ) sin ( α ) sin ( θ H ) cos ( α )
If M is also expressed in P with M { P } = x M { P } y M { P } z M { P } T , the expressions become (Equation (3)):
θ H = arctan y M { P } x M { P } H { P } = ( M { P } T · w H { P } ) w H { P }
If M { P } T · w H { P } < 0 then the projected point is in the lower part of the cone. Working with the upper part of the cone, when this condition is true, the nearest point is the vertex of the O P cone (example in Figure 6). The orthogonal distance h is therefore (Equation (4)):
h = ( M { P } H { P } ) T ( M { P } H { P } ) si M { P } T w H { P } 0 M { P } T M { P } si M { P } T w H { P } < 0

3.1.3. Quadratic Form of the Cone

To obtain this relationship, one must first note that if the angle between the vectors ( X O P ) and d or ( X O P ) and d is equal to the angle α , then the point X belongs to the cone. This is equivalent to the following scalar product:
( X O P ) T d = ± X O P cos ( α )
When squared, the expression becomes:
( X O P ) T d 2 = X O P cos ( α ) 2 ( ( X O P ) T d ) ( ( X O P ) T d ) = ( X O P ) T ( X O P ) cos 2 ( α ) ( X O P ) T d d T ( X O P ) ( X O P ) T ( X O P ) cos 2 ( α ) = 0 ( X O P ) T d d T cos 2 ( α ) I 3 Q ( X O P ) = 0
Thus, one can obtain a quadratic function of the cone in the general case (Equation 5), directly showing the parameters d , α and O P which define it:
Q ( X ) = ( X O P ) T Q ( X O P )
This gives the relationships between the point X and the cone (Equations (6a)–(6f)):
(6a)    If Q ( X ) = 0 then X belongs to the cone (6b)    If Q ( X ) < 0 then X is inside the cone (6c) If Q ( X ) > 0 then X is outside the cone (6d) If Q ( X ) = 0 and ( X O P ) T d > 0 then X belongs to the upper part of the cone (6e) If Q ( X ) < 0 and ( X O P ) T d > 0 then X is inside the upper part of the cone (6f) If Q ( X ) > 0 or ( X O P ) T d < 0 then X is outside the upper part of the cone .

3.2. Projector Calibration

Projector calibration is an essential step, since it involves estimating the parameters of the cone used in this method and in the Equations (1) to (6f). As a reminder, the parameters of the cone to be estimated are its vertex O P , its direction vector d and its half-angle of aperture α .
Since O P and d represent the relative pose of the cone with respect to the camera, the projector must be fixed with respect to the camera during calibration and during any experimentation. Furthermore, if the projector has a variable half-angle aperture, it is also necessary to ensure that this is fixed.

3.2.1. Three-Dimensional Point Generation for Cone Estimation

The first step in the calibration method is to generate a cloud of 3D points belonging to the cone. To achieve it, the chosen method is to capture several light projection images on a flat surface.
This light projection is the result of the intersection between the projector cone and the flat surface, which is by definition a conic. In this case, it will be an ellipse because one wants a closed shape. The projection of this ellipse into the image is also an ellipse.
If one can estimate the relative pose of this surface with respect to the camera and extract the light contour in the image, which is an ellipse, then one can obtain the real ellipse from the ellipse in the image (Figure 7).
To estimate the relative pose, a chessboard pattern is attached to the surface to estimate its relative pose. This relative pose is obtained by estimating the homography between the calibration pattern and the image plane [43]. As the camera is calibrated, rotation and translation can be estimated from the homography. To extract the ellipse from the image, a contour detection method is used, presented in [44,45]. Then, from the contour points obtained, the best ellipse is estimated using the method of [46], which is based on least squares optimisation.
It is therefore possible to obtain n elliptical sections of the cone from n images taken at different distances from the surface. For each ellipse extracted in an image, the ellipse in the surface is obtained via the estimated homography between the image plane and this surface. The estimated relative pose allows one to obtain the elliptical section, i.e., the set of 3D points expressed in the camera frame of reference belonging to the ellipse of the surface.
Figure 8 shows an example where the camera has captured three images of an ellipse surrounding a chessboard pattern on the flat surface in three different poses. This is equivalent to obtaining three elliptical sections of the same cone.
Consequently, if n elliptical sections are obtained where m 3D points are extracted per section, one can generate a cloud of n × m 3D points belonging to the cone.

3.2.2. Cone Estimation

The aim is now to estimate the cone that best approximates the 3D point cloud. This is performed by geometric fitting, i.e., by minimizing the orthogonal distances between the 3D points and the cone.
Let p be the vector containing the six parameters of the cone, which are:
  • the three coordinates of its vertex O P ;
  • the two angles yaw and pitch in ZYX-Euler convention of its direction vector d , the angle roll not being necessary since a cone has an axis of symmetry;
  • its opening half-angle α .
To estimate the p vector, the Levenberg–Marquardt (LM) iterative method [47] is used. For the first iteration, one needs to choose an initial parameter vector p 0 . This could be the null vector, for example. In practice, measurements will be used, taken with a caliper for the various parameters of p 0 .
At each iteration i, LM receives a solution vector p i and the function for calculating the vector of orthogonal distances to be minimized. This function first calculates the rotation and translation of the cone relative to the camera from O P and the yaw and pitch angles of the vector p i . It then expresses the 3D points in the cone reference frame and calculates the orthogonal distance vector between the 3D points and the cone using the Equation (4).
So from p i and this function, LM returns a new solution vector p i + 1 .
When the difference between the sum of the squared orthogonal distances obtained with p j and the calculus obtained with p j + 1 is less than an arbitrary threshold, this means that LM has reached convergence at iteration j. The best solution in the least squares sense is therefore the vector p j .
It is important to note that there are an infinite number of cones that share the same elliptical section. On the other hand, there is only one cone that passes through an elliptical section and through a point outside that section. Consequently, this minimization can only work if one has at least two elliptical sections.

3.3. Intersection with a Ray

Now that the method to determine the cone parameters using the calibration method is known, the next step is to be able to calculate the intersections between a camera ray and the cone; we are only interested in the intersections with the top of the cone.
A ray is a half-line that can be expressed parametrically using the following Equation (7):
X ( t ) = O + t u
where O is the starting point of the ray, u is its unit direction vector and t is a positive real number such that X ( 0 ) = O . Note that it only deals with cases where the point O is outside the cone. This is because the starting point of the camera rays is the optical center in this situation. The point O can nevertheless be inside the lower part of the cone.
If X ( t ) is an intersection point, then it belongs to the cone. Therefore, to determine the intersections, one can simply calculate the values of t which cancel the general polynomial of the cone (Equation (5)) after replacing X by the expression for X ( t ) :
Q ( X ( t ) ) = ( X ( t ) O P ) T Q ( X ( t ) O P ) = 0 ( O + t u O P ) T Q ( O + t u O P ) = 0 t 2 ( u T Q u ) + t ( 2 u T Q ( O O P ) ) + ( O O P ) T Q ( O O P ) = 0 a t 2 + b t + c = 0 with a =    u T Q u b =    2 u T Q ( O O P ) c =    ( O O P ) T Q ( O O P ) = Q ( O )

3.4. Projection of the Generatrices of the Cone in the Image

To understand the relationship between a point on the cone and its projection into the image, one needs to look at the projections of the generatrices of the cone. Since the cone represents the projector, only the upper half of the cone is considered, which means that the generatrices can be limited to the half-lines that start at the vertex O P . Note that the study is limited to the case where the cone is oriented in the same direction as the camera, as required by this method. Geometrically, this condition is satisfied if the angle called μ between z C and d = z P is in the interval ] π 2 α ; π 2 + α [ (see Figure 9). In practice, for the camera to be able to see the light projection, the angle μ will be close to 0.
  • Calculating the projections of the generatrices of the cone
Let us call Π g the plane which passes through the optical center O C and through at least one generatrix of the cone. The intersection of this plane with the image plane is the line g .
Let X be a point in the image such that X = K C { u } , where u is the direction vector of the camera ray associated with X . If X is the projection into the image of a 3D point belonging to a generatrix resulting from the intersection between the cone and the plane Π g , then X belongs to the line g . However, if X belongs to the line g , it is not necessarily the projection of a point on the cone. To obtain the equation of g , the simplest way is to select two points belonging to one of the generatrices (or the generatrix) through which the Π g passes, then project them into the image to find the parameters of the line. The same result would be obtained by projecting two points of the plane Π g which do not belong to the same camera radius. However, knowing the line g is not enough to understand the projection of a generatrix.
  • Observation when Π g is not tangent to the cone
When the plane Π g is not tangent to the cone, it passes through two generatrices g 1 and g 2 , with g 1 chosen as the generatrix closest to the optical center O C (see Figure 10).
The direction vectors of these two generatrices are w 1 and w 2 such that:
w 1 { P } = sin ( α ) cos ( θ 1 ) sin ( α ) sin ( θ 1 ) cos ( α ) et w 2 { P } = sin ( α ) cos ( θ 2 ) sin ( α ) sin ( θ 2 ) cos ( α )
where θ 1 and θ 2 are the angle between the vector w 1 and w 2 , respectively, projected in the plane x P y P and the vector x P (see Figure 10). As the plane Π g is chosen here as not tangent to the cone, one has θ 1 θ 2 . In this configuration, if X is a point on the line g then it can be both the projection of a point X 1 belonging to g 1 and of a point X 2 belonging to g 2 . The generatrix g 1 therefore contains the first intersections of the camera ray with the cone, while g 2 contains the second intersections.
  • Calculating the two special generatrices when Π g is tangent to the cone
There are only two orientations for which the plane Π g is tangent to the cone. When this happens, then θ 1 = θ 2 and the generatrices g 1 and g 2 are superimposed. The two possible cases are illustrated in Figure 11. The two angles θ A and θ B define the two generatrices g A and g B through which pass the two planes tangent to the cone Π g A and Π g B . It is important to note that the angles θ A and θ B depend solely on the parameters of the cone and its pose relative to the camera T { E C } , which means that they are unique and fixed for a given configuration.
To determine these two angles, a plane tangent to the cone is used, called Π T , which passes through a generatrix g whose direction vector is w defined by an angle θ . It is represented in Figure 10 passing through the generatrix g 2 purely for reasons of visibility. The orientation of the plane Π T is characterized by the unit normal vector n T such that:
n T { P } = cos ( α ) cos ( θ ) cos ( α ) sin ( θ ) sin ( α )
It is now possible to look for the two solutions of θ so that the plane Π T is superimposed on the plane Π g A or Π g B . This superposition only occurs when the plane Π T passes through the optical center O C and therefore when the vector n T is orthogonal to the vector ( O C O P ). This leads us to look for θ when (Equation (9)):
O C { P } T n T { P } = 0 x O C { P } y O C { P } z O C { P } cos ( α ) cos ( θ ) cos ( α ) sin ( θ ) sin ( α ) = 0 x O C { P } cos ( α ) cos ( θ ) + y O C { P } cos ( α ) sin ( θ ) z O C { P } sin ( α ) = 0
where O C { P } is the optical center expressed in the reference frame of the P projector. To solve the Equation (9), one can pose (Equation (10)):
s = cos ( α ) x O C { P } y O C { P } and t = cos ( θ ) sin ( θ )
One can rewrite Equation (9) using the scalar product between the vectors s and t and the angles defined in Figure 12:
s T t = cos ( α ) x O C { P } 2 + y O C { P } 2 s t = s cos ( γ ) = z O C { P } sin ( α )
The angle γ between the vectors s and t is therefore:
γ = ± arccos ( z O C { P } sin ( α ) cos ( α ) x O C { P } 2 + y O C { P } 2 ) = ± arccos ( z O C { P } x O C { P } 2 + y O C { P } 2 tan ( α ) )
As for the angle β which defines the orientation of the vector s , its expression is:
β = atan 2 ( y O C { P } , x O C { P } )
The solution to Equation (9) is therefore (Equation (11)):
θ A , B = β ± γ = atan 2 ( y O C { P } , x O C { P } ) ± arccos ( z O C { P } x O C { P } 2 + y O C { P } 2 tan ( α ) )
Note that the two angles θ A and θ B only depend on the half-angle of the opening α of the cone and the translation between the cone and the camera (translation expressed in the equation by O C { P } ). From the expression of these two angles, one can obtain the expression of the two unit direction vectors w A and w B of the two special generatrices g A and g B . The projections of these two generatrices belong to the lines g A and g B which are the intersections of the image plane with the planes Π g A and Π g B .
  • Splitting the cone into two areas using the two special generatrices
Using the generatrices g A and g B , one can divide the cone into two surfaces as shown in Figure 13. The cyan surface named S 1 (and the magenta surface named S 2 , respectively) contains the set of generatrices containing all the first (and second, respectively) possible intersections between a camera beam and the cone. These generatrices were previously named g 1 (and g 2 , respectively).
In the image plane, this shows that one can delimit the areas where the camera rays can intersect the cone. The cyan area, surrounded by the ellipse E , contains the projections of part of the set of 3D points belonging to the surface S 1 . The magenta area contains the projections of the other part of the set of 3D points belonging to the surface S 1 and contains the projections of the set of 3D points belonging to the surface S 2 . These areas are bounded by:
  • the lines g A and g B ;
  • the ellipse E which is the set of vanishing points of the generatrices. These points define a first bound on all the projections of the generatrices;
  • the projection of the vertex O P into the image plane, named O P . It defines a second bound on all the projections of the generatrices.
In the configuration illustrated in Figure 13, each generatrix projection is therefore a segment whose limits are the vanishing point of the generatrix belonging to the ellipse E and the point O P .
However, there are other configurations where the projections of the generatrices onto the image plane are not segments but lines. Indeed, if z O P { C } = 0 , then the line passing through O P and O C is parallel to the plane, which implies that the projection of the vertex O P into the image plane is a point at infinity (or ideal point) and that the lines g A and g B are parallel.
The case where z O P { C } < 0 is illustrated in Figure 14 which shows the lower part of the cone behind the camera. This part is shown in brown and its projection in the image plane is also in brown. This projection is obviously fictitious, as this part cannot be seen by the camera even if it had an infinite field of view. The point O P , also fictitious, is still the intersection of the lines g A and g B , but this time it is on the other side in comparison to that shown in Figure 13.
As for the magenta and cyan areas in the image plane, they are still delimited by the lines g A and g B and the ellipse E but extend towards infinity on the side where the lines g A and g B diverge.
Therefore, if z O P { C } 0 , then each generatrix projection is a half-line whose only boundary is its vanishing point belonging to the ellipse E . However, working with real cameras and not with the image plane, which is infinite, each generatrix will have a segment as its projection into the image.
This division of the cone into two surfaces leads to several results:
  • Any ray associated with a point in the cyan area has its unique intersection with the cone a 3D point belonging to the surface S 1 .
  • Any ray associated with a point in the magenta area has two intersections with the cone, the first of which belongs to the surface S 1 and the second to the surface S 2 . These two intersections are superimposed if the point in the magenta area belongs to the line g A or to the line g B .
  • Any ray associated with a point outside the magenta area and the cyan area does not have any intersection with the cone.

3.5. Intersection Selection and Triangulation

Now, everything is ready to develop a method for determining which intersection to choose when the ray associated with a pixel in the contour has two intersections with the cone. To better understand the various stages of the procedure, let us take as an example the 3D contour of light whose associated 3D curve is circular and results from the intersection between a plane and the cone.
The example of the circular contour is illustrated in Figure 13 which contains the following elements:
  • C is the closed 2D curve corresponding to the contour in the image with X a 2D point on the curve.
  • C is our circular 3D curve corresponding to the contour of the light in the scene, where X is a 3D point on the curve which has X as its projection. The point X therefore corresponds to the correct intersection between the cone and the camera ray associated with X .
  • The areas containing the first and second intersections with the cone are delimited by the two generatrices g A and g B which divide the cone into two surfaces S 1 and S 2 . They also divide C into two curves, one in cyan called C 1 and the other in magenta called C 2 .
  • A and B are the two intersections of g A and g B , respectively, with the curve C and are thus the only two 3D points common to the curves C 1 and C 2 .
  • The cyan curve named C 1 and the magenta curve named C 2 are the projections of the curves C 1 and C 2 , respectively.
  • A and B are the projections of A and B and are therefore the only two points common to the curves C 1 and C 2 . They belong to both the curve C and the lines g A and g B . So, in theory, they correspond to the two unique points of tangency of the lines g A and g B with C .
Using the points of tangency, A and B , will allow one to divide the 2D curve C in two via the line ( A B ) . However, in practice:
  • the curve C is discrete,
  • the light contours are extracted in a perfectible way,
  • the camera calibration and the cone estimation have uncertainties, so the lines g A and g B also have uncertainties.
All this means that the lines g A and g B are not exactly tangent to the curve C . The method for determining the points A and B must therefore be adapted to this constraint.
Once these two points have been obtained and the curve C has been divided in two, it remains to determine which of the two curves is C 1 or C 2 . To do this, one needs to look at the relative position of the vertex O P in relation to the camera. Figure 15 shows that the vertex O P and the curve C 2 lie in the same half-space bounded by the plane passing through O C and the line ( A B ) . This property is always true whatever the pose of the cone.
In the image plane, this property can be expressed by defining a point p ref such that p ref and the curve C 2 lie in the same half-plane bounded by the line ( A B ) .
There are an infinite number of solutions for p ref in the image plane which satisfy this condition. Therefore p ref is forced to belong to the circumscribed circle of the rectangle delimiting the image, as can be seen in Figure 15. The segment between p ref and the center of the image forms an angle with the horizontal axis of the image called ν . This gives (Equation (12)):
p ref = w 2 + h 2 cos ( ν ) sin ( ν )
where w and h are the width and height of the image in pixels, respectively. For the condition to be true, one possible solution for ν is to indicate the direction of O P relative to the optical center O C . To obtain this angle, O P is orthogonally projected onto the plane normal to the camera axis passing through O C , which gives the point H P illustrated in Figure 15. This gives ν which is the angle formed by the vectors ( H P O C ) and x C (Equation (13)):
( H P O C ) = x O P { C } y O P { C } 0 donc ν = arctan y O P { C } x O P { C }
Calculating this angle ν and the point p ref allows one to deduce which of the two curves is C 2 .
Once the curves C 1 and C 2 have been obtained, it is then possible to obtain the 3D contour. Indeed, if a point X belongs to C 1 , then the first intersection of the camera ray associated with it must be chosen, and if it belongs to C 2 , the second must be chosen. This is because any point X of C 1 is the image of a 3D point X of C 1 , and any point X of C 2 is the image of a 3D point X of C 2 .

3.6. Test of the Method in Simulation

  • Contour simulation
To illustrate the different stages of the method, a gallery model will be used. Figure 16 shows the camera represented by a pyramid and the projector represented by a cone positioned inside the model. In this example, the axis of the cone is parallel to the axis of the camera. The vertex O P is positioned above and to the right of the camera; it belongs to the plane which passes through the optical center O C and which is orthogonal to the camera axis. Consequently, the projection of O P in the image plane is a point at infinity (an ideal point) and the projections of the generatrices all belong to lines parallel to each other.
The 3D curve C is the intersection between the cone and the model. The set of 3D points on this curve is thus obtained from the intersections of the generatrices of the cone with the model, and they are shown in black in Figure 16. The number of 3D points is equal to the number of generatrices chosen to represent the cone. These 3D points are then projected onto the image to obtain a set of 2D points belonging to the curve c u r v e P r o j C o n e . These 2D points are shown in black in Figure 16.
The idea now is to obtain the 3D points of the curve C from the 2D points of the curve C . Assume that the parameters of the cone and the camera are known.
  • Calculating the generatrices and the lines containing their projection
The first step is to calculate the direction vectors w A and w B of the generatrices g A and g B . To do this, one needs to calculate the angles θ A and θ B using the Equation (11).
Once the direction vectors have been obtained, two 3D points are selected on g A and g B , projected into the image, and the parameters of the lines g A and g B are computed. The two lines are defined by their respective unit direction vectors u A and u B and by A 1 and B 1 , two points belonging to the two lines, respectively.
  • Determining the points A and B
Now, one needs to obtain the points A and B which are theoretically the points of tangency of the lines g A and g B with the curve C . The approach will be only presented for the point A since it will be the same for B .
A first solution would be to say that A is the point belonging to the line g A closest to the curve C . However, a slightly more generic approach has been chosen, the result of which is illustrated in Figure 17. It consists of considering A as the point belonging to the line g A which minimizes the sum of the squared distances between itself and the x% of the points on the curve C closest to the line g A . Let t be the parameter of the line. Therefore, looking for t such that (Equation (14)):
t = argmin t i = 1 n ( A X i ) T ( A X i ) h ( X i ) with A = A 1 + t u A
where:
  • X i are the x% of the points on the curve C closest to the line g A .
  • h : R 3 R + is a function whose aim is to reduce the impact of the points X i far from the line g A in minimization.
Figure 17 shows the points A and B obtained, with 40% of the points closest to g A in yellow and 40% of the points closest to g B in orange. Developing the sum to be minimized in Equation (14) gives:
i = 1 n ( A X i ) T ( A X i ) h ( X i ) = i = 1 n ( A 1 + t u A X i ) T ( A 1 + t u A X i ) h ( X i ) = i = 1 n ( t u A + v i ) T ( t u A + v i ) h ( X i ) with v i = A 1 X i = i = 1 n ( t 2 u A T u A + 2 t u A T v i + v i T v i ) h ( X i ) = t 2 u A T u A i = 1 n 1 h ( X i ) + t 2 i = 1 n u A T v i h ( X i ) + i = 1 n v i T v i h ( X i ) = a t 2 + 2 b t + c
with
a = u A T u A i = 1 n 1 h ( X i ) b = i = 1 n u A T v i h ( X i ) c = i = 1 n v i T v i h ( X i )
Since a must be positive, the polynomial a t 2 + 2 b t + c has a minimum when its derivative cancels, i.e., when:
2 a t + 2 b = 0 t = b a
This gives:
A = A 1 b a u A
Using the same procedure, the point B can be obtained. This produces the line ( A B ) (Figure 17) which will be used to divide the curve C in two.
  • Separation of the curve C for intersection selection
The points A and B are determined in a perfectible way in reality. Therefore, a minimum distance is defined around the line ( A B ) called e which the points of the curve C must respect. Points whose distance orthogonal to the line is less than e are considered indeterminate and will therefore not be taken into account in the 3D reconstruction of the contour. In practice, the value of e will depend on the quality of the measurements taken.
This separation of the C curve into two curves is illustrated in Figure 18 with C 1 in cyan and C 2 in magenta.
To obtain C 1 or C 2 , the point p ref is used (Equation (12)) obtained via the angle ν (Equation (13)) and by applying the following two conditions:
X C 2 if sign n A B T ( X A ) = sign n A B T ( p ref A ) X C 1 if sign n A B T ( X A ) = sign n A B T ( p ref A )
with n A B , the normal vector to the line ( A B ) .
  • Three-dimensional reconstruction
Now, points X which are on the curve C are known as they belong to the curve C 1 or the curve C 2 . One therefore can apply what was presented in Section 3.3 and Section 3.5 to reconstruct these points in 3D.
Figure 19 takes the 3D scene from Figure 16 and adds the reconstructed 3D points. The calculated differences between these reconstructed 3D points and the initial 3D points of the simulation (obtained via the intersection between the generatrices of the cone and the gallery model) are below 10 8 m, which is very small. These deviations are due to numerical errors in the calculations at the various stages. This proves that these reconstructed 3D points are the same as the initial 3D points and that our method therefore works. Now the method will be tested with real data.

4. Results

The local heritage offers access, between two laboratory buildings, to the Saint-Clément aqueduct (more commonly known as the Arceaux aqueduct) which was built in the 18th century. It is no longer in use, but in dry periods it is an ideal experimental platform. The gallery is accessible via trapdoors, which, once closed, plunge one into total darkness. This confined space recreates experimental conditions similar to those in a karstic environment, but in a dry environment whose dimensions are easy to measure. Indeed, the aqueduct is simply a long corridor whose left and right walls are parallel and separated by a width of 62 cm, which were measured with a laser rangefinder.

4.1. Experimental Setup

The camera used is the Nikon D7000 reflex with a 18 mm focal length and a resolution of 2464 × 1632 pixels (PlongeImage, Bordeaux, France).
The projector is a DIVEPRO M35 (La Palanquee, Palavas, France) dive light with an aperture angle of 145 ° in air and 90 ° in water. As the aperture angle is too large to capture the full projection of light in this experiment, a custom-made plastic cylinder (obtained from the 3D printer) was added to reduce it (Figure 20), as can be seen in Figure 21. Its internal diameter is the same as the diameter of the projector, i.e., d = 53.8 mm. The cylinder protrudes from the projector by l = 93 mm. It is possible to calculate the half-angle of aperture α as a function of the lengths l and d:
α = arctan d / 2 l = arctan 26.9 93 = 16.13 °
This value will be used to check the order of magnitude of future α values for the cone estimation.
The camera and projector are placed on a rigid support as shown in Figure 21. They had to be fixed to the support because the slightest rotation of one in relation to the other alters the transformation between the camera and the projector, which is estimated during the projector calibration.
To form an idea of the magnitude of this transformation, a caliper was used to measure the translations between the camera and the projector along the camera’s x, y and z axes. The rotation between the two is supposed to be small because an identical orientation for both was chosen. Here is a summary of the various measurements of the projector parameters (lengths in m):
α     16.13 ° O P { C }    0.2 0 0.07 T O P { C }    0.212 m d { C }    0 0 1 T

4.2. Camera Calibration Results

To calibrate the camera, Zhang’s method [42] was used with the help of eight images of a flat chessboard test pattern measuring 7 × 10 squares and 36 mm wide (Figure 22). In each of the eight images, the position of the test pattern relative to the camera is different (Figure 23), which is a 3D representation of the test pattern in eight different positions. Here are the camera parameters estimated via calibration:
  • the intrinsic matrix K obtained is:
    K = 1908.56 0.0 1227.53 0.0 1909.94 832.48 0.0 0.0 1.0
  • the resulting focal length is 18.2 mm.
It can be seen that the orders of magnitude of the various camera parameters are consistent with the manufacturer’s parameters. In addition, the average reprojection errors are of the order of 1/10 of a pixel. This level of error is quite satisfactory for all eight shots, each containing 56 calibration points.

4.3. Projector Calibration Results

To calibrate the projector, i.e., estimate the parameters of the cone, a white wall was used on which a chessboard pattern was hung. The first step is to obtain 3D points belonging to the cone by detecting the contour of the light projected onto the wall, as explained in the Section 3.2. To maximise the contrast of the light projected onto the white wall, the only light source in the room is the projector.
For this calibration, five images were captured of light projected onto the wall with the axis of the cone almost orthogonal to the wall (Figure 24). The shape of the projected light therefore resembles a circle.
Each detected contour is transformed into an ellipse, shown in blue in Figure 24. To obtain the ellipses, the method detailed in [46] was used, but iteratively with the RANSAC algorithm [48] to eliminate certain outlying contour points. The contour points eliminated by RANSAC are shown in red in Figure 24 while the points used to obtain the ellipses are in green.
Using the pattern detected in each image, the equations for the five planes were obtained. From these five planes and the five ellipses obtained, the five elliptical sections are calculated. They are represented in 3D in Figure 25 with the camera and the test pattern.
The image on the left shows all the elements expressed in the chessboard pattern frame, so one can see the different camera positions used to capture the five images.
The image on the right shows all the elements expressed in the camera frame. It is in this frame of reference that the elliptical sections must be expressed in order to estimate the cone.
Three-dimensional points are extracted and expressed in the camera frame of reference for each elliptical section. In this case, 200 points per section were recovered, making a total of 1000 3D points. Note that these 3D points are at a distance from the camera frame of between 1.50 m and 2.17 m.
Now that 1000 3D points have been obtained, which are supposed to belong to the cone, the next step is to define an initial cone for the iterative minimization algorithm. This is shown in red in Figure 26. An initial cone is chosen, that is far from the solution for greater visibility in Figure 26 and also to prove that the method is capable of converging even when the initial cone is far from the solution. Obviously, this is an experimental result and not a mathematical proof. This optimisation problem is probably non-convex; therefore, an initial cone close to the solution should ideally be chosen using the aperture angle previously calculated and the relative pose measurement between the projector and the cone.
The algorithm is then applied to obtain the optimised cone shown in gray. Its estimated parameters (lengths in meters) are:
  • α = 14.79 °
  • O P { C } = 0.1956 0.0092 0.0759 T
  • O P { C } = 0.210 m
  • d { C } = 0.013 0.080 0.997 T
To ensure the validity of the cone obtained, the estimated parameters are compared with the previously measured parameters. Table 1 compares the estimated opening angle and the estimated gap with the measurements. As the measurements can be improved, this comparison at least proves that the order of magnitude of the estimated parameters is consistent.
Another point to check is the minimization error for estimating the cone. In this case, this error is directly linked to the orthogonal distances between the 3D points of the elliptical sections and the cone, since it is the root mean square (RMS) of these distances that the algorithm is trying to minimize during estimation. In Figure 26, the color of each 3D point is defined by its orthogonal distance from the cone via the colormap Jet (the colormap bar is shown in the figure) and the following function:
f : R + [ 0 , 1 ] d d d + 0.1
where d is the orthogonal distance, and m is chosen as the median of all these distances (here, m = 1.4 mm).
The statistics for these distances can be found in Table 2. In particular, a maximum distance of 4.2 mm and an overall mean of 1.5 mm are obtained. The overall RMS is 1.9 mm, which corresponds to approximately 0.13% of 1.50 m which is the smallest distance between a 3D point of an elliptical section and the camera frame. Table 3 represents the statistics between the 3D points of each elliptical section and the reconstructed 3D points. A maximum distance of 226.1 mm and an overall mean of 29.0 mm are obtained.
In view of the analysis of the orthogonal distances, the minimization is satisfactory, confirming once again that the estimated parameters of the cone are at least consistent with their true values.

4.4. Three-Dimensional Results

The shape of the aqueduct is a long narrow corridor, which makes it possible to fully distinguish the light projection in the images. Six images were captured, taken at different distances from a chessboard pattern placed in the aqueduct. For each image, the light contours were checked manually, to isolate the triangulation method from the contour detection method, which is still in need of improvement. For each contour extracted, part of the contour points were manually delimited on the left wall and another part on the right wall. These delimitations can be seen in Figure 27. It shows an example of an extracted contour in two images, with contour points on the left wall in red, on the right wall in blue and the rest of the contour points in green.
Only the contours of the left and right walls will be used for triangulation, as this will allow us to check that the 3D points obtained respect the dimensions of the aqueduct corridor. To obtain these 3D points, it is necessary, as usual, to define which intersection is relevant for each radius associated with the contour points. In this situation, one can see that each contour point on the left wall implies a first intersection; additionally, each contour point on the right wall implies a second intersection. Figure 28 illustrates the method presented in Section 3.4 and Section 3.5 for the two images in Figure 27:
  • Lines g A in yellow, g B in orange and their intersection with the curve (Figure 28a,d)
  • Separation of the curves (Figure 28b,e)
  • Intersections with the walls left in cyan and right in magenta (Figure 28c,f)
Figure Figure 29a shows the result of the reconstruction of the contours of the six images. The shape of the 3D reconstructions of the contours is close to a hyperbola, which is consistent since the intersection between a plane and a cone in this configuration is supposed to be a hyperbola.
A comparison of this 3D reconstruction with reality is now possible. The aqueduct is a corridor whose left and right walls are parallel and separated by a width of 62 cm. One can therefore check that:
  • the 3D points resulting from the left part and the right part, respectively, of the contour belong to the same plane so they are supposed to be coplanar,
  • If a plane is estimated from the left 3D points and a plane from the right 3D points, they must be parallel and separated by a distance of approximately 62 cm.
As shown in Figure 29b, estimations are obtained for the left wall in red and for the right wall in blue by using a least square method.
To evaluate coplanarity, the orthogonal distance between 3D points and their associated plane is computed. For left points, Table 4a presents the results for the six positions of the sensor and Table 4b presents the results for all positions with an average deviation of 13.9 mm. For right points, Table 4c presents the results for the six positions of the sensor and Table 4d presents the results for all positions with an average deviation of 24.1 mm.
To check the gap between the two estimated planes, the distances are computed between the left 3D points and the right 3D points, respectively, and the right and left plane, respectively. To estimate the distance between the left points and the right plane, Table 5a presents the results for the six positions of the sensor and Table 5b presents the results for all positions with an average distance of 584.9 mm. To estimate the distance between the left points and the right plane, Table 5c presents the results for the six positions of the sensor and Table 5d presents the results for all positions with an average distance of 583.6 mm. Both distances are near the measured distance of 620 mm.
The two estimated planes are almost parallel since the angle calculated between the two normal vectors is 178.9 ° .

5. Discussion

  • Method
This original approach is based on the principles of structured light, leveraging the precise projection of light contours onto the karstic surfaces. It functions by capturing the resultant light patterns with a camera leading to reconstruct detailed three-dimensional representations of the subsurface structures. A classical method is used for the camera calibration. However, for the projector calibration estimation, each step is precisely detailed, namely the 3D point generation, the cone estimation, the estimation of the 3D points obtained by the intersection of the cone and its projection in the camera image plane. The main limitation of this approach is the projector contour detection. Indeed, since the boundary between the light and dark zones is not well defined, a detection method based on threshold brightness will be imprecise. It could be solved using a laser cone projector; however, due to the experimental conditions and the fact that the system will be carried by divers, it will be dangerous for their eyes. An active contour model such as that presented in ref. [49] could be used.
  • Experimental setup
For the validation of the method, a reflex Nikon D7000 with an underwater housing was used, allowing a good resolution (2464 × 1632 pixels), but also allowing for the possibility of adjusting the many settings that can be changed on this type of camera. However, its size remains substantial. The projector (DIVEPRO M35) has a wide angle of aperture and needs a custom-made cylinder to reduce its size. Maybe a dedicated projector with a correct aperture could be better. Also, to explore a new configuration of this system, some simulations were made with one projector and four cameras (Figure 30) to solve the error of the transition zone (see Figure 19). Another idea concerns using several projectors with localization sensors with respect to one or more cameras to have different configurations and measurement redundancy.
An embedded computer would allow the detection and 3D reconstruction directly.
  • Calibration results
For the camera, the results in dry conditions with a standard chessboard calibration pattern are correct enough, with fewer than 1/10 pixels. However, for underwater calibration, the results are dependent on water turbidity. A solution could be to measure the refraction index of the water and then apply a correction to the camera parameters. For the projector, the same kind of remarks can be made. A first estimation of the cone parameters is performed in dry conditions allowing for an estimation of the errors on its parameters, namely 8% for the angle α and 1% for its position with respect to the camera. Performing the experiment in dry conditions, also knowing the refraction index, could provide useful information to compute the new value of the aperture angle α , avoiding complex experimentation.This calibration is also dependent on the camera calibration; therefore, if the camera calibration is correct, the projector calibration will be accurate.
  • Three-dimensional reconstruction
The experimentation was conducted in dry conditions with known dimensions of the “cave” (aqueduct). The planes estimation was obtained with a mean value less than 2.5 cm, the distance estimation between the planes was obtained with a mean value less than 3 cm and the angle estimation between the normal range of the planes was obtained with a mean value near 180 ° . Validation of the method in wet conditions should also be acceptable. However, experimentation must be performed in dry conditions, first in a pool with good visibility conditions. Then a second experiment must be performed in a real cave. One can be reminded that the sites are dependent on administrative authorizations and especially weather conditions; currently, heavy rains make them impassable. Experiments were carried out in a pool a few months ago, but unfortunately the turbidity was such that the detection of the halo contours was problematic, not allowing for a presentation of these results. A new experiment is planned in the coming weeks.

6. Conclusions

The aim of this paper was to propose a 3D reconstruction solution designed for underwater galleries found in karstic environments. The study of these environments represents a considerable challenge for the future, as they could provide part of humanity’s water needs. Exploration remains the best approach for acquiring reliable data on the structure, and it is mainly carried out by qualified divers using topography methods with Ariane wires and section-by-section surveys. However, on extensive karstic networks, this approach presents a major risk, as the duration of these surveys is time-consuming, leading divers to long stops. The future is therefore more likely to lie with robotics, but there are still many technical challenges to be overcome before underwater drones that are sufficiently autonomous can be sent out to explore this type of environment. The present research proposed an original method of 3D scanning in this type of environment. This approach uses a combination of a camera and a cone-shaped projector. The parameters of the light cone have been fixed, namely its center, its direction vector and its angular aperture. A method for calibrating the cone was presented, enabling the position of the 3D points obtained by intersecting the visual rays of the contour of the projector halo as seen by the camera with the cone whose parameters have been estimated. The calibration of the camera is performed thanks to a classical approach. A simulation of this method was presented and the results validate the proposal. The first experiment in real conditions was conducted in a disused aqueduct (without water) which had the advantage of reproducing an environment close to that of a karstic aquifer, namely narrow, without light and above all without water to continue the evaluation. The camera calibration was validated with an error reconstruction less than 0.1 pixel and camera parameters close to technical values. The cone calibration led to an angles of error less than 2 degrees and distance with respect to the camera of less than 2 mm. The results were validated by testing the coplanarity of the aqueduct walls to within 25 mm and estimating the known distance between these walls to within 36 mm. Experiments have been performed in real karstic environments; however, due to complex meteorological conditions, visual acquisition was not possible. Thus, future experiments will be performed in a more controlled environment, such as a swimming pool, and then finished in a real aquifer with clearer water.

Author Contributions

Conceptualization, Q.M., S.D. and J.T.; methodology, Q.M., S.D. and J.T.; software, Q.M. and S.D.; validation, Q.M., S.D. and J.T.; formal analysis, Q.M., S.D. and J.T.; investigation, Q.M., S.D. and J.T.; resources, Q.M., S.D. and J.T.; data curation, Q.M., S.D. and J.T.; writing—original draft preparation, J.T.; writing—review and editing, Q.M., S.D. and J.T.; visualization, Q.M., S.D. and J.T.; supervision, J.T.; project administration, J.T.; funding acquisition, J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be available upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gleick, P.H. Water in Crisis Pacific Institute for Studies in Dev, Environment & Security Stockholm Env Institute; Oxford Univ. Press: Oxford, UK, 1993; Volume 9. [Google Scholar]
  2. Bakalowicz, M. Karst groundwater: A challenge for new resources. Hydrogeol. J. 2005, 13, 148–160. [Google Scholar] [CrossRef]
  3. Bakalowicz, M. Karst and karst groundwater resources in the Mediterranean. Environ. Earth Sci. 2015, 74, 5–14. [Google Scholar] [CrossRef]
  4. Cong, Y.; Gu, C.; Zhang, T.; Gao, Y. Underwater robot sensing technology: A survey. Fundam. Res. 2021, 1, 337–345. [Google Scholar] [CrossRef]
  5. Chutia, S.; Kakoty, N.M.; Deka, D. A Review of Underwater Robotics, Navigation, Sensing Techniques and Applications. In Proceedings of the 2017 3rd International Conference on Advances in Robotics, New Delhi, India, 28 June–2 July 2017; ACM: New York, NY, USA, 2017. [Google Scholar] [CrossRef]
  6. Massot-Campos, M.; Oliver-Codina, G. Optical sensors and methods for underwater 3D reconstruction. Sensors 2015, 15, 31525–31557. [Google Scholar] [CrossRef] [PubMed]
  7. Hu, K.; Wang, T.; Shen, C.; Weng, C.; Zhou, F.; Xia, M.; Weng, L. Overview of Underwater 3D Reconstruction Technology Based on Optical Images. J. Mar. Sci. Eng. 2023, 11, 949. [Google Scholar] [CrossRef]
  8. Snyder, J. Doppler Velocity Log (DVL) navigation for observation-class ROVs. In Proceedings of the OCEANS 2010 MTS/IEEE SEATTLE, Seattle, WA, USA, 20–23 September 2010; IEEE: New York, NY, USA, 2010; pp. 1–9. [Google Scholar] [CrossRef]
  9. Lee, C.M.; Lee, P.M.; Hong, S.W.; Kim, S.M.; Seong, W. Underwater navigation system based on inertial sensor and doppler velocity log using indirect feedback kalman filter. Int. J. Offshore Polar Eng. 2005, 15, ISOPE-05-15-2-088. [Google Scholar]
  10. Brown, C.J.; Blondel, P. Developments in the application of multibeam sonar backscatter for seafloor habitat mapping. Appl. Acoust. 2009, 70, 1242–1247. [Google Scholar] [CrossRef]
  11. Gary, M.; Fairfield, N.; Stone, W.C.; Wettergreen, D.; Kantor, G.; Sharp, J.M., Jr. 3D mapping and characterization of Sistema Zacatón from DEPTHX (DE ep P hreatic TH ermal e X plorer). In Proceedings of the Sinkholes and the Engineering and Environmental Impacts of Karst, Tallahassee, FL, USA, 22–26 September 2008; pp. 202–212. [Google Scholar] [CrossRef]
  12. Kantor, G.; Fairfield, N.; Jonak, D.; Wettergreen, D. Experiments in navigation and mapping with a hovering AUV. In Field and Service Robotics; Springer: Berlin/Heidelberg, Germany, 2008; pp. 115–124. [Google Scholar] [CrossRef]
  13. Mallios, A.; Ridao, P.; Ribas, D.; Carreras, M.; Camilli, R. Toward autonomous exploration in confined underwater environments. J. Field Robot. 2016, 33, 994–1012. [Google Scholar] [CrossRef]
  14. Hartley, R.I.; Sturm, P. Triangulation. Comput. Vis. Image Underst. 1997, 68, 146–157. [Google Scholar] [CrossRef]
  15. Oleari, F.; Kallasi, F.; Rizzini, D.L.; Aleotti, J.; Caselli, S. An underwater stereo vision system: From design to deployment and dataset acquisition. In Proceedings of the OCEANS 2015, Genova, Italy, 18–21 May 2015; IEEE: New York, NY, USA, 2015. [Google Scholar] [CrossRef]
  16. Massot-Campos, M.; Oliver-Codina, G. One-shot underwater 3D reconstruction. In Proceedings of the 2014 IEEE Emerging Technology and Factory Automation (ETFA), Barcelona, Spain, 16–19 September 2014; IEEE: New York, NY, USA, 2014; pp. 1–4. [Google Scholar] [CrossRef]
  17. Chi, S.; Xie, Z.; Chen, W. A Laser Line Auto-Scanning System for Underwater 3D Reconstruction. Sensors 2016, 16, 1534. [Google Scholar] [CrossRef]
  18. Bräuer-Burchardt, C.; Heinze, M.; Schmidt, I.; Kühmstedt, P.; Notni, G. Underwater 3D Surface Measurement Using Fringe Projection Based Scanning Devices. Sensors 2015, 16, 13. [Google Scholar] [CrossRef]
  19. Meline, A.; Triboulet, J.; Jouvencel, B. Comparative study of two 3D reconstruction methods for underwater archaeology. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012; IEEE: New York, NY, USA, 2012; pp. 740–745. [Google Scholar] [CrossRef]
  20. Onmek, Y.; Triboulet, J.; Druon, S.; Meline, A.; Jouvencel, B. Evaluation of underwater 3D reconstruction methods for Archaeological Objects: Case study of Anchor at Mediterranean Sea. In Proceedings of the 2017 3rd International Conference on Control, Automation and Robotics (ICCAR), Nagoya, Japan, 22–24 April 2017; IEEE: New York, NY, USA, 2017; pp. 394–398. [Google Scholar] [CrossRef]
  21. Brandou, V.; Allais, A.G.; Perrier, M.; Malis, E.; Rives, P.; Sarrazin, J.; Sarradin, P.M. 3D reconstruction of natural underwater scenes using the stereovision system IRIS. In Proceedings of the OCEANS 2007-Europe, Aberdeen, Scotland, 18–21 June 2007; IEEE: New York, NY, USA, 2007; pp. 1–6. [Google Scholar] [CrossRef]
  22. Beall, C.; Lawrence, B.J.; Ila, V.; Dellaert, F. 3D reconstruction of underwater structures. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; IEEE: New York, NY, USA, 2010; pp. 4418–4423. [Google Scholar] [CrossRef]
  23. Zhou, Y.; Li, Q.; Ye, Q.; Yu, D.; Yu, Z.; Liu, Y. A binocular vision-based underwater object size measurement paradigm: Calibration-Detection-Measurement (C-D-M). Measurement 2023, 216, 112997. [Google Scholar] [CrossRef]
  24. Weidner, N.; Rahman, S.; Li, A.Q.; Rekleitis, I. Underwater cave mapping using stereo vision. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May 29–3 June 2017; IEEE: New York, NY, USA, 2017; pp. 5709–5715. [Google Scholar] [CrossRef]
  25. Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  26. Rahman, S.; Li, A.Q.; Rekleitis, I. Svin2: An underwater slam system using sonar, visual, inertial, and depth sensor. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; IEEE: New York, NY, USA, 2019; pp. 1861–1868. [Google Scholar] [CrossRef]
  27. Hidalgo, F.; Braunl, T. Review of underwater SLAM techniques. In Proceedings of the 2015 6th International Conference on Automation, Robotics and Applications (ICARA), Queenstown, New Zealand, 17–19 February 2015; IEEE: New York, NY, USA, 2015. [Google Scholar] [CrossRef]
  28. Zhao, W.; He, T.; Sani, A.Y.M.; Yao, T. Review of SLAM Techniques For Autonomous Underwater Vehicles. In Proceedings of the 2019 International Conference on Robotics, Intelligent Control and Artificial Intelligence, Shenyang, China, 8–11 August 2019; ACM: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  29. Joshi, B.; Xanthidis, M.; Roznere, M.; Burgdorfer, N.J.; Mordohai, P.; Li, A.Q.; Rekleitis, I. Underwater Exploration and Mapping. In Proceedings of the 2022 IEEE/OES Autonomous Underwater Vehicles Symposium (AUV), Singapore, 19–21 September 2022; pp. 1–7. [Google Scholar] [CrossRef]
  30. am Ende, B.A. 3D mapping of underwater caves. IEEE Comput. Graph. Appl. 2001, 21, 14–20. [Google Scholar] [CrossRef]
  31. Fan, H.; Qi, L.; Chen, C.; Rao, Y.; Kong, L.; Dong, J.; Yu, H. Underwater Optical 3-D Reconstruction of Photometric Stereo Considering Light Refraction and Attenuation. IEEE J. Ocean. Eng. 2022, 47, 46–58. [Google Scholar] [CrossRef]
  32. Papp, R.Z.; Szentpéteri, K.; Balázs, G.; Topa, B.A.; Zajzon, N. 3D photogrammetry of flooded mines and caves with the UX-1 series underwater exploration robots—The UNEXUP Project. In Proceedings of the EGU General Assembly, Online, 19–30 April 2021. [Google Scholar] [CrossRef]
  33. Hou, J.; Ye, X. Real-time Underwater 3D Reconstruction Method Based on Stereo Camera. In Proceedings of the 2022 IEEE International Conference on Mechatronics and Automation (ICMA), Guilin, China, 7–10 August 2022; IEEE: New York, NY, USA, 2022. [Google Scholar] [CrossRef]
  34. Ju, Y.; Lam, K.M.; Xie, W.; Zhou, H.; Dong, J.; Shi, B. Deep Learning Methods for Calibrated Photometric Stereo and Beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 1–19. [Google Scholar] [CrossRef] [PubMed]
  35. Abdullah, A.; Barua, T.; Tibbetts, R.; Chen, Z.; Islam, M.J.; Rekleitis, I. CaveSeg: Deep Semantic Segmentation and Scene Parsing for Autonomous Underwater Cave Exploration. arXiv 2023, arXiv:2309.11038. [Google Scholar] [CrossRef]
  36. Wu, X.; Wang, J.; Li, J.; Zhang, X. Retrieval of siltation 3D properties in artificially created water conveyance tunnels using image-based 3D reconstruction. Measurement 2023, 211, 112586. [Google Scholar] [CrossRef]
  37. Fan, J.; Wang, X.; Zhou, C.; Ou, Y.; Jing, F.; Hou, Z. Development, Calibration, and Image Processing of Underwater Structured Light Vision System: A Survey. IEEE Trans. Instrum. Meas. 2023, 72, 1–18. [Google Scholar] [CrossRef]
  38. Wang, W.; Joshi, B.; Burgdorfer, N.; Batsosc, K.; Lid, A.Q.; Mordohaia, P.; Rekleitisb, I. Real-Time Dense 3D Mapping of Underwater Environments. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; IEEE: New York, NY, USA, 2023. [Google Scholar] [CrossRef]
  39. Bräuer-Burchardt, C.; Munkelt, C.; Bleier, M.; Heinze, M.; Gebhart, I.; Kühmstedt, P.; Notni, G. Underwater 3D Scanning System for Cultural Heritage Documentation. Remote Sens. 2023, 15, 1864. [Google Scholar] [CrossRef]
  40. Yu, B.; Tibbetts, R.; Barna, T.; Morales, A.; Rekleitis, I.; Islam, M.J. Weakly Supervised Caveline Detection for AUV Navigation Inside Underwater Caves. In Proceedings of the 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA, 1–5 October 2023; IEEE: New York, NY, USA, 2023. [Google Scholar] [CrossRef]
  41. Hu, C.; Zhu, S.; Song, W. Real-time Underwater 3D Reconstruction Based on Monocular Image. In Proceedings of the 2022 IEEE International Conference on Robotics and Biomimetics (ROBIO), Jinghong, China, 5–9 December 2022; pp. 1239–1244. [Google Scholar] [CrossRef]
  42. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  43. Marchand, E.; Uchiyama, H.; Spindler, F. Pose estimation for augmented reality: A hands-on survey. IEEE Trans. Vis. Comput. Graph. 2015, 22, 2633–2651. [Google Scholar] [CrossRef] [PubMed]
  44. Massone, Q.; Druon, S.; Triboulet, J. An original 3D reconstruction method using a conical light and a camera in underwater caves. In Proceedings of the 4th International Conference on Control and Computer Vision, Macau, China, 13–15 August 2021; pp. 126–134. [Google Scholar] [CrossRef]
  45. Massone, Q.; Druon, S.; Breux, Y.; Triboulet, J. Contour-based approach for 3D mapping of underwater galleries. In Proceedings of the Global Oceans 2020: Singapore–US Gulf Coast, Singapore, 5–30 October 2020; IEEE: New York, NY, USA, 2020; pp. 1–6. [Google Scholar] [CrossRef]
  46. Halır, R.; Flusser, J. Numerically stable direct least squares fitting of ellipses. In Proceedings of the 6th International Conference in Central Europe on Computer Graphics and Visualization, Plzen-Bory, Czech Republic, 9–13 February 1998; WSCG, Citeseer: Princeton, NJ, USA, 1998; Volume 98, pp. 125–132. [Google Scholar]
  47. Moré, J.J. The Levenberg-Marquardt algorithm: Implementation and theory. In Numerical Analysis; Springer: Berlin/Heidelberg, Germany, 1978; pp. 105–116. [Google Scholar] [CrossRef]
  48. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  49. Kass, M.; Witkin, A.; Terzopoulos, D. Snakes: Active contour models. Int. J. Comput. Vis. 1988, 1, 321–331. [Google Scholar] [CrossRef]
Figure 1. Earth’s water distribution.
Figure 1. Earth’s water distribution.
Sensors 24 04024 g001
Figure 2. Three-dimensional reconstruction methods in underwater environments.
Figure 2. Three-dimensional reconstruction methods in underwater environments.
Sensors 24 04024 g002
Figure 3. Flowchart of the camera + projector method.
Figure 3. Flowchart of the camera + projector method.
Sensors 24 04024 g003
Figure 4. Diagram of the system consisting of the light projector represented by a cone of revolution and the camera observing the light projection on a plane.
Figure 4. Diagram of the system consisting of the light projector represented by a cone of revolution and the camera observing the light projection on a plane.
Sensors 24 04024 g004
Figure 5. Parameterization of the cone using its generatrices. With this parameterization, one can find the closest generatrix to an external point M and thus find the orthogonal projection H of this point on the cone.
Figure 5. Parameterization of the cone using its generatrices. With this parameterization, one can find the closest generatrix to an external point M and thus find the orthogonal projection H of this point on the cone.
Sensors 24 04024 g005
Figure 6. Example showing the closest points on the upper part of the cone (in red) to points outside the surface of the cone (in blue).
Figure 6. Example showing the closest points on the upper part of the cone (in red) to points outside the surface of the cone (in blue).
Sensors 24 04024 g006
Figure 7. Example of detecting the contour of a halo of light on a wall using a chessboard to estimate the plane pose. The contour is in green and the adjusted ellipse in blue.
Figure 7. Example of detecting the contour of a halo of light on a wall using a chessboard to estimate the plane pose. The contour is in green and the adjusted ellipse in blue.
Sensors 24 04024 g007
Figure 8. Method for obtaining several elliptical sections of the cone.
Figure 8. Method for obtaining several elliptical sections of the cone.
Sensors 24 04024 g008
Figure 9. Representation of the camera/cone pair in an orthogonal projection 2D view where the projection axis is perpendicular to z P and z C . The relative orientation of the cone with the camera can be defined here by a single angle called μ . This is only true if one considers that the camera has an infinite field of view and is therefore symmetrical about its axis defined by z C (the cone is basically symmetrical about its axis defined by z P = d ).
Figure 9. Representation of the camera/cone pair in an orthogonal projection 2D view where the projection axis is perpendicular to z P and z C . The relative orientation of the cone with the camera can be defined here by a single angle called μ . This is only true if one considers that the camera has an infinite field of view and is therefore symmetrical about its axis defined by z C (the cone is basically symmetrical about its axis defined by z P = d ).
Sensors 24 04024 g009
Figure 10. Representation of the plane Π g when it is not tangent to the cone passing through the optical center and the two generatrices g 1 and g 2 .
Figure 10. Representation of the plane Π g when it is not tangent to the cone passing through the optical center and the two generatrices g 1 and g 2 .
Sensors 24 04024 g010
Figure 11. Representation of Π g A (a) and Π g B (b), the only two planes tangent to the cone passing through the optical center.
Figure 11. Representation of Π g A (a) and Π g B (b), the only two planes tangent to the cone passing through the optical center.
Sensors 24 04024 g011
Figure 12. Visualization of the vectors t and s and their angles.
Figure 12. Visualization of the vectors t and s and their angles.
Sensors 24 04024 g012
Figure 13. Representation in the image plane and in the cone of the areas containing the first intersections (in cyan) and the second intersections (in magenta).
Figure 13. Representation in the image plane and in the cone of the areas containing the first intersections (in cyan) and the second intersections (in magenta).
Sensors 24 04024 g013
Figure 14. Representation in the image plane and in the cone of the areas containing the first intersections (in cyan) and the second intersections (in magenta) in the case where part of the cone (in brown) is behind the camera.
Figure 14. Representation in the image plane and in the cone of the areas containing the first intersections (in cyan) and the second intersections (in magenta) in the case where part of the cone (in brown) is behind the camera.
Sensors 24 04024 g014
Figure 15. Illustration of how to obtain p ref , the reference point in the image plane which indicates the relative position of the cone, and which is used to obtain the curves C 1 and C 2 .
Figure 15. Illustration of how to obtain p ref , the reference point in the image plane which indicates the relative position of the cone, and which is used to obtain the curves C 1 and C 2 .
Sensors 24 04024 g015
Figure 16. The camera (represented by a pyramid) and the projector (represented by a cone) arranged inside the model of our gallery to simulate a contour of light in an image. This contour is obtained by projecting the intersections between the generatrices of the cone and the model into the image.
Figure 16. The camera (represented by a pyramid) and the projector (represented by a cone) arranged inside the model of our gallery to simulate a contour of light in an image. This contour is obtained by projecting the intersections between the generatrices of the cone and the model into the image.
Sensors 24 04024 g016
Figure 17. Determining the points A and B for our previous example contour.
Figure 17. Determining the points A and B for our previous example contour.
Sensors 24 04024 g017
Figure 18. Determining the curves C 1 and C 2 for the previous example contour.
Figure 18. Determining the curves C 1 and C 2 for the previous example contour.
Sensors 24 04024 g018
Figure 19. The reconstructed 3D points added to the 3D scene of Figure 19 with the first intersections in cyan and the second intersections in magenta.
Figure 19. The reconstructed 3D points added to the 3D scene of Figure 19 with the first intersections in cyan and the second intersections in magenta.
Sensors 24 04024 g019
Figure 20. The diving lamp and its angle of aperture.
Figure 20. The diving lamp and its angle of aperture.
Sensors 24 04024 g020
Figure 21. The system used, consisting of a camera and a conical-shaped projector, to which a black cylinder has been added.
Figure 21. The system used, consisting of a camera and a conical-shaped projector, to which a black cylinder has been added.
Sensors 24 04024 g021
Figure 22. Image of one of the eight chessboard patterns taken at different poses for camera calibration using Zhang’s method.
Figure 22. Image of one of the eight chessboard patterns taken at different poses for camera calibration using Zhang’s method.
Sensors 24 04024 g022
Figure 23. Three-dimensional view of the chessboard in its eight poses in relation to the camera (X, Y, Z).
Figure 23. Three-dimensional view of the chessboard in its eight poses in relation to the camera (X, Y, Z).
Sensors 24 04024 g023
Figure 24. Two images of light projected onto the wall from the five where the axis of the projector is almost orthogonal to the wall (images 1 and 5).
Figure 24. Two images of light projected onto the wall from the five where the axis of the projector is almost orthogonal to the wall (images 1 and 5).
Sensors 24 04024 g024
Figure 25. Three-dimensional view of the elliptical sections, the camera and the chessboard pattern. In the image on the left, the elements are expressed in the pattern frame. In the image on the right, the elements are expressed in the camera frame (X, Y, Z).
Figure 25. Three-dimensional view of the elliptical sections, the camera and the chessboard pattern. In the image on the left, the elements are expressed in the pattern frame. In the image on the right, the elements are expressed in the camera frame (X, Y, Z).
Sensors 24 04024 g025
Figure 26. Representation of the estimated cone in relation to elliptical sections where the color of the points depends on the distance orthogonal to the cone (X, Y, Z).
Figure 26. Representation of the estimated cone in relation to elliptical sections where the color of the points depends on the distance orthogonal to the cone (X, Y, Z).
Sensors 24 04024 g026
Figure 27. Two images in the aqueduct, at different distances, with light contours.
Figure 27. Two images in the aqueduct, at different distances, with light contours.
Sensors 24 04024 g027aSensors 24 04024 g027b
Figure 28. Intersection selection. (a) Generatrices projection g A , g B (2.17 m). (b) Separation of the curves (2.17 m). (c) Intersections: left in cyan and right in magenta (2.17 m). (d) Generatrix projection l g a , l g b (1.50 m). (e) Separation of the curves (1.50 m). (f) Intersections: left in cyan and right in magenta (1.50 m).
Figure 28. Intersection selection. (a) Generatrices projection g A , g B (2.17 m). (b) Separation of the curves (2.17 m). (c) Intersections: left in cyan and right in magenta (2.17 m). (d) Generatrix projection l g a , l g b (1.50 m). (e) Separation of the curves (1.50 m). (f) Intersections: left in cyan and right in magenta (1.50 m).
Sensors 24 04024 g028
Figure 29. Three-dimensional reconstruction of the contours extracted from the six images (X, Y, Z).
Figure 29. Three-dimensional reconstruction of the contours extracted from the six images (X, Y, Z).
Sensors 24 04024 g029aSensors 24 04024 g029b
Figure 30. New sensor configuration with four cameras.
Figure 30. New sensor configuration with four cameras.
Sensors 24 04024 g030
Table 1. Comparison of estimated and measured cone parameters.
Table 1. Comparison of estimated and measured cone parameters.
Estimated ValuesMeasured ValuesRelative Deviation from Measurement
α 14.79 ° 16.13 ° 1.34 ° (−8%)
O P { C } 0.210 m0.212 m−0.002 m (−0.9%)
Table 2. Statistics of the orthogonal distances in mm between the 3D points of each elliptical section and the estimated cone.
Table 2. Statistics of the orthogonal distances in mm between the 3D points of each elliptical section and the estimated cone.
Dist. (mm)Max.Min.Med.Aver.RMS
Section 04.20.01.41.52.0
Section 13.30.01.51.51.8
Section 24.10.01.41.62.0
Section 33.40.01.61.61.8
Section 43.60.01.31.51.8
All4.20.01.41.51.9
Table 3. Distance statistics in mm between the 3D points of each elliptical section and the reconstructed 3D points.
Table 3. Distance statistics in mm between the 3D points of each elliptical section and the reconstructed 3D points.
Dist. (mm)Max.Min.Med.Aver.RMS
Section 094.00.428.527.732.6
Section 1149.60.023.630.942.8
Section 2226.10.118.633.651.3
Section 3179.30.118.630.044.7
Section 4114.30.119.823.029.1
All226.10.020.629.040.9
Table 4. Assessment of coplanarity of 3D points.
Table 4. Assessment of coplanarity of 3D points.
(a) Orthogonal distances between 3D left points from image i and estimated left plane i
Dist. (mm)Max.Min.Med.Aver.RMS
image 036.40.02.13.65.6
image 160.10.04.78.013.2
image 289.90.09.013.320.3
image 324.60.05.45.76.7
image 436.10.05.15.87.6
image 533.60.03.95.67.9
(b) Orthogonal distances between 3D left points from all images and the global estimated left plane
Dist. (mm)Max.Min.Med.Aver.RMS
All220.50.011.013.919.9
(c) Orthogonal distances between 3D right points from image i and estimated right plane i
Dist. (mm)Max.Min.Med.Aver.RMS
Image 027.50.05.26.69.0
Image 144.20.018.516.620.0
Image 229.30.010.110.212.1
Image 321.30.07.17.58.8
Image 426.70.03.45.78.0
Image 546.40.06.59.112.3
(d) Orthogonal distances between 3D right points from all images and the global estimated right plane
Dist. (mm)Max.Min.Med.Aver.RMS
All113.90.020.424.130.5
Table 5. Measuring the distance between the two walls.
Table 5. Measuring the distance between the two walls.
(a) Orthogonal distances between 3D left points from image i and estimated right plane i
Dist. (mm)Max.Min.Med.Aver.RMS
image 0641.4576.6591.4593.1593.2
image 1622.0484.1607.9596.5597.1
image 2757.7542.6582.2592.6593.3
image 3596.6485.5574.8567.7568.2
image 4598.0482.0579.6569.7570.3
image 5619.4496.7592.0581.8582.5
(b) Orthogonal distances between 3D left points from all images and the global estimated right plane
Dist. (mm)Max.Min.Med.Aver.RMS
All778.4492.1581.9584.9585.2
(c) Orthogonal distances between 3D right points from image i and estimated left plane i
Dist. (mm)Max.Min.Med.Aver.RMS
Image 0628.6581.2594.0599.1599.2
Image 1602.9466.0584.5573.4574.4
Image 2767.6563.9573.9586.2587.5
Image 3597.1508.0565.2564.2564.7
Image 4625.6529.5609.6597.5598.1
Image 5603.2500.7578.2568.9569.6
(d) Orthogonal distances between 3D right points from all images and the global estimated left plane
Dist. (mm)Max.Min.Med.Aver.RMS
All691.1476.9585.2583.6584.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Massone, Q.; Druon, S.; Triboulet, J. A Novel 3D Reconstruction Sensor Using a Diving Lamp and a Camera for Underwater Cave Exploration. Sensors 2024, 24, 4024. https://doi.org/10.3390/s24124024

AMA Style

Massone Q, Druon S, Triboulet J. A Novel 3D Reconstruction Sensor Using a Diving Lamp and a Camera for Underwater Cave Exploration. Sensors. 2024; 24(12):4024. https://doi.org/10.3390/s24124024

Chicago/Turabian Style

Massone, Quentin, Sébastien Druon, and Jean Triboulet. 2024. "A Novel 3D Reconstruction Sensor Using a Diving Lamp and a Camera for Underwater Cave Exploration" Sensors 24, no. 12: 4024. https://doi.org/10.3390/s24124024

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop