Previous Article in Journal
A Metrological Approach to Developing Quality Testing Standards for Emerging Advanced Materials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Geometric Calibration of Thermal Infrared Cameras: A Comparative Analysis for Photogrammetric Data Fusion

by
Neil Sutherland
1,*,
Stuart Marsh
1,
Fabio Remondino
2,
Giulio Perda
2,
Paul Bryan
1,† and
Jon Mills
3
1
Nottingham Geospatial Institute, University of Nottingham, Nottingham NG7 2TU, UK
2
3D Optical Metrology (3DOM) Unit, Fondazione Bruno Kessler (FBK), Via Sommarive 18, 38121 Trento, Italy
3
School of Engineering, Newcastle University, Newcastle upon Tyne NE1 7RU, UK
*
Author to whom correspondence should be addressed.
This author is retired.
Metrology 2025, 5(3), 43; https://doi.org/10.3390/metrology5030043
Submission received: 10 April 2025 / Revised: 13 June 2025 / Accepted: 25 June 2025 / Published: 8 July 2025

Abstract

The determination of precise and reliable interior (IO) and relative (RO) orientation parameters for thermal infrared (TIR) cameras is critical for their subsequent use in photogrammetric processes. Although 2D calibration boards have become the predominant approach for TIR geometric calibration, these targets are susceptible to projective coupling and often introduce error through manual construction methods, necessitating the development of 3D targets tailored to TIR geometric calibration. Therefore, this paper evaluates TIR geometric calibration results obtained from 2D board and 3D field calibration approaches, documenting the construction, observation, and calculation of IO and RO parameters. This includes a comparative analysis of values derived from three popular commercial software packages commonly used for geometric calibration: MathWorks’ MATLAB, Agisoft Metashape, and Photometrix’s Australis. Furthermore, to assess the validity of derived parameters, two InfraRed Thermography 3D-Data Fusion (IRT-3DDF) methods are developed to model historic building façades and medieval frescoes. The results demonstrate the success of the proposed 3D field calibration targets for the calculation of both IO and RO parameters tailored to photogrammetric data fusion. Additionally, a novel combined TIR-RGB bundle block adjustment approach demonstrates the success of applying ‘out-of-the-box’ deep-learning neural networks for multi-modal image matching and thermal modelling. Considerations for the development of TIR geometric calibration approaches and the evolution of proposed IRT-3DDF methods are provided for future work.

1. Introduction

The determination of precise and reliable intrinsics for thermal infrared (TIR) cameras is critical for their subsequent application in photogrammetric processes. Whilst geometric calibration represents a largely routine procedure for cameras sensing in the visible spectrum (RGB) [1], where 2D boards or 3D coded test fields are commonly used, the process presents several challenges when undertaken for TIR cameras. Notably, these comprise the following: (1) the inherent low resolution of commercial TIR sensors (typically 180 × 120–640 × 480 pix), (2) the construction of accurate calibration targets with sufficient difference between material emissivities ( ε ), and (3) the imaging of control points with high contrast and low noise [2]. Consequently, the adoption of 2D calibration boards from the computer vision (CV) community has become the predominant approach for TIR geometric calibration, made using low-emissivity materials ( ε < 0.5 ; e.g., aluminum foil, copper, and stainless steel) contrasted with high-emissivity materials ( ε > 0.8 ; e.g., rubber, wood, and velvet) [3]. Such methods have been utilised to calibrate infrared sensors across different spectral wavelengths [4,5,6], for passive and active thermography [7,8,9], and for indoor and outdoor applications [10,11,12]. However, the overall accuracy of existing 2D calibration boards, often constructed through manual methods to generate calibration patterns, has required new methods for geometric calibration tailored to TIR cameras.
In a bid to improve the quality of calibration boards, several works have looked to reduce human intervention by employing superior manufacturing processes for accurate board construction. Usamentiaga et al.’s [2] critical study demonstrates the use of DiBond®, an aluminium composite material, and UV-printing to produce a cost-efficient, easy-to-manufacture, and accurate 2D calibration board. Similarly, ElSheikh et al.’s [13] comprehensive literature review analysed the best materials, heating techniques, and associated errors of existing geometric calibration methods for TIR cameras, using a stainless-steel (SS304) base plate with UV-printing and laser-engraving for calibration boards. Zhang et al. [14] demonstrated the suitability of high-tolerance 3D printing to improve accuracy, constructing 3D-printed elements (PLA and aluminium plate) configurable into a 2D checker-board pattern. Roshan et al. [15] introduced high-tolerance aluminium, ArUco, [16] components into a traditional checker-board, mitigating the established issues of partial pattern observation and extraction for 2D calibration boards. Although these approaches provide considerable improvements in target construction, 2D calibration boards remain prone to the projective coupling of interior (IO) and exterior (EO) orientation parameters [7], with higher correlations between parameters being a likely by-product of poorer network geometry typically imposed by low resolution TIR cameras.
Conversely, effective 3D calibration approaches reduce the propensity for correlations between (and across) IO and EO parameters, leading to greater overall accuracy for subsequent photogrammetric applications [1,7]. Much like 2D boards, 3D calibration fields have been constructed using various materials and pattern configurations for TIR geometric calibration, with Westfeld et al. [17] creating a 3D test field using velour and silver foil and Yastikli and Guler [18] using an iron calibration rig with 77 plastic markers of contrasting depths. Rizzi et al. [19] constructed a 3D test field from custom markers to calibrate a FLIR P640 thermal camera (640 × 480 pix), calculating material emissivity values to validate material choice. In a bid to calibrate thermal stereo set-ups, Rangel et al. [20] and Adán et al. [21] used reflective markers for the calibration of thermal cameras fixed with a depth camera and laser scanner, respectively. Both Luhmann et al. [22] (2D board and 3D field) and Dlesk et al. [23] (2D board, and 2-level and 3-level 3D fields) compared multiple calibration approaches, determining that 3D fields are superior for subsequent 3D thermal modelling. However, opportunities are still present to determine the performance of new 3D test fields for the following: the calculation of IO parameters, their comparison against existing 2D board approaches, and their application for InfraRed Thermography 3D-Data Fusion (IRT-3DDF) [24].
In parallel, the estimation of the relative orientation (RO) between sensors has played a crucial role in the development of data fusion methods using TIR cameras. By establishing a fixed RO between independent cameras [25], or exploiting tandem sensors within the same camera housing [26], the mapping of TIR-RGB image pairs can be realised through effective calibration. This relationship is of particular importance when there are no means of determining the poses of TIR images independently, often due to low contrast or noisy images providing few discernible features for photogrammetric image matching. Furthermore, if TIR and RGB images are to be captured concurrently, this fixed relationship can help mitigate sub-optimal conditions for each image modality during image acquisition. By assuming a fixed RO between multiple cameras or sensors, such information can be used as an a priori constraint for sensor fusion [27], to determine a universal transformation key between image pairs [23], or as a means of additional sensor integration [28]. Therefore, the determination of reliable approaches for relative pose estimations is critical for the use of multi-modal imagery within IRT-3DDF.
The exploitation of fixed stereo systems within IRT-3DDF has enabled calibration approaches to determine the RO between sensors and cameras for data fusion applications. For sensors in fixed relative orientation within the same camera housing, Lecomte et al. [29] and Macher and Landes [30] used MATLAB’s Stereo Camera Calibrator tool to estimate the relative pose of TIR and RGB sensors, the latter comparing results to known camera dimensions. Fixing several cameras on a stereo rig, Lerma et al. [31] present a distance-constrained self-calibration approach, obtaining more precise IO and EO parameters for a FLIR B4 thermal camera when integrating known baseline constraints. Ursine et al. [32] assessed the accuracy of RO parameters using a 2D board, determining on-site calibration as a suitable approach under favourable conditions. Notably, the development of reliable RO calibration methods and the application of derived parameters for novel IRT-3DDF methods remains overlooked and warrants further investigation.
Whilst the standard error of image coordinate residuals [13] and individual calibration parameters [7] have become common precision metrics for estimating the performance of CV geometric calibration, these cannot be used as a proxy for accuracy within subsequent photogrammetric applications [33]. Therefore, the metric evaluation of derived IO and RO parameters must be undertaken both during and after calibration for meaningful comparisons to be made regarding the quality of different TIR geometric calibration processes [22,34]. Furthermore, the appraisal of calibration methods requires emphasis to be placed on the use of IO and RO parameters for IRT-3DDF, particularly when specific challenges and novel applications rely on precise geometric calibration [23]. Finally, as a wealth of photogrammetric CV software capable of performing geometric calibration now exists, a comparative analysis of popular commercial software benchmarked against dedicated metrology software may indicate the level of precision, control, and information being sacrificed in order to achieve greater procedural automation [1,35].

1.1. Research Aims

This paper aims to comparatively evaluate TIR geometric calibration approaches for the determination of IO and RO parameters. Moreover, the performance of calibration approaches applied to two IRT-3DDF methods is assessed to determine the efficacy of calibration processes for modelling architectural and cultural heritage. The research reported in this paper will accomplish the following:
1.
Compare 2D board and 3D field calibration processes, detailing production, observation, and calculation of IO and RO parameters using three commercial software packages: MathWorks’ MATLAB, Agisoft Metashape, and Photometrix’s Australis;
2.
Assess the accuracy of derived IO parameters applied to IRT-3DDF using a combined TIR-RGB bundle block adjustment for a historic building façade (Method 1);
3.
Assess the accuracy of derived RO parameters applied to IRT-3DDF using a relative pose implementation on medieval frescoes (Method 2).

1.2. Paper Structure

This paper is structured as follows: Section 2 outlines the methodology for 2D board and 3D field calibration targets, comparing the generated IO and RO results from the specified software packages; Section 3 applies these parameters to two IRT-3DDF methods, evaluating their performance using a combined TIR-RGB bundle block adjustment and relative pose implementation; and Section 4 reviews the performance of both 2D board and 3D field calibration processes, providing recommendations for future adoption, refinement, and applications.

2. Geometric Calibration

The following section outlines the processes involved in the geometric calibration of all cameras used in subsequent IRT-3DDF methods. This includes the following: (1) the construction of both the 2D board and 3D field calibration targets; (2) the calculation of the IO parameters for a combined TIR-RGB bundle block adjustment (Method 1); and (3) the determination of RO parameters for a relative pose implementation (Method 2). Importantly, the TIR and RGB cameras are not fixed in RO for Method 1, meaning imagery for the calibration processes and IRT-3DDF method were acquired asynchronously. For Method 2, independent cameras were fixed on a stereo rig to determine the IO and RO parameters for the eventual IRT-3DDF methods.

2.1. Cameras

For the capture of TIR images in both methods, a Workswell® WIRIS Pro UAV camera (Prague, Czech Republic), featuring both thermal (640 × 512 pix) and visible (1920 × 1080 pix) spectrum sensors within the same camera housing, was configured to be used terrestrially (Figure 1). This camera provides thermal–visible image pairs captured simultaneously during each imaging survey. Although the Workswell WIRIS Pro features both an infrared and visible sensor, WWP is used as the designation for the infrared sensor within the camera. To differentiate between visible spectrum images derived from the sensor housed within the Workswell WIRIS Pro, as opposed to images captured from independent digital cameras, the term VIS is adopted for the visible spectrum sensor housed within the Workswell WIRIS Pro thermal camera, whilst the term RGB is used to denote separate digital cameras. For RGB images, two different digital cameras were used for each data fusion method listed in Section 1.1, including a Sony (Tokyo, Tokyo) α 7R II (7952 × 5304 pix) (RGB1) and a Nikon D750 (Tokyo, Tokyo) (6016 × 4016 pix) (RGB2). Specifications for each camera are listed in Table 1.

2.2. Calibration Targets

Where calibration methods for IRT-3DDF differ from other TIR approaches is that targets must be observable in both infrared and visible spectral ranges, allowing common points to be identified across image modalities for the estimation of RO parameters. Therefore, construction and observation processes must appreciate the duality of the required geometric calibration, creating targets capable of calibrating multiple sensors precisely and simultaneously. Although 2D calibration boards have become the established standard for IRT-3DDF, forming the comparative basis within this study, this work aims to introduce a novel, low-cost, and multi-purpose 3D field that is precise in manufacturing and construction, easy to assemble and configure, and provides clear markers in both image modalities.
For both calibration targets, the ASTM standard ‘Practice for Measuring and Compensating for Emissivity Using Infrared Imaging Radiometers’ (ASTM E1933-14) was used to determine emissivity values of each target material [36]. Here, ASTM E1933-14’s non-contact thermometer method (NTM) was modified to utilise ThermoLab (Version 0.9.4), Workswell’s freely available post-processing software allowing emissivity values to be derived retroactively. For each calibration target, eight perpendicular images were captured at varying temperatures, with each image constituting four points where a surface-modifying material was adhered (TESTO SE & Co., Alton, UK, emission tape) ( ε = 0.95 ). The emissivity values for both 2D board and 3D field targets can be seen in Table 2. For all 2D and 3D approaches, a temperature-controlled water bath was utilised to submerge the 2D board and 3D field markers to excite thermal contrast. Components were submerged in the water bath for 1 min 30 s at 40 °C prior to positioning, constituting a passive thermography approach once image capture was undertaken [2,13].

2.2.1. 2D Board

For the 2D calibration target, a checker-board was designed using DiBond®, an aluminium composite material, and an industrial signage UV printer [2] (Figure 2). DiBond, comprising two aluminium sheets sandwiching a polymer core, provides several benefits over previously listed materials as it is (1) perfectly planar; (2) a good heat conductor; (3) inexpensive; (4) available in several material finishes; (5) able to be cut to any size; (6) appropriate for high-resolution UV printing (i.e., <600 dpi). An A3 piece of silver, brushed-finish DiBond was utilised, printed on with black ink using two passes of a SwissQ, city: Kriessern, country: Switzerland Nyala 4 UV printer (1350 dpi, 53.1 dots/mm). As both checker-board [13] and circle-board [2] patterns have been deemed suitable for high-precision calibration targets, a 5 × 8 ( 35 × 35 mm) checker-board pattern was printed.

2.2.2. 3D Field

For the 3D calibration targets, two separate multi-level test fields were configured using custom markers. Importantly, the novelty of the 3D test fields, designed to be low-cost, easy-to-manufacture, and highly configurable, is that the multi-purpose markers solve several open challenges for IRT-3DDF, notably (1) their use as configurable calibration markers for lab- and field-based geometric calibration procedures, (2) their use as survey markers for photogrammetric scaling, referencing, and validation, and (3) their use as ground truth for radiometric calibration, correction, and verification. Each test field featured markers of 60 mm × 60 mm constructed from milled aluminium and rubber squares fixed to a polymer baseplate. These provide a crosshair generated by materials with different colours and emissivity values (Table 2), clearly identifiable in both image modalities. A total of 16 markers provided a suitable configuration that covered both image formats and could be positioned after active heating. Figure 3 shows the two 3D calibration targets developed for the determination of IO and RO parameters in Method 1 and Method 2, respectively. For the metric evaluation of the IO values in Method 1, each marker was measured using a Leica TS10 total station (angular accuracy: 0.1 mgon; distance accuracy (non-prism): 2 mm + 2 ppm), although 3D world coordinates were not included within the self-calibrating bundle adjustment. For Method 2, a series of scale bars were measured between markers to provide scaling and reference for the relative pose estimation.

2.3. Interior Orientation

For the geometric calibration of TIR cameras, the WWP can be defined as a standard photogrammetric camera following the Brown model [37], merely observing a different spectral range (7.5–13.5 µm). However, as the configuration of the WWP’s germanium lens and solid-state focal plane array (FPA) sensor is optimised for radiometry as opposed to geometry [7], the determination of accurate distortion parameters is paramount. Therefore, to test the importance of the distortion coefficients, parameters were determined for each calibration target with either two or three radial distortion coefficients ( k 1 , k 2 , k 3 ), in addition to tangential distortion coefficients ( p 1 , p 2 ). After initial investigations, the FPA sensor was deemed perpendicular, meaning no differential scaling ( b 1 ) or non-orthogonality ( b 2 ) values were included as required calibration parameters [33]. Finally, the optics of the WWP were fixed during manufacturing, with the camera focussed at infinity and unchanged during each calibration procedure [38]. As an appropriate camera network geometry is fundamental for effective calibration using both 2D and 3D calibration targets [39], images were taken with high convergence, orthogonal roll (+/−90°), varying depths along the principal axis, and the entirety of the target within the image format [1] (Figure 4).
For the calculation of IO parameters, three separate commercial software packages were employed for geometric calibration. Firstly, MathWorks’ MATLAB, R2024b (v24.2.0.2773142) a popular software for CV camera calibration, was used for the geometric calibration using the 2D board. Secondly, for the 3D field calibration target, Agisoft Metashape Pro (v2.1.3) an industry-standard software widely adopted for Structure-from-Motion (SfM) and Multi-View-Stereo (MVS) workflows, was used to assess the feasibility of the proposed methods to work with commonly employed photogrammetric software. Finally, to assess the validity of proposed 3D field methods for high-precision metrological applications, Photometrix’s Australis (v8.555) was also used to undertake calibration processes. Although both 3D field software are capable of undertaking calibration using marker-less feature-based self-calibration [40], only 2D marker coordinates were used within the calibration pipelines. For both software, the same 2D image coordinates, 3D world coordinates, and accompanying constraints were used for implementation. The 2D image coordinates of the test field markers were user-defined in Agisoft Metashape and weighted at 0.33 pix. Similarly, 3D marker coordinates, used solely for metric evaluation, were measured with a Leica, Heerbrugg, Switzerland TS10 total station, with a weight of 2 mm applied to account for propagated error values.

2.3.1. 2D Board (MATLAB)

For the 2D board, MATLAB’s Camera Calibrator app (2024b) was used for the calibration of the WWP, using Zhang’s algorithm [41]. This method involves the capture of a 2D planar target from which detected points can be compared to known metric pattern dimensions. Geometric calibration is achieved through a closed-form solution, followed by a non-linear refinement based on the maximum likelihood criterion. Of the 42 images captured for calibration, 2 were discarded due to partial pattern recognition not meeting a nominal minimum threshold (MinCornerMetric = 0.3), with the remaining 40 being used to calculate IO parameters and accompanying errors. The 2D board distortion profiles (Figure 5) and distortion models (Figure 6) can be seen below, with the resulting IO parameters compared in Table 3.

2.3.2. 3D Field (Agisoft Metashape)

As the Metashape GUI does not allow for purely marker-based image orientation, several procedures were implemented to undertake a self-calibrating bundle adjustment prioritising marker 2D coordinates for the 3D field. Firstly, nominal values for the focal length and pixel size were provided, with redundant calibration parameters fixed at zero (i.e., b 1 , b 2 , and extraneous k coefficients). Secondly, images were oriented in a standard photogrammetric workflow, including weighted 2D marker coordinates as part of the initial bundle adjustment. Once all images had been oriented, all tie points were removed (apart from one to allow for optimisation) and cameras were optimised in an additional self-calibrating bundle adjustment. This allowed for the markers to be the predominant information for the bundle adjustment as opposed to ambiguous tie points generated from local features. The 3D field distortion profiles (Figure 5) and distortion models (Figure 6) can be seen below, with the resulting IO parameters compared in Table 3.

2.3.3. 3D Field (Australis)

Finally, Photometrix’s Australis was included to provide a robust metrological solution for calculating the IO parameters of the WWP. Australis’ workflow was implemented with a series of spatial resections and bundle adjustments undertaken to orient all images using the weighted 2D marker coordinates. Once completed, an additional free network self-calibrating bundle adjustment was undertaken to refine the initial IO parameters. Of the 37 images used for calibration, all featured a minimum of eight intersecting rays per marker, with the convergence limit and number of bundle adjustment iterations set at 0.005 and 12, respectively. The results generated by Australis for the 3D field, including distortion profiles (Figure 5), distortion models (Figure 6), and IO values (Table 3) can be seen below.
To assess the performance of each calibration approach, the standard error of IO parameters ( σ ), Mean Reprojection Error (MRE) (pix) in 2D image space, and standard error ( S O ) (mm) of camera poses in 3D object space are used [7]. The results in Table 3 show notable differences between the 2D board and 3D field calibration approaches. Firstly, it must be noted that the scale of the 2D board (A3-sized) and resulting camera networks are significantly smaller then the 3D calibration targets (∼4×), with the derived MRE and S O being heavily scale-dependent. MATLAB calculates a focal length ∼13 µm smaller than the 3D field approaches, which show comparable focal lengths across all four distortion coefficient tests (all within 8 µm). This is likely a product of insufficient camera-to-object distance variation in the 2D board network geometry, with the planar target providing less depth in feature points and individual camera positions [33]. In addition, the higher intrinsic standard errors are largely due to the ambiguous extraction of checker-board corners (MRE = 0.50 pix) propagating into greater IO imprecision. Most significantly, the results generated by the 2D board show significantly different distortion models from that of the 3D field calibrations (Figure 6), the culmination of higher IO standard errors, poor feature point extraction, and the difficulty in filling the image format effectively whilst maintaining preferable network geometry.
In contrast, the observable similarities in the IO values, distortion models, and MRE from both Agisoft Metashape and Australis are expected, utilising the same 2D image coordinates for their respective bundle adjustments and obtaining marginally different radial and tangential profiles (Figure 5). Both software packages determined similar IO values across calibration tests, capturing the convex germanium lens geometry equating to a distortion of ∼0.25 pix on the right of the image format. Australis obtains the best MRE and S O results across both the k 2 and k 3 coefficients. Notably, Australis provides greater control over the self-calibrating bundle adjustment (inc. iterations, convergence limit, and minimum number of rays per marker), which, when coupled with native access to the covariance matrix, allowed for results to be refined for greater IO precision and metric accuracy. However, the validity of these metrics must not be viewed in isolation for the assessment of calibration processes, but taken into account when intrinsics are applied independently in subsequent photogrammetric applications.

2.4. Relative Orientation

The estimation of the relative pose between fixed stereo cameras determines the translation and rotation vectors necessary for mapping one image onto another. In the case of IRT-3DDF, this allows accurate geometries generated from high-quality RGB images to be textured or assigned temperature information from proximal TIR images. For the comparison of 2D board and 3D field calibration targets, two separate baselines were assessed. Firstly, as the Workswell WIRIS Pro thermal camera features both a infrared sensor and a visible sensor within the same camera housing, the baseline between both internal sensors was determined. However, as the resolution of the VIS sensor is sub-optimal for detailed modelling (1920 × 1080 pix), a stereo rig fixing the WWP and RGB2 was assembled to acquire additional data for Method 2. For each calibration set-up, as with the calculation of the IO parameters, baselines were determined with two or three radial distortion coefficients to assess their importance in determining precise and reliable RO parameters for IRT-3DDF (Figure 7).

2.4.1. 2D Board (MATLAB)

MATLAB’s Stereo Camera Calibrator App (2024a) was used to compute the 2D board calibration. Firstly, IO values for the Workswell WIRIS Pro sensors (WWP and VIS) and RGB2 were re-generated using the same processes outlined in Section 2.3.1. These values are subsequently used as fixed values for relative pose estimation. MATLAB’s Stereo Camera Calibrator is limited by its inability to calibrate two cameras of different resolution, meaning that imagery and associated IO values (i.e., f x , f y , c x , and c y ) for VIS and RGB2 had to be resampled to that of the WWP to accommodate stereo calibration (ultimately rescaled upon completion). For the VIS-WWP camera and RGB2-WWP stereo rig, 37 and 20 image pairs were processed, respectively. The estimated relative translation and rotation values for each stereo pair can be seen in Table 4, showing the pose of the Workswell WIRIS Pro thermal sensor with respect to the VIS/RGB2 sensor.

2.4.2. 3D Field (Agisoft Metashape)

For the relative pose estimation using Agisoft Metashape, using a different 3D field than that used for the calculation of the initial camera parameters (Figure 3), IO values for all sensors and cameras were re-calculated using the approach outlined in Section 2.3.2. Derived IO parameters were subsequently used as fixed values for Agisoft Metashape’s Multi-Camera System for the VIS-WWP sensors and RGB2-WWP stereo rig. Much like the IO method, ambiguous tie points were removed after initial orientation, with an additional self-calibrating bundle adjustment undertaken to refine RO values. Results were then scaled using a collection of scale bars allowing error metrics to be derived. For the VIS-WWP camera and RGB2-WWP stereo rig, 37 and 31 image pairs were processed, respectively. Results for both baseline tests can be seen in Table 4, similarly showing the mapping of the WWP onto VIS/RGB2.

2.4.3. 3D Field (Australis)

For the calculation of RO parameters in Australis, the initial IO values for all sensors were obtained through the processes outlined in Section 2.3.3. Subsequently, image pairs were imported into Australis for stereo calibration with the same 2D marker coordinates and constraints as those used in Agisoft Metashape. A self-calibrating bundle adjustment was then undertaken to obtain the extrinsic parameters of the cameras, with scale bars used to scale the scene, derive the RO parameters, and calculate error metrics. Results for both baseline tests in Australis can be seen in Table 4.
From the results in Table 4, an indication of the overall performance of the derived RO parameters can be determined, with the translation (X, Y, Z) and rotation ( ω , ϕ , κ ) of the TIR sensor with respect to the reference sensor (VIS or RGB2). Here, X Y Z are aligned with the image coordinate system (x, y), where X is aligned with x pointing right, Y is aligned with y pointing down, and Z points forwards from the perspective centre. To gain some understanding of the performance of relative pose estimation, the baselines (horizontal displacement X in camera space) between the fixed sensors was utilised. According to Workswell s.r.o, the baseline between the centre of both lenses within the Workswell WIRIS Pro is 40.7 mm. Similarly, the measured baseline between the lenses of the RGB2 and WWP stereo rig was 184 mm, with a depth offset of 10 mm. All results achieve a reasonable estimation for the baseline (X) between the VIS-WWP and RGB2-WWP sensors, with MRE mostly <1 pix and 3D S O < 6 mm for all tests. Expectedly, a greater proportional range is present within the translation values and S O for the VIS-WWP across each software package and distortion coefficients than that of the RGB2-WWP. This is largely due to the shorter baseline and poorer resolution of the VIS sensor with respect to RGB2. However, the RO results for both 3D field tests show markedly similar results for the stereo rig baseline, although the σ values for the Australis captured within the RO results are notably higher than both MATLAB and Agisoft Metashape across all results. Whilst the derived RO parameters and error metrics suggest adequate performance for all calibration processes, their application for IRT-3DDF is needed to determine the value of the calibration processes for specific data fusion and heritage-specific applications.

3. InfraRed Thermography 3D Data Fusion (IRT-3DDF)

The following section outlines the application of the derived IO and RO parameters for the documentation of architectural and cultural heritage using IRT-3DDF. Method 1 executes a combined TIR-RGB bundle block adjustment using the IO parameters obtained in Section 2.3 as fixed camera intrinsics, utilising multi-modal image matching to generate correspondence across image modalities. Method 2, a relative pose implementation, applies the RO parameters from Section 2.4 to derive a textured 3D thermal model of medieval wall paintings. Furthermore, insights into the radiometric calibration, correction, and verification of thermal infrared cameras are highlighted, although the focus of this section remains on geometric calibration and resulting network geometry.

3.1. Method 1: Combined TIR-RGB Bundle Block Adjustment

For the evaluation of the IO parameters obtained in Section 2.3, Lenton Lodge Gatehouse, a 19th century Grade II * listed property in Nottingham, UK, was used as a case study (Figure 8). Built by Jeffry Wyatville and largely comprising Mansfield sandstone, the structure represents one of the finest examples of Elizabethan Revival architecture in England. Currently, Lenton Lodge Gatehouse shows visible signs of material degradation, notably around the parapet, decorative traceried panels, and turrets. Facing East-South-East, thermal images were acquired in January 2024 between 10:30 and 11:00 a.m. using passive thermography, maximising the baking of the structure from solar radiation and minimising shadows on the façade. A total of 20 WWP images were captured in a single strip with a Ground Sample Distance (GSD) of 9.81 mm, whilst 25 RGB1 images were captured in two strips with GSDs of 1.03 mm and 1.29 mm. Importantly, imagery from the WWP and RGB1 were captured independently with no fixed relative orientation, meaning no consistent baseline between the cameras can be assumed.
Additionally, a network of control points, utilising the custom markers from the 3D calibration test fields and additional natural points, were surveyed using a Leica, Heerbrugg, country: Switzerland TS10 total station to provide four ground control points (GCPs) and four check points (CPs) for subsequent referencing, scaling, and metric evaluation. The WWP was radiometrically calibrated by Workswell s.r.o prior to the execution of the imaging survey, in line with recommendations for routine servicing provided by the manufacturer. Preliminary lab experiments determined a 45 min acclimatisation time was needed for the IR sensor and CPU to generate stable temperature readings, in line with similar findings using the WWP [42]. Furthermore, non-uniformity corrections (NUCs) were undertaken at 3 min intervals during the survey to maintain consistent temperature values [43]. Finally, thermometer readings on the custom markers were taken as ground truth for quantitative temperature verification, comparing known and derived 3D thermal model temperature values.
With the proliferation of deep-learning approaches for both mono- [44] and multi-modal [45] image matching, the possibility of using ‘out-the-box’ neural networks for photogrammetric thermal modelling is becoming a reality for IRT-3DDF [46]. Therefore, to explore multi-modal image matching between TIR and RGB images, Deep-Image-Matching (DIM), an open-source Python library, Deep-Image-Matching (DIM) (v1.0.0) designed for robust multi-view image matching, was utilised [47]. DIM provides a modular workflow to configure and customise both hand-crafted and deep-learning feature matchers, allowing correspondence to be exported for seamless integration into various photogrammetric software. To facilitate effective multi-modal matching, WWP images underwent a histogram normalisation and UnSharp Masking to enhance detectable features. Additionally, RGB images were downsampled from their original resolution (7952 × 5304 pix) to 1 8 th size (994 × 663 pix) to more closely mirror the resolution of the WWP images. For all WWP and RGB images, 8000 keypoints were detected using a SuperPoint [48] feature extractor and LightGlue feature matcher [49] (Sp + Lg) (Figure 9). Utilising DIM, all 45 images were matched exhaustively, resulting in a possible 990 image pairs across the dataset.
For the combined TIR-RGB bundle block adjustment, the generated database from DIM was imported into COLMAP for image orientation [50]. Here, each block of WWP and RGB1 images were assigned a distinct camera model, using the IO parameters from the TIR geometric calibration in Section 2.3. To keep the focus of the experiments on the performance of the TIR geometric calibration (independent variable), intrinsics for RGB1 were calculated through self-calibration using the 25 images of the Lenton Lodge Gatehouse and held constant for each calibration subset. Importantly, IO values for the RGB1 focal length and principal point were divided by eight to match the resolution of the downsampled images provided to DIM. For each calibration set-up, an independent bundle adjustment was undertaken, fixing the IO values for both the WWP and RGB1 to assess the accuracy of the geometric calibration. Finally, to determine the possibility of providing no initial IO values for a combined TIR-RGB bundle block adjustment, a self-calibration approach using two radial distortion coefficients was undertaken wherein camera parameters were refined during the bundle adjustment. For all cases, derived sparse 3D tie points and camera poses were imported into Agisoft Metashape for metric evaluation (Figure 10).
Results for the combined TIR-RGB bundle block adjustment are listed in Table 5, separating RGB1 and WWP image blocks into their respective modalities for metric evaluation. For the WWP blocks, Root-Mean Squared Error (RMSE) on the CPs is separated into horizontal (X), depth (Y), and vertical (Z) components, with the RMSEWWP TOT representing the total RMSE of results in 3D space. Additionally, the RMSE of the RGB blocks (RMSERGB1) and the MRE of the WWP blocks (MREWWP) are included for a comprehensive assessment of the combined TIR-RGB bundle block adjustment and camera intrinsics. Firstly, the results of the WWP image blocks show notable differences between calibration approaches, with the RMSEWWP TOT obtained from both 2D board calibration tests roughly twice that of the 3D field results. Calculating a focal length ranging from 13 to 16 µm less than that obtained from the 3D field methods (Table 3), coupled with an inability to capture the unique distortion of the WWP’s germanium lens (Figure 6), the 2D board shows significant error in the depth (Y) component of the RMSEWWP. Consequently, the results obtained from the the 2D board calibration targets are similar to that of COLMAP where no prior calibration is undertaken and nominal IO values are corrected via ‘on-the-job’ self-calibration. The poorer MREWWP and RMSEWWP of the 2D board using MATLAB is likely due to the scale of the calibration network geometry being insufficient for upscaling to the 3D scene and the imprecision within the calibration rendering IO values that are heavily scene-dependent [22].
In contrast, the results of the 3D fields using both Agisoft Metashape and Australis show positive results, obtaining similar MREWWP values for the combined TIR-RGB bundle block adjustment, and all RMSEWWP TOT values are in line with the GSD of the WWP block (GSD = 9.81 mm). Agisoft Metashape achieved the best RMSEWWP TOT results with two radial distortion coefficients, with Australis achieving comparable results with the best RMSERBG1 when three radial distortion coefficients were used. Interestingly, the MREWWP and RMSEWWP results obtained using either two or three radial distortion coefficients show little discernible difference, which, when combined with the similar distortion models and profiles (Figure 5 and Figure 6), suggests that for germanium lenses where no complex distortion is present, either two or three distortion coefficients could be used for effective TIR geometric calibration.
Furthermore, as each combined TIR-RGB bundle block adjustment featured the same RGB1 IO values and multi-modal correspondence, all RGB blocks present similar RMSERGB1 values of <3 mm, with mono-modal matches (those between RGB images) from Sp + Lg allowing for successful orientation of the downsampled RGB block. However, a strong correlation is also apparent between the WWP (RMSEWWP TOT) and RGB1 (RMSERGB1) image blocks (r = 0.86), suggesting the influence of the WWP (by virtue of the fixed IO values) on multi-modal image orientation. Similarly, the MREWWP across all tests shows notable correlation, closely mirroring the RMSERGB1 and RMSEWWP values. This reiterates the suitability for image and object space metrics to be utilised in tandem for the assessment of the combined TIR-RGB bundle block adjustment and camera intrinsics [51].
The results obtained with COLMAP, undertaking simultaneous IO and EO via ‘on-the-job’ self-calibration, proves the benefit that can be provided by effective geometric calibration for sensor fusion pipelines [34]. Given the nominal focal length and image centre as starting IO values, the results of the combined TIR-RGB bundle block adjustment perform worse than if successful pre-calibration is undertaken, suggesting the benefit that can be gained from calibration using multi-modal correspondence. Notably, RMSEWWP TOT were the same for when nominal values ( f x , f y , u o , and v o ) or full IO values from the 3D field ( f x , f y , c x , c x , k 1 , k 2 , p 1 , and p 2 ) were refined, with the bundle adjustment deriving the same IO values for both the WWP and RGB1 regardless of the initial starting point. The resulting 3D models, generated from the RGB1 images and textured with WWP thermal images, can be seen in Figure 11.

3.2. Method 2: Relative Pose Implementation

For the evaluation of the RO parameters, Castello Di Arco, located in the Sarca Valley of Trentino, northern Italy, was chosen as a case study. The 14th century medieval frescoes, depicting aristocracy and court life, have undergone several phases of investigation and restoration, largely due to water ingress and plaster delamination caused by the castle’s rock wall foundations (Figure 12). In a bid to reduce this phenomena, a space heater has been utilised to actively heat the façades after heavy rain and reduce the impact of the moisture on the frescoes. Whilst active IRT provides established non-destructive testing (NDT) techniques for cultural heritage wall paintings [52], to determine the extent of the deterioration and assess the current conservation practices, IRT-3DDF was employed to localise visible signs of delamination and discolouration with quantitative temperature differences. Therefore, to assess the impact of current conservation practices, the imaging survey was undertaken after active thermography, with the space heater removed just prior to data acquisition. The survey was undertaken in October 2024 after successive days of dry weather (avg. temp. 13 °C). The fresco corner, facing North-East, was subject to solar radiation from 06:52 (sunrise) until the commencement of the survey (10:45–11:30 a.m.).
To utilise the findings generated in Section 2.4, stereo results were used from both the in-built thermal camera and purpose-built stereo rig baselines. For both baselines (VIS-WWP and RGB2-WWP), 40 stereo pairs were captured of the fresco corner and adjoining walls (Figure 12). Captured at a distance ∼2.5 m away from the frescoes, the GSD for the WWP, VIS, and RGB2 sensors were 3.3 mm, 2.0 mm, and 0.3 mm, respectively. As previously stated, the benefit of utilising known RO parameters is essential in cases where no discernible thermal contrast can be identified or generated through active thermography. For Castello Di Arco, presenting a temperature range of ∼3 °C, the derived thermal images presented low contrast and high noise, meaning few suitable correspondences could be used for feature-based matching (Figure 13). Initial tests proved the inability for WWP images to be oriented mono-modally, necessitating RO parameters with a reference camera/sensor for effective modelling. To mitigate these challenges, a network of the 3D field custom markers, actively heated and positioned in front of the walls (Figure 12), provided manual tie points to aid in the orientation of the WWP block and generate multi-modal correspondence [10]. To generate control for the resulting models, measured scale bars were taken between positioned markers, with two used as control and two reserved for validation. To undertake the modelling, Agisoft Metashape’s Multi-Camera System was utilised, with VIS/RGB2 set as the ‘primary’ camera (geometry) and the WWP set as the ‘secondary’ camera (texture). IO values were fixed from the initial calibration process (Section 2.4), with the X, Y, and Z translations and ω , ϕ , and κ rotations input as initial constraints for the bundle adjustment. In addition, 2D image coordinates (custom markers and tie points) were both weighted at 1.0 pix and 1.5 pix for the RGB2-WWP and VIS-WWP baselines, respectively.
Table 6 shows the results from each baseline calibration test specifying the distortion coefficients used for the WWP during calibration ( k 2 or k 3 ), with the generated textured frescoes visible in Figure 14. The RMSE on the measured scale bars are separated for each modality to provide an assessment of the image orientation in each imaging domain. Additionally, as with the results reported in Section 3.1, a self-calibration process has been undertaken where no initial IO or RO parameters were provided, with IO, RO, and EO parameters obtained simultaneously via a self-calibrating bundle adjustment. For the VIS-WWP baseline, the success of the obtained RO parameters is evident, with all approaches able to orient the 40 stereo pairs successfully. For the sensors within the Workswell WIRIS Pro, calibration results are comparable across all calibration targets, obtaining similar RMSEVIS-WWP results for the reference VIS sensor ranging from 1.5 to 2.7 mm. Interestingly, the 2D board obtained the best RMSEVIS-WWP value amongst VIS blocks, showing the suitability of 2D boards for standard CV geometric calibration. Importantly, all results (apart from MATLAB) provide a better RMSEVIS-WWP for the WWP image blocks than that of pure self-calibration, suggesting the superiority of pre-calibration for sensor systems with short baselines.
For the stereo rig (RGB2-WWP), all approaches achieve a similar RMSERGB2-WWP for the RGB2 image blocks, showing the suitability of each calibration approach (across different distortion coefficients) to obtain camera intrinsics for traditional digital cameras. However, notable differences for the 2D board calibration target using MATLAB are identifiable, with the IO and RO parameters leading to a markedly higher RMSERGB2-WWP for the WWP block of ∼9 mm. This is likely due to the different Field-of-View (FOV) between the fixed cameras during RO calibration, with the different focal lengths of the sensors (50 mm vs. 13 mm) making it more difficult to fill both image formats with the 2D board during simultaneous image capture [32]. However, we see comparable results from both 3D field calibration approaches, obtaining an RMSERGB2-WWP of <2 mm for the WWP image block. Significantly, Australis was able to obtain the lowest MRE, S O , and RMSE for the WWP blocks in RO calibration methods across both distortion coefficient subsets. Much like the findings of Section 3.1, the results from both 3D calibration methods demonstrate the superiority of the 3D field (and accompanying software) to obtain precise and reliable IO values that are image invariant, scene-independent, and metrically accurate for photogrammetric reconstruction.
As self-calibration was used as a benchmark to determine the ability for IO, RO, and EO parameters to be obtained simultaneously in the same bundle adjustment, it is important to understand the performance of these results when compared to the RO pre-calibration processes. Firstly, the block RMSE values from the VIS-WWP baseline were favourable, establishing intrinsics in line with pre-calibrated values and an estimated baseline (X) of −41.2 mm (∼0.5 mm from the reference value). However, the results of the RGB2-WWP baseline, obtaining a better RMSERGB2-WWP for the WWP block than any calibration approach, generates a markedly different distortion model for the WWP (notably c x , c y and p 1 , p 2 ) and translation values (X = −187.8 mm; Z = −0.8 mm). This suggests the susceptibility for ‘on-the-job’ self-calibration to find any scene-dependent solution for IO and RO parameters that fit with the extracted tie points and included markers, especially when fewer than 200 tie points exist in the WWP block. This reiterates the importance of known IO and RO parameters for scenes where self-calibration is more impressionable, with parameters (and logical constraints) a necessary feature when insufficient correspondence is present or when modelled sensors present unique distortion profiles.

4. Discussion

4.1. TIR Geometric Calibration

4.1.1. Calibration Targets

As historical methods of TIR geometric calibration have been overly reliant on the manual construction of 2D boards, the ambition of this work was to determine a novel 3D test field design, preferable for photogrammetric camera calibration, which could be compared against established contemporary approaches. Critically, the success of all TIR geometric calibration approaches is in the ability for control points to be extracted with high contrast and low noise, with the greatest contributing factor being the emissivity of the materials used. As seen in Table 2, the difference in emissivity between the DiBond and UV-printing ink was notably small ( Δ ε = 0.20 ). This made it difficult to excite thermal contrast and extract accurate checker-board corners, resulting in uncertainty propagating into imprecise IO values and larger error values (i.e., IO σ , MRE, and S O ) (Table 3). Although scale, network geometry, and calibration set-up still play a role, material emissivity is likely the largest contributing factor for the poor calibration and IRT-3DDF results obtained with the 2D board, demonstrating the importance of material choice and emissivity calculation as part of effective TIR geometric calibration.
In contrast, the emissivity difference for the 3D test field provided well-defined crosshairs with high contrast throughout the entirety of the imaging surveys ( Δ ε = 0.65 ). Highly-configurable in terms of quantity, geometry, scale, and setting, the novelty of the proposed custom markers is characterised by their suitability as (1) crosshair targets for 3D test fields; (2) additional tie points for photogrammetric multi-modal image matching; (3) GCPs and CPs for photogrammetric evaluation; and (4) known points for quantitative temperature verification. However, the proposed method is still limited by the manual identification of markers within the imaging surveys as opposed to fully automatic feature detection. Roshan et al.’s [15] 2D ChArUco checker-board provides a logical progression of the current thermal markers, with milled aluminium coded markers as a novel approach for both future calibration and survey applications.
However, although coded markers [15]; distinctive shapes [53]; speckles [54]; actively heated elements [9]; and mono- or multi-modal local features [40] may provide avenues for future exploration, the ability for these to be observed by commercial thermal cameras with the inherently low resolution (typically 180 × 120–640 × 480 pix), common for three-dimensional thermal imaging systems (TTISs) incorporating multiple image- and range-based sensors [55], remains an open challenge. Finally, whilst the use of active thermography provided the necessary contrast for effective control point identification in both calibration targets, the considerable limitations of existing and alternate active thermography approaches include the following: the difficulty in constructing electrical targets, the inaccuracy of control point extraction, the inability to take devices into the field, and the need for repeated heating cycles [14,54].

4.1.2. Interior Orientation

The derived IO parameters, obtained by comparing calibration targets, the number of distortion coefficients, and software packages, point to several considerations for TIR geometric calibration. As previously highlighted, the quality of the 2D board drastically affects the derived intrinsics, incapable of capturing the distortion of the germanium lens and producing significantly larger IO σ , MRE, and S O values (Table 3). In addition, whilst an initial set of 20 images were captured for interior orientation using MATLAB, an additional 20 (totalling 40 images) were needed to refine IO values within expected values (as seen in Dlesk et al. [23]). Although MRE and S O errors obtained during calibration would indicate agreeable performance of the 2D board, these metrics cannot be viewed in isolation, with the comparison and application of IO values in Section 3.1 demonstrating the inability for the 2D board to achieve precise TIR camera intrinsics for scene-independent photogrammetric use. As a result, even though 2D boards are established as a method for TIR geometric calibration, the susceptibility of projective coupling from sub-optimal calibration network geometry, the tendency for derived IO values to be image-variant, and the observed scene-dependency of calibration processes means that the use of 2D boards should be limited for metrically accurate IRT-3DDF.
In contrast, the 3D field, presenting a strong configuration of multi-level points, was able to capture the unique distortion of the WWP optics with either two or three radial distortion coefficients. The acknowledged benefits of a 3D calibration field, including the ability for the image format to be properly filled, the greater scale and depth information present, and the high convergence between point observations, allowed for more precise and reliable estimates of IO values. The comparable performance of both Agisoft Metashape and Australis software packages, both in terms of distortion profiles (Figure 5) and distortion models (Figure 6), is likely due to the identical camera network geometry, 2D marker coordinates, and associated constraints. However, for these results to be achieved, the inability for the Agisoft Metashape GUI to run a purely marker-based bundle adjustment required the removal of ambiguous tie points in the ‘textureless’ calibration target after initial alignment and for an additional optimising bundle adjustment to be undertaken. Therefore, when deciding between software packages, considerations between automation and accuracy must be made in accordance with the camera network for TIR geometric calibration, with Australis allowing for greater access to the self-calibrating bundle adjustment and accompanying parameter correlations (Australis was the only software to provide access to both IO and IO/EO covariance matrices for the assessment of intra- and inter-image projective coupling). Finally, regarding the choice of distortion coefficients or software package, there is little in the IO or RO calibration performance metrics or IRT-3DDF results that would suggest a preference for two or three radial distortion coefficients; however, this will largely be TIR camera- and calibration dependent.
For the self-calibration results in Table 5, the influence that imprecise intrinsics can have for multi-modal image matching is apparent. The obtained principal point for the self-calibration presented a difference from calibrated IO values in c x and c y of ∼10 pix and 4 pix, respectively; which, even though capable of determining a similar distortion profile to the 3D field approaches, resulted in expectedly higher RMSETIR TOT [56]. As discussed in Section 3.1, the resulting IO values from the combined TIR-RGB bundle block adjustment self-calibration were practically the same when either nominal- (simple pinhole camera model), full- (OpenCV camera model), or pre-calibrated IO values were refined [50]. This demonstrates a considerable opportunity to understand the influence of bundle iterations and initial intrinsics on multi-modal image orientation. Therefore, the exploration of on-site self-calibration [57], utilising varying initial camera intrinsics and constraints, is a logical progression for Section 3.1 [34].
Finally, the difference in TIR intrinsics obtained for Section 2.3 and Section 2.4 must be noted, stressing the need for TIR cameras to be regularly calibrated prior to photogrammetric use (IO and RO calibration experiments were undertaken six months apart). For example, an observable offset in the principal point and tangential distortion coefficients was present for WWP IO values used in the Section 3.1 and Section 3.2. This phenomena could be caused by mechanical stress or component instability within the Workswell WIRIS Pro, a possible side effect of rotational movements during calibration or camera handling and transportation [7]. Furthermore, the influence of sensor temperature and ambient conditions during geometric calibration can be seen to have an influence on the derived TIR images, causing an observable increase in temperature values within the obtained calibration datasets (Figure 3). As for the sensor, camera housing and environmental temperatures have an established impact on resulting camera intrinsics [58] and radiometric accuracy [42]; the analysis of this effect on FPA sensors (detecting the very radiation that will be causing instability) warrants future investigation.

4.1.3. Relative Orientation

The results of the relative pose estimations in Section 2.4 and Section 3.2 demonstrate the success of the proposed calibration approaches and the importance of pre-calibrated IO and RO parameters for IRT-3DDF. Firstly, the results for the 2D board provided an agreeable estimation of the relative pose for both baselines, notably for the VIS-WWP baseline where the focal lengths and FOVs of the VIS-WWP image pairs made filling the image format more achievable. However, the results from both 3D field calibration approaches demonstrate the superiority of a strong camera network and multi-level calibration target for TIR relative pose estimation, with both Agisoft Metashape and Australis achieved successful results when IO and RO parameters were applied (notably for the RGB2-WWP stereo rig <2 mm for all image blocks). Importantly, both Section 2.4 and Section 3.2 show the capability for RO parameters to be generated across varying baselines and sensors, critical for implementation within IRT-3DDF where a vast array of resolutions and TTISs is present.
The importance of pre-calibration in Section 3.2 is evident, where self-calibration was highly unsuccessful in obtaining simultaneous IO, RO, and EO parameters in a low-contrast setting. Even though IO values for VIS and RGB2 were in line with pre-calibrated values, both VIS-WWP and RGB2-WWP results failed to provide reliable IO values for the WWP, with VIS-WWP results determining a focal length ∼100 µm from pre-calibrated values and RGB2-WWP results achieving a principal point (282, 243) significantly off from the image centre (320, 256). Clearly, Agisoft Metashape’s Multi-Camera System calculates IO values for the ‘primary’ camera as a priority, fitting the ‘secondary’ camera within the remaining constraints. In addition, altering the weights given to WWP tie points and user-defined coordinates of the custom markers, imperative for generating multi-modal correspondence, provided notable differences in obtained IO and RO parameters, likely due to the extremely low number of tie points in the WWP image block (No.WWP < 200). Therefore, additional investigations assessing the stability of relative pose estimations for IRT-3DDF via ‘on-the-job’ self-calibration are needed, notably the inclusion of the following: (1) initial estimates and constraints for IO parameters; (2) initial estimates and constraints for RO parameters (most likely including horizontal displacement in X) [31]; and (3) greater TIR mono- and TIR-RGB multi-modal tie points (in a similar vein to Section 3.1).

4.2. InfraRed Thermography 3D-Data Fusion (IRT-3DDF)

Whilst the use of IRT as part of a broader catalogue of NDT techniques has long been established for architectural heritage [59], the presented examples of IRT-3DDF reveal opportunities where 2D imaging techniques can embrace synergies with 3D modelling practices [55]. The ability to generate rapid, real-time, and repeatable thermal models creates opportunities for the assessment of architectural heritage through the comparison of 3D thermal models over time, the localisation of thermal anomalies, and the quantification of material degradation. Similarly, the results from Castello di Arco, with known and observable discrepancies in the temperature of the medieval frescoes, point to possible interventions and areas for investigation. Firstly, it is evident that the use of a space heater to alleviate the effects of the water seeping through the wall has been insufficient to address a pressing issue in the upper corner, with a temperature difference of ∼1.5 °C from the fresco average still apparent. The use of IRT-3DDF helps demonstrate the extent of the temperature difference and the proximity of affected areas in relation to the space heater and identifies additional areas on the wall where conservation work should be considered. However, whilst the focus of this work has been on geometric calibration, for additional insights to be extracted, future emphasis on radiometric calibration, correction, and validation is needed for the implementation of effective IRT-3DDF [60].
The results from Section 3.1 also demonstrate the success of a novel deep-learning-based multi-modal image matching method for thermal modelling, representing a real opportunity for the development of IRT-3DDF by exploiting ‘out-of-the-box’ neural networks not trained on TIR images. This demonstrates the following: (1) the ability for feature correspondence to be determined across image modalities, (2) the strength of these points for simultaneous image orientation as part of a combined TIR-RGB bundle block adjustment, and (3) the increased automation of IRT-3DDF without a reliance on known survey points or manual registration. Similarly, the results in Section 3.2 represent a critical method for IRT-3DDF where insufficient temperature differences means little thermal contrast is visible within a given scene. The ability for scenes to exploit proximal sensors and reliable calibration procedures represents a critical opportunity for indoor thermal modelling, where the generation of contrast is a likely obstacle. However, the manoeuvrability and construction of the RGB2-WWP stereo rig provides considerable limitations in terms of data capture and synchronisation, likely exacerbated when the scene increases in scale and complexity. The findings from Section 3 demonstrate the benefit of using a high-resolution digital camera to generate the ‘base’ geometry for IRT-3DDF, with most commercial thermal cameras (and accompanying visible sensors) developed for building inspection as opposed to photogrammetry. Therefore, the logical evolution of this research should focus on combining the complementary methods of Section 3 for the establishment of TIR geometric calibration as a fundamental component of IRT-3DDF. This would further develop the novelty of multi-modal correspondence for orienting TIR and RGB images whilst utilising the proximal sensors within thermal cameras, helping mitigate environments where both mono- and multi-modal TIR correspondence may not be determined.

5. Conclusions

The determination of reliable IO and RO parameters for TIR cameras is critical for their subsequent use in photogrammetric processes. The performance of calibration procedures cannot be used as a proxy for their success in eventual 3D reconstruction, notably for data fusion methods utilising sensors of varying resolutions and modality. This work has comparatively evaluated existing 2D- and novel 3D calibration targets, detailing construction, calibration, and applications for IRT-3DDF. Results suggest the superiority of a 3D field for the estimation of TIR intrinsics and provide insights regarding appropriate camera parameters, software packages and and future refinements. In addition, the estimation of the relative pose between two baselines was determined to assess the accuracy of image orientation utilising in-built sensors from a singular thermal camera or a constructed stereo rig. The application of IO and RO parameters to two IRT-3DDF methods demonstrates the opportunities to be gained for NDT within architectural heritage exploiting thermal modelling. The first method, proposing a novel combined TIR-RGB bundle block adjustment utilising deep-learning multi-modal correspondence, demonstrates the capability of using neural networks generalised outside of their training domain. This presents an automatic data fusion method where simultaneous image orientation can be achieved without a reliance on GCPs or manual registration. The second method, utilising a fixed stereo rig for the generation of a textured 3D thermal model, provides a workflow demonstrating the importance of proximal sensors and effective RO parameters when low thermal contrast and few TIR correspondence can be determined. Future work will look to combine developed methods to exploit the resolution and flexibility of commercial digital cameras whilst utilising the fixed sensors within a thermal camera for instances when TIR images require additional support for successful image orientation.

Author Contributions

Conceptualization, N.S., and F.R.; data curation, N.S., and F.R.; formal analysis, N.S., and G.P.; funding acquisition, S.M., and J.M.; methodology, N.S., and G.P.; project administration, N.S., S.M., and J.M.; resources, S.M., F.R., and J.M.; supervision, S.M., F.R., P.B., and J.M.; validation, N.S.; visualisation, N.S.; writing—original draft, N.S.; writing—review and editing, N.S., S.M., F.R., G.P., P.B., and J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the UK Research and Innovation (UKRI) Engineering and Physical Sciences Research Council (EPSRC) as part of the Geospatial Systems Centre of Doctoral Training (CDT), grant number: EP/S023577/1, https://gtr.ukri.org/projects?ref=EP%2FS023577%2F1 (accessed on 22 May 2025). Sensor equipment used in the research was funded by UKCRIC—UK Collaboratorium for Research in Infrastructure & Cities: Newcastle Laboratories, grant number: EP/R010102/1, https://gtr.ukri.org/projects?ref=EP%2FS023577%2F1 (accessed on 26 January 2025).

Data Availability Statement

The original data presented in this study is openly available in the Thermitage GitHub repository at https://github.com/nneilsutherland/Thermitage (accessed on 1 July 2025).

Acknowledgments

We would like to extend our thanks to Roberto Zampiccoli (Arco municipality) and the colleagues at the Castello Di Arco for their kindness, hospitality, and support during the image acquisitions. We would also like to thank Jon Bedford and Historic England’s Geospatial Survey Team for lending the Sony α 7R II used in the Lenton Lodge Gatehouse surveys.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of this study; in the collection, analyses, or interpretation of data; in the writing of this manuscript; or in the decision to publish the results.

References

  1. Remondino, F.; Fraser, C. Digital Camera Calibration Methods: Considerations and Comparisons. In Proceedings of the ISPRS Commission V Symposium ‘Image Engineering and Vision Metrology’, Dresden, Germany, 25–27 September 2006; pp. 266–272. [Google Scholar] [CrossRef]
  2. Usamentiaga, R.; Garcia, D.; Ibarra-Castanedo, C.; Maldague, X. Highly Accurate Geometric Calibration for Infrared Cameras using Inexpensive Calibration Targets. Measurement 2017, 112, 105–116. [Google Scholar] [CrossRef]
  3. Fryskowska-Skibniewska, A.; Delis, P.; Kedzierski, M.; Matusiak, D. The Conception of Test Fields for Fast Geometric Calibration of the FLIR VUE PRO Thermal Camera for Low-Cost UAV Applications. Sensors 2022, 22, 2468. [Google Scholar] [CrossRef] [PubMed]
  4. St-Laurent, L.; Mikhnevich, M.; Bubel, A.; Prévost, D. Passive Calibration Board for Alignment of VIS-NIR, SWIR and LWIR Images. In Proceedings of the 13th International Conference on Quantitative InfraRed Thermography (QIRT 2016), Gdansk, Poland, 4–8 July 2016; pp. 777–784. [Google Scholar] [CrossRef]
  5. Swamidoss, I.N.; Bin Amro, A.; Sayadi, S. Systematic Approach for Thermal Imaging Camera Calibration for Machine Vision Applications. Optik 2021, 247, 168039. [Google Scholar] [CrossRef]
  6. Kubota, Y.; Ke, Y.; Hayakawa, T.; Moko, Y.; Ishikawa, M. Optimal Material Search for Infrared Markers under Non-Heating and Heating Conditions. Sensors 2021, 21, 6527. [Google Scholar] [CrossRef]
  7. Luhmann, T.; Fraser, C.; Maas, H.G. Sensor Modelling and Camera Calibration for Close-Range Photogrammetry. ISPRS J. Photogramm. Remote Sens. 2016, 115, 37–46. [Google Scholar] [CrossRef]
  8. Rankin, A.; Huertas, A.; Matthies, L.; Bajracharya, M.; Assad, C.; Brennan, S.; Bellutta, P.; Sherwin, G.W. Unmanned Ground Vehicle Perception Using Thermal Infrared Cameras. In Proceedings of the International Society for Optical Engineering, Orlando, FL, USA, 25–29 April 2011; Volume 8045, p. 804503. [Google Scholar] [CrossRef]
  9. Sun, S.; Wei, W.; Yuan, X.; Zhou, R. Research on Calibration Methods of Long-Wave Infrared Camera and Visible Camera. J. Sens. 2022, 2022, 8667606. [Google Scholar] [CrossRef]
  10. Bison, P.; Bortolin, A.; Cadelano, G.; Ferrarini, G.; Furlan, K.; Grinzato, E. Geometrical Correction and Photogrammetric Approach in Thermographic Inspection of Buildings. In Proceedings of the 11th International Conference on Quantitative InfraRed Thermography (QIRT 2012), Naples, Italy, 11–14 June 2012; pp. 1–9. [Google Scholar] [CrossRef]
  11. Harguess, J.; Strange, S. Infrared Stereo Calibration for Unmanned Ground Vehicle Navigation. In Proceedings of the International Society for Optical Engineering, Baltimore, MD, USA, 5–9 May 2014; Volume 9084, p. 90840S. [Google Scholar] [CrossRef]
  12. Motayyeb, S.; Samadzedegan, F.; Dadrass Javan, F.; Hosseinpour, H. Fusion of UAV-based Infrared and Visible Images for Thermal Leakage Map Generation of Building Façades. Heliyon 2023, 9, e14551. [Google Scholar] [CrossRef]
  13. ElSheikh, A.; Abu-Nabah, B.A.; Hamdan, M.O.; Tian, G.Y. Infrared Camera Geometric Calibration: A Review and a Precise Thermal Radiation Checkerboard Target. Sensors 2023, 23, 3479. [Google Scholar] [CrossRef]
  14. Zhang, C.; Zou, Y.; Dimyadi, J.; Chang, R. Thermal-textured BIM Generation for Building Energy Audit with UAV Image Fusion and Histogram-based Enhancement. Energy Build. 2023, 301, 113710. [Google Scholar] [CrossRef]
  15. Roshan, M.C.; Isaksson, M.; Pranata, A. A Geometric Calibration Method for Thermal Cameras using a ChArUco Board. Infrared Phys. Technol. 2024, 138, 105219. [Google Scholar] [CrossRef]
  16. Garrido-Jurado, S.; Muñoz-Salinas, R.; Madrid-Cuevas, F.; Marín-Jiménez, M. Automatic Generation and Detection of Highly Reliable Fiducial Markers under Occlusion. Pattern Recognit. 2014, 47, 2280–2292. [Google Scholar] [CrossRef]
  17. Westfeld, P.; Mader, D.; Maas, H.G. Generation of TIR-attributed 3D Point Clouds from UAV-based Thermal Imagery. Photogramm.—Fernerkund.—Geoinf. 2015, 2015, 381–393. [Google Scholar] [CrossRef]
  18. Yastikli, N.; Guler, E. Performance Evaluation of Thermographic Cameras for Photogrammetric Measurements. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2013, XL-1/W1, 383–387. [Google Scholar] [CrossRef]
  19. Rizzi, A.; Voltolini, F.; Girardi, S.; Gonzo, L.; Remondino, F. Digital Preservation, Documentation and Analysis of Paintings, Monuments and Large Cultural Heritage with Infrared Technology, Digital Cameras and Range Sensors. In Proceedings of the 21st International CIPA Symposium, Athens, Greece, 1–6 October 2007; pp. 1–6. [Google Scholar]
  20. Rangel, J.; Soldan, S. 3D Thermal Imaging: Fusion of Thermography and Depth Cameras. In Proceedings of the 12th International Conference on Quantitative InfraRed Thermography (QIRT 2014), Bordeaux, France, 7–11 July 2014; pp. 1–10. [Google Scholar] [CrossRef]
  21. Adán, A.; Prado, T.; Prieto, S.A.; Quintana, B. Fusion of Thermal Imagery and LiDAR Data for Generating TBIM Models. In Proceedings of the IEEE SENSORS 2017, Glasgow, UK, 29 October–1 November 2017; pp. 1–3. [Google Scholar] [CrossRef]
  22. Luhmann, T.; Piechel, J.; Roelfs, T. Geometric Calibration of Thermographic Cameras. In Thermal Infrared Remote Sensing; Kuenzer, C., Dech, S., Eds.; Springer: Dordrecht, The Netherlands, 2013; Volume 17, pp. 27–42. [Google Scholar] [CrossRef]
  23. Dlesk, A.; Vach, K.; Pavelka, K. Transformations in the Photogrammetric Co-Processing of Thermal Infrared Images and RGB Images. Sensors 2021, 21, 5061. [Google Scholar] [CrossRef]
  24. Sutherland, N.; Marsh, S.; Priestnall, G.; Bryan, P.; Mills, J. InfraRed Thermography and 3D-Data Fusion for Architectural Heritage: A Scoping Review. Remote Sens. 2023, 15, 2422. [Google Scholar] [CrossRef]
  25. Alba, M.I.; Barazzetti, L.; Scaioni, M.; Rosina, E.; Previtali, M. Mapping Infrared Data on Terrestrial Laser Scanning 3D Models of Buildings. Remote Sens. 2011, 3, 1847–1870. [Google Scholar] [CrossRef]
  26. Ham, Y.; Golparvar-Fard, M. An Automated Vision-Based Method for Rapid 3D Energy Performance Modeling of Existing Buildings Using Thermal and Digital Imagery. Adv. Eng. Inform. 2013, 27, 395–409. [Google Scholar] [CrossRef]
  27. Patrucco, G.; Cortese, G.; Giulio Tonolo, F.; Spanò, A. Thermal and Optical Data Fusion Supporting Built Heritage Analyses. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2020, XLIII-B3-2020, 619–626. [Google Scholar] [CrossRef]
  28. Adamopoulos, E.; Volinia, M.; Girotto, M.; Rinaudo, F. Three-Dimensional Thermal Mapping from IRT Images for Rapid Architectural Heritage NDT. Buildings 2020, 10, 187. [Google Scholar] [CrossRef]
  29. Lecomte, V.; Macher, H.; Landes, T. Combination of Thermal Infrared Images and Laserscanning Data for 3D Thermal Point Cloud Generation on Buildings and Trees. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2022, XLVIII-2/W1-2022, 129–136. [Google Scholar] [CrossRef]
  30. Macher, H.; Landes, T. Combining TIR Images and Point Clouds for Urban Scenes Modelling. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2022, XLIII-B2-2022, 425–431. [Google Scholar] [CrossRef]
  31. Lerma, J.L.; Navarro, S.; Cabrelles, M.; Seguí, A.E. Camera Calibration with Baseline Distance Constraints. Photogramm. Rec. 2010, 25, 140–158. [Google Scholar] [CrossRef]
  32. Ursine, W.; Calado, F.; Teixeira, G.; Diniz, H.; Silvino, S.; De Andrade, R. Thermal/Visible Autonomous Stereo Visio System Calibration Methodology for Non-controlled Environments. In Proceedings of the 11th International Conference on Quantitative InfraRed Thermography (QIRT 2012), Naples, Italy, 11–14 June 2012; pp. 1–11. [Google Scholar] [CrossRef]
  33. Fraser, C. Automatic Camera Calibration in Close Range Photogrammetry. Photogramm. Eng. Remote Sens. 2013, 79, 381–388. [Google Scholar] [CrossRef]
  34. Hastedt, H.; Luhmann, T. Investigations on the Quality of the Interior Orientation and its Impact in Object Space for UAV Photogrammetry. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2015, XL-1/W4, 321–328. [Google Scholar] [CrossRef]
  35. Duran, Z.; Atik, M.E. Accuracy Comparison of Interior Orientation Parameters from Different Photogrammetric Software and Direct Linear Transformation Method. Int. J. Eng. Geosci. 2021, 6, 74–80. [Google Scholar] [CrossRef]
  36. ASTM E1933-14(2018); Practice for Measuring and Compensating for Emissivity Using Infrared Imaging Radiometers. ASTM International: West Conshohocken, PA, USA, 2022. [CrossRef]
  37. Brown, D.C. Close-Range Camera Calibration. Photogramm. Eng. 1971, 37, 855–866. [Google Scholar]
  38. Workswell s.r.o. Workswell WIRIS Pro Datasheet. 2021. Available online: https://my.workswell.eu/download/file/33/en (accessed on 4 February 2025).
  39. Luhmann, T.; Robson, S.; Kyle, S. (Eds.) Close-Range Photogrammetry: Principles, Methods and Applications, 2nd ed.; Whittles Publishing: Dunbeath, UK, 2006. [Google Scholar]
  40. Perda, G.; Morelli, L.; Remondino, F.; Fraser, C.; Luhmann, T. Analyzing Target-, Handcrafted- and Learning-Based Methods for Automated 3D Measurement and Modelling. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2024, XLVIII-2/W7-2024, 105–112. [Google Scholar] [CrossRef]
  41. Zhang, Z. A Flexible New Technique for Camera Calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  42. Wan, Q.; Brede, B.; Smigaj, M.; Kooistra, L. Factors Influencing Temperature Measurements from Miniaturized Thermal Infrared (TIR) Cameras: A Laboratory-Based Approach. Sensors 2021, 21, 8466. [Google Scholar] [CrossRef]
  43. Kelly, J.; Kljun, N.; Olsson, P.O.; Mihai, L.; Liljeblad, B.; Weslien, P.; Klemedtsson, L.; Eklundh, L. Challenges and Best Practices for Deriving Temperature Data from an Uncalibrated UAV Thermal Infrared Camera. Remote Sens. 2019, 11, 567. [Google Scholar] [CrossRef]
  44. Jin, Y.; Mishkin, D.; Mishchuk, A.; Matas, J.; Fua, P.; Yi, K.M.; Trulls, E. Image Matching Across Wide Baselines: From Paper to Practice. Int. J. Comput. Vis. 2020, 129, 517–547. [Google Scholar] [CrossRef]
  45. Velesaca, H.O.; Bastidas, G.; Rouhani, M.; Sappa, A.D. Multimodal Image Registration Techniques: A Comprehensive Survey. Multimed. Tools Appl. 2024, 83, 63919–63947. [Google Scholar] [CrossRef]
  46. Elias, M.; Weitkamp, A.; Eltner, A. Multi-Modal Image Matching to Colorize a SLAM-based Point Cloud with Arbitrary Data from a Thermal Camera. ISPRS Open J. Photogramm. Remote Sens. 2023, 9, 100041. [Google Scholar] [CrossRef]
  47. Morelli, L.; Ioli, F.; Maiwald, F.; Mazzacca, G.; Menna, F.; Remondino, F. Deep-Image-Matching: A Toolbox for Multiview Image Matching of Complex Scenarios. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2024, XLVIII-2/W4-2024, 309–316. [Google Scholar] [CrossRef]
  48. DeTone, D.; Malisiewicz, T.; Rabinovich, A. SuperPoint: Self-Supervised Interest Point Detection and Description. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 337–349. [Google Scholar] [CrossRef]
  49. Lindenberger, P.; Sarlin, P.E.; Pollefeys, M. LightGlue: Local Feature Matching at Light Speed. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 17581–17592. [Google Scholar] [CrossRef]
  50. Schonberger, J.L.; Frahm, J.M. Structure-from-Motion Revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar] [CrossRef]
  51. Sledz, A.; Heipke, C. Joint Bundle Adjustment of Thermal Infra-Red and Optical Images Based on Multimodal Matching. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2022, XLIII-B1-2022, 157–165. [Google Scholar] [CrossRef]
  52. Williams, J.; Corvaro, F.; Vignola, J.; Turo, D.; Marchetti, B.; Vitali, M. Application of Non-Invasive Active Infrared Thermography for Delamination Detection in Fresco. Int. J. Therm. Sci. 2022, 171, 107185. [Google Scholar] [CrossRef]
  53. Zhang, J.; Zhang, S.; Peng, G.; Zhang, H.; Wang, D. 3DRadar2ThermalCalib: Accurate Extrinsic Calibration between a 3D mmWave Radar and a Thermal Camera Using a Spherical-Trihedral. In Proceedings of the 25th International Conference on Intelligent Transportation Systems (ITSC), Macau, China, 8–12 October 2022; pp. 2744–2749. [Google Scholar] [CrossRef]
  54. Lagüela, S.; González-Jorge, H.; Armesto, J.; Arias, P. Calibration and Verification of Thermographic Cameras for Geometric Measurements. Infrared Phys. Technol. 2011, 54, 92–99. [Google Scholar] [CrossRef]
  55. Adán, A.; Pérez, V.; Vivancos, J.L.; Aparicio-Fernández, C.; Prieto, S.A. Proposing 3D Thermal Technology for Heritage Building Energy Monitoring. Remote Sens. 2021, 13, 1537. [Google Scholar] [CrossRef]
  56. Clarke, T.A.; Wang, X.; Fryer, J.G. The Principal Point and CCD Cameras. Photogramm. Rec. 1998, 16, 293–312. [Google Scholar] [CrossRef]
  57. Senn, J.A.; Mills, J.; Walsh, C.L.; Addy, S.; Peppa, M.V. On-Site Geometric Calibration of RPAS Mounted Sensors for SfM Photogrammetric Geomorphological Surveys. Earth Surf. Processes Landforms 2022, 47, 1615–1634. [Google Scholar] [CrossRef]
  58. Daakir, M.; Zhou, Y.; Pierrot Deseilligny, M.; Thom, C.; Martin, O.; Rupnik, E. Improvement of Photogrammetric Accuracy by Modeling and Correcting the Thermal Effect on Camera Calibration. ISPRS J. Photogramm. Remote Sens. 2019, 148, 142–155. [Google Scholar] [CrossRef]
  59. Tejedor, B.; Lucchi, E.; Bienvenido-Huertas, D.; Nardi, I. Non-Destructive Techniques (NDT) for the Diagnosis of Heritage Buildings: Traditional Procedures and Futures Perspectives. Energy Build. 2022, 263, 112029. [Google Scholar] [CrossRef]
  60. Adán, A.; Pérez, V.; Ramón, A.; Castilla, F.J. Correction of Temperature from Infrared Cameras for More Precise As-Is 3D Thermal Models of Buildings. Appl. Sci. 2023, 13, 6779. [Google Scholar] [CrossRef]
Figure 1. Cameras used for TIR geometric calibration and IRT-3DDF methods, comprising the following: (a) Workswell WIRIS Pro (WWP and VIS), (b) Sony α 7R II (RGB1), and (c) Nikon D750 (RGB2).
Figure 1. Cameras used for TIR geometric calibration and IRT-3DDF methods, comprising the following: (a) Workswell WIRIS Pro (WWP and VIS), (b) Sony α 7R II (RGB1), and (c) Nikon D750 (RGB2).
Metrology 05 00043 g001
Figure 2. 2D checker-board imaged in both (a) VIS and (b) TIR domains.
Figure 2. 2D checker-board imaged in both (a) VIS and (b) TIR domains.
Metrology 05 00043 g002
Figure 3. 3D fields imaged in both VIS and TIR image domains for the (a,b) IO and (c,d) RO calibration procedures.
Figure 3. 3D fields imaged in both VIS and TIR image domains for the (a,b) IO and (c,d) RO calibration procedures.
Metrology 05 00043 g003
Figure 4. IO camera networks obtained for the (a) 2D board using MATLAB, (b) 3D field using Agisoft Metashape, and (c) 3D field using Australis.
Figure 4. IO camera networks obtained for the (a) 2D board using MATLAB, (b) 3D field using Agisoft Metashape, and (c) 3D field using Australis.
Metrology 05 00043 g004
Figure 5. Gaussian radial and tangential distortion profiles generated from the obtained IO parameters with the 2D board using MATLAB (MAT), 3D field using Agisoft Metashape (AM), and 3D field using Australis (AUST). Here, the graphs represent the distortion originating at the image centre ( u o , v o ) and finishing at the maximum radial distance.
Figure 5. Gaussian radial and tangential distortion profiles generated from the obtained IO parameters with the 2D board using MATLAB (MAT), 3D field using Agisoft Metashape (AM), and 3D field using Australis (AUST). Here, the graphs represent the distortion originating at the image centre ( u o , v o ) and finishing at the maximum radial distance.
Metrology 05 00043 g005
Figure 6. Distortion models for the (a,b) 2D board using MATLAB, (c,d) 3D field using Agisoft Metashape, and (e,f) 3D field using Australis, showing distortion magnitude (pix), distortion normals (↗), principal point (), and image centre (320,256) (×).
Figure 6. Distortion models for the (a,b) 2D board using MATLAB, (c,d) 3D field using Agisoft Metashape, and (e,f) 3D field using Australis, showing distortion magnitude (pix), distortion normals (↗), principal point (), and image centre (320,256) (×).
Metrology 05 00043 g006
Figure 7. RO camera networks obtained for the RGB2-WWP stereo baseline with (a) 2D board using MATLAB, (b) 3D field using Agisoft Metashape, and (c) 3D field using Australis.
Figure 7. RO camera networks obtained for the RGB2-WWP stereo baseline with (a) 2D board using MATLAB, (b) 3D field using Agisoft Metashape, and (c) 3D field using Australis.
Metrology 05 00043 g007
Figure 8. Lenton Lodge Gatehouse, Nottingham, UK.
Figure 8. Lenton Lodge Gatehouse, Nottingham, UK.
Metrology 05 00043 g008
Figure 9. An example WWP-RGB1 image pair generated from SuperPoint + LightGlue (Sp + Lg) in Deep-Image-Matching (DIM).
Figure 9. An example WWP-RGB1 image pair generated from SuperPoint + LightGlue (Sp + Lg) in Deep-Image-Matching (DIM).
Metrology 05 00043 g009
Figure 10. Combined TIR-RGB bundle block adjustment camera network for the Lenton Lodge Gatehouse, showing the single strip of WWP images and two strips of RGB1 images.
Figure 10. Combined TIR-RGB bundle block adjustment camera network for the Lenton Lodge Gatehouse, showing the single strip of WWP images and two strips of RGB1 images.
Metrology 05 00043 g010
Figure 11. Generated 3D models from the combined TIR-RGB bundle block adjustment of the Lenton Lodge Gatehouse, textured with (a) RGB1 and (b) WWP thermal images.
Figure 11. Generated 3D models from the combined TIR-RGB bundle block adjustment of the Lenton Lodge Gatehouse, textured with (a) RGB1 and (b) WWP thermal images.
Metrology 05 00043 g011
Figure 12. Castello Di Arco’s medieval frescoes, Trentino, Italy, featuring (a) the fresco corner under investigation, (b) the stereo rig featuring the WWP and RGB2 cameras, and (c) the stereo rig camera network.
Figure 12. Castello Di Arco’s medieval frescoes, Trentino, Italy, featuring (a) the fresco corner under investigation, (b) the stereo rig featuring the WWP and RGB2 cameras, and (c) the stereo rig camera network.
Metrology 05 00043 g012
Figure 13. An example of geometrically un-verified SIFT matches obtained in COLMAP between sequential mono-modal (a) RGB2 and (b) WWP images, resulting in 5924 and 22 matches, respectively.
Figure 13. An example of geometrically un-verified SIFT matches obtained in COLMAP between sequential mono-modal (a) RGB2 and (b) WWP images, resulting in 5924 and 22 matches, respectively.
Metrology 05 00043 g013
Figure 14. Generated textured models of Castello Di Arco’s medieval frescoes, with the (a) RGB results showing the visible signs of deterioration and (b) showing the temperature range visualising the most significant (cold) areas of water ingress.
Figure 14. Generated textured models of Castello Di Arco’s medieval frescoes, with the (a) RGB results showing the visible signs of deterioration and (b) showing the temperature range visualising the most significant (cold) areas of water ingress.
Metrology 05 00043 g014
Table 1. Camera specifications for the Workswell WIRIS Pro (WWP and VIS), Sony α 7R II (RGB1), and Nikon D750 (RGB2) cameras.
Table 1. Camera specifications for the Workswell WIRIS Pro (WWP and VIS), Sony α 7R II (RGB1), and Nikon D750 (RGB2) cameras.
Workswell WIRIS Pro Infrared Sensor (WWP)
Resolution640× 512 pix
Sensor Size (FPA)10.88 × 8.71 mm
Pixel Size17.00 µm
Nominal Focal Length13.00 mm
Spectral Range (LWIR)7.5–13.5 µm
Temp. Sensitivity0.05 °C
Temp. Accuracy±2 °C
Workswell WIRIS Pro Visible Sensor (VIS)
Resolution1920 × 1080 pix
Sensor Size (CMOS)5.23 × 2.94 mm
Pixel Size2.72 µm
Nominal Focal Length3.50 mm
Sony α 7R II (RGB1)
Resolution7952× 5304 pix
Sensor Size (CMOS)35.90× 24.00 mm
Pixel Size4.50 µm
Nominal Focal Length35.00 mm
Nikon D750 (RGB2)
Resolution6016× 4016 pix
Sensor Size (CMOS)35.90 × 24.00 mm
Pixel Size5.95 µm
Nominal Focal Length50.00 mm
Table 2. Calculation of material emissivity ( ε ), standard deviation ( σ ), and difference in emissivity of materials ( ε Δ ) used for both 2D board and 3D field calibration targets.
Table 2. Calculation of material emissivity ( ε ), standard deviation ( σ ), and difference in emissivity of materials ( ε Δ ) used for both 2D board and 3D field calibration targets.
Material ε σ Δ ε
2D BoardDiBond®0.620.040.20
UV-printing0.820.04
3D FieldAluminium0.200.030.65
Rubber0.850.05
Table 3. IO parameters for WWP obtained with the 2D board using MATLAB, 3D Field using Agisoft Metashape, and 3D Field using Australis.
Table 3. IO parameters for WWP obtained with the 2D board using MATLAB, 3D Field using Agisoft Metashape, and 3D Field using Australis.
CoefficientIntrinsics2D Board
(MATLAB)
3D Field
(Agisoft Metashape
3D Field
(Australis)
Value σ Value σ Value σ
k 2 f (mm)13.0210.03213.0380.01513.0340.014
c x (pix)311.2191.821312.5850.626312.9260.740
c y (pix)255.5581.928257.7550.614257.9210.721
k 1 0.0440.009−0.0430.004−0.0390.003
k 2 0.0050.0540.3680.0140.3470.013
p 1 −0.0000.001−0.0010.0000.0000.000
p 2 0.0010.0010.0000.000−0.0000.000
MRE (pix)0.500.160.12
S o  (mm)3.203.252.06
k 3 f (mm)13.0170.03213.0330.01513.0300.014
c x (pix)311.2321.823312.5410.623312.9080.730
c y (pix)255.4821.926257.7410.612257.9400.711
k 1 0.0550.017−0.0310.008−0.0310.008
k 2 −0.1770.2470.2470.0780.2700.073
k 3 0.7911.0470.3500.2210.2140.204
p 1 −0.0000.001−0.0010.0000.0000.000
p 2 0.0010.0010.0000.000−0.0000.000
MRE (pix)0.500.160.12
S o  (mm)3.653.252.08
Table 4. RO parameters for WWP, VIS, and RGB2 using the 2D board and 3D field calibration targets.
Table 4. RO parameters for WWP, VIS, and RGB2 using the 2D board and 3D field calibration targets.
CoefficientBaselineParameters2D Board
(MATLAB)
3D Field
(Agisoft Metashape)
3D Field
(Australis)
Value σ Value σ Value σ
k 2 VIS-WWPX (mm)−41.5300.191−39.7380.396−40.2712.691
Y (mm)−1.9920.1490.4090.363−0.1482.845
Z (mm)0.3090.1141.1420.2196.1241.456
ω (°)−0.0800.013−0.0920.008−0.1750.073
ϕ (°)0.5790.0160.7170.0090.7010.071
κ (°)−0.0800.0100.0010.0050.0000.021
MRE (pix)0.760.550.22
S o  (mm)4.483.202.32
RGB2-WWPX (mm)−184.4640.331−183.6590.629−183.0841.493
Y (mm)−0.2160.2411.5010.5781.2422.023
Z (mm)6.9250.2529.7370.50111.1541.276
ω (°)0.9170.0150.5860.0150.4240.080
ϕ (°)3.6440.0203.2430.0153.2080.066
κ (°)−0.0460.015−0.1470.014−0.1750.296
MRE (pix)0.871.100.14
S o  (mm)5.262.170.77
k 3 VIS-WWPX (mm)−41.5030.191−39.7210.395−40.5402.889
Y (mm)−1.9900.1490.4060.362−0.6142.895
Z (mm)0.3710.1141.2720.219−1.3381.094
ω (°)−0.0800.013−0.0920.0080.0250.075
ϕ (°)0.5790.0160.7130.0090.7150.072
κ (°)−0.0800.0100.0010.005−0.0070.019
MRE (pix)0.760.550.29
S o  (mm)4.483.203.54
RGB2-WWPX (mm)−184.4460.331−183.6440.629−183.1431.495
Y (mm)−0.2150.2411.4900.5781.0642.143
Z (mm)7.0170.2529.3400.50211.1321.263
ω (°)0.9170.0140.5650.0150.4180.085
ϕ (°)3.6440.0203.2670.0153.2130.068
κ (°)−0.0460.015−0.1460.014−0.1540.311
MRE (pix)0.871.100.14
S o  (mm)5.262.110.76
Table 5. Results of the combined TIR-RGB bundle block adjustment for the Lenton Lodge Gatehouse, separating results based on the software and distortion coefficients utilised.
Table 5. Results of the combined TIR-RGB bundle block adjustment for the Lenton Lodge Gatehouse, separating results based on the software and distortion coefficients utilised.
ParametersRMSERGB1
(mm)
MREWWP
(pix)
RMSEWWP (mm)
XYZ TOT
Self-Calibration k 2 2.860.324.0216.319.0319.07
MATLAB k 2 2.590.403.8814.1110.5518.04
k 3 2.600.313.9812.398.6315.62
Agisoft
Metashape
k 2 2.440.272.725.746.128.82
k 3 2.470.282.456.915.699.28
Australis k 2 2.320.293.055.446.358.89
k 3 2.190.293.065.606.288.95
Table 6. RO results for Castello Di Arco’s medieval frescoes, with the measured scale bar RMSE (mm) for each baseline (VIS-WWP or WWP-NGB2) separated into each image modality.
Table 6. RO results for Castello Di Arco’s medieval frescoes, with the measured scale bar RMSE (mm) for each baseline (VIS-WWP or WWP-NGB2) separated into each image modality.
ParametersRMSE VIS-WWP (mm)RMSE RGB2-WWP (mm)
VISWWPRGB2WWP
Self-Calibration k 2 2.434.301.461.42
MATLAB k 2 1.495.042.208.72
k 3 1.515.012.178.82
Agisoft
Metashape
k 2 2.724.121.261.87
k 3 2.744.241.261.89
Australis k 2 2.253.831.261.58
k 3 2.283.951.261.65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sutherland, N.; Marsh, S.; Remondino, F.; Perda, G.; Bryan, P.; Mills, J. Geometric Calibration of Thermal Infrared Cameras: A Comparative Analysis for Photogrammetric Data Fusion. Metrology 2025, 5, 43. https://doi.org/10.3390/metrology5030043

AMA Style

Sutherland N, Marsh S, Remondino F, Perda G, Bryan P, Mills J. Geometric Calibration of Thermal Infrared Cameras: A Comparative Analysis for Photogrammetric Data Fusion. Metrology. 2025; 5(3):43. https://doi.org/10.3390/metrology5030043

Chicago/Turabian Style

Sutherland, Neil, Stuart Marsh, Fabio Remondino, Giulio Perda, Paul Bryan, and Jon Mills. 2025. "Geometric Calibration of Thermal Infrared Cameras: A Comparative Analysis for Photogrammetric Data Fusion" Metrology 5, no. 3: 43. https://doi.org/10.3390/metrology5030043

APA Style

Sutherland, N., Marsh, S., Remondino, F., Perda, G., Bryan, P., & Mills, J. (2025). Geometric Calibration of Thermal Infrared Cameras: A Comparative Analysis for Photogrammetric Data Fusion. Metrology, 5(3), 43. https://doi.org/10.3390/metrology5030043

Article Metrics

Back to TopTop