High georeferencing accuracy is a crucial requirement for any remote sensing analysis. A new approach to georeferencing improvement is proposed, specifically designed for thermal IR nighttime imagery. The proposed method was tested on a set of nighttime ECOSTRESS LST images of the East African Rift, and a georeferencing improvement of 12.0 pixels on average was noted, while an average accuracy of ±2.9 pixels was achieved. The improvement in accuracy is supported by
Figure 6, which presents a comparison between the mean error per image from the original processing and from the proposed method. There are only two cases in which the proposed method led to larger errors than in the original processing, but the absolute values of these errors are rather small (
Figure 6). On average, the georeferencing error decreased by 12.0 pixels, from 14.9 pixels to 2.9 pixels.
4.1. Analysis of the Method
The proposed method assumes that water body edges can be easily identified in TIR nighttime imagery and that the same edges can be identified in the reference. The first assumption is usually true due to the heat capacity and thermal inertia differences between land and water; however, in cases where the LST difference between land and the water body is small, adjustments to edge detection should be made. The Canny edge detection algorithm allows the adaptation of thresholds for the hysteresis procedure (in this implementation, it was adapted by using the parameter “threshold_sigma”; see
Appendix C Table A2 for details). As for the similarity with the reference, the water bodies in the land cover classification are easy to identify due to their strong spectral differences from land masses and vegetation. The LST of water bodies at night is higher than that of the land masses, and therefore, the edges in the reference and target images are similar enough for matching purposes (the few exceptions to this are discussed below).
The proposed method can be applied globally, provided that some water bodies are present and visible in the image. Since ECOSTRESS scenes have a 384 km swath width, chances are that at least some water bodies are present in an image, although areas with snow cover probably need to be excluded.
4.1.1. Implementation
As there is no dedicated method for improving the georeferencing of nighttime TIR imagery, the results obtained were compared to the standard processing of ECOSTRESS. The tie point residuals (presented in
Table 2) complement the information provided by manual checkpoint setting; these values have an average accuracy of ±1.3 pixels. Since the residuals are provided as metadata for each processed image, they can be used by all users for the preliminary assessment of the transformation accuracy and reliability. If Euclidean distances are similarly low for all tie points, georeferencing is likely to be accurate, whereas singular larger distances suggest a larger error, e.g., due to rotation error.
Limitations of the validation procedure need to be considered in the analysis. The distribution of check points is supposed to be as homogeneous as possible, and some check points should be placed at the edges of each image. However, cloud cover, radiometric artefacts, and the lack of unambiguous features to serve as check points influenced the checkpoint distribution and, therefore, may have introduced a bias into the validation dataset.
It is important to note that the proposed method considers the most up-to-date reference image but does not accommodate rapid events (e.g., weather or anthropogenic) that change the contours of water bodies. In the processing, a temporal resolution of one month was used, which is sufficient for the analysed study area, where land cover changes happen over the course of seasons. However, for other study areas, the time window may need to be adjusted. For instance, if there is a dynamic change in the water level, a temporal resolution of one month may not be enough. At the same time, it is important to remember that decreasing the time window can come at the cost of data gaps, because fewer datasets will be available to create a cloud-free mosaic.
A special case of dynamic change to consider is a seashore area with high tides. If enough inland water bodies are present, sea areas can be masked and disregarded in the matching. In principle, a reference with the same tide level can be used; however, finding a cloud-free reference with the same tide level may be challenging.
4.1.2. Data Quality
The proposed method depends on the visibility of matching objects on the Earth’s surface; therefore, the largest weakness in the processing comes from imprecise cloud masking in both target and reference data. If clouds are not masked properly, their edges will be used in feature matching, thereby introducing errors. The addition of a statistics-based cloud mask based on the LST histogram was introduced to reduce errors from cloud edges. Several tie point evaluation steps were implemented to remove invalid tie points based on the matching of cloud edges, but some erroneous tie points may nevertheless remain.
Additionally, some errors can appear due to georeferencing errors of cloud masks, because cloud masks of both the target and reference are used on both datasets. Thus, if a cloud covers a water body in the target image, but the georeferencing of these masks has an offset, the matching may be faulty. Such errors are possible despite the evaluation of tie points; however, no such cases were observed in the images processed.
Errors can also appear due to other reasons. The reference images, which are binary masks, were resampled from their original spatial resolution of 20 m to match the spatial resolution of ECOSTRESS imagery at 70 m, which can potentially introduce a shift in the geolocation of water body edges. Additionally, the reference imagery was acquired at different wavelengths from those used to obtain the target images. Differences may especially appear in areas where land cover is more complex than the classes provided in the SCL, such as vegetation floating on water. In the SCL, such areas will be classified as vegetation, but in ECOSTRESS images, they are seen as warmer than land surfaces, and in processing, they are treated as water. Additionally, assumptions were made to create the classification layer, and these assumptions may lead to differences in the delimitation of a water body.
Lastly, misclassification errors in the SCL additionally add to the overall error per image. The risk of maintaining an error in the reference image, however, is minimised due to the fact that for each month of acquisitions, a separate reference layer is created.
4.1.3. Finding and Evaluating Tie Points
The proposed matching algorithm uses a brute force principle by comparing image fragments and shifting these fragments pixel by pixel. This allows the x-offset, y-offset, and rotation to be derived for each image. In the manual assessment of ECOSTRESS imagery, no images were encountered for which a transformation including scaling or shearing would be necessary, so we decided to opt for fewer transformation parameters and a lower minimum number of tie points for the improvement process to take place. However, since ECOSTRESS has a swath width of approximately 400 km, errors can appear in higher off-nadir angles, especially in the most extreme pixels. The implementation of the method, however, allows for fitting scaling and shearing. The number of parameters that can be fitted depends on the number of valid tie points found, so the adaption of the parameter “minimum_num_tp” would be required to fit additional parameters.
The proposed method does not consider rotation for each water body separately, and the geometric resolution of a tie point is limited to a pixel. The results presented in
Figure 5 suggest an overcorrection in cases where rotation exceeds 0.2
. However, due to different manoeuvres taking place on the ISS, rotation in the imagery can appear, so disregarding this parameter does not seem to be a correct solution.
In this implementation, two parameters limiting the water body choice have been defined: the minimum size of the water body and the minimum number of detected edge pixels. In the study area, numerous small, round water bodies can be found, and their similarity may lead to difficulties in fitting transformation parameters. We decided to reduce the number of matching objects to large water bodies, which typically have more complex shapes, in order to avoid introducing mismatches. Potentially, algorithms, such as Iterative Closest Point, could enhance the precision of matching. In areas with multiple small water bodies, constellations thereof could be used for matching, with a similar principle to that used in star trackers.
The reliability of the proposed georeferencing method strongly depends on the number of matches found during the process. It appears that images with the highest error values were processed with only a few tie points (
Table 2 and
Table 3). Generally, the more tie points used for fitting the transformation parameters, the more reliable the method becomes. However, keeping even a single wrong tie point means that all transformation parameters for the complete image will be wrong. Therefore, we decided to opt for using fewer tie points, but with higher restrictions on accuracy. This principle works well as long as no scaling or shearing is present. Nevertheless, errors can also appear if tie points in a group are located within small distances to each other and one tie point is located in the distant part of the image, because the matching error for this single point is much harder to identify. Such a situation may lead to the overcorrection of rotation, which is possibly the case for 20210717T025340, 20210724T234916, 20211207T180801, and 20201005T192641. This is visible when analysing the standard deviation of errors (
Table 3). For instance, in image 20201005T192641, one out of three tie points (
Table 2) is an invalid match but was not picked out in the tie point evaluation step. This wrong tie point influenced the whole transformation matrix, resulting in an average checkpoint error of 4.3 pixels for this image (
Table 3).
The parameters in the evaluation criteria were empirically derived and possibly need adaption for other study areas. There is a trade-off between the strictness of the evaluation and the number of images that will not be processed at all due to insufficient remaining tie points. The user needs to decide upon the rigour of the tie point evaluation, considering the application.
4.2. Future Directions
The proposed method can be optimised in the future, e.g., by using algorithms such as Iterative Closest Point, including the rotation of each individual water body, accounting for sub-pixel offsets, and using constellations of small water bodies for matching.
The method can potentially be applied to imagery from different sensors, because water bodies maintain a high contrast with the surrounding land masses throughout the night. As new operational thermal sensors appear (e.g., TRISHNA, Surface Biology and Geology and Land Surface Temperature Monitoring, as well as low-cost commercial constellations), applications using nighttime TIR imagery will gain importance. The ability to accurately georeference imagery in an automated way enables the provision of automated solutions on a global scale, and the proposed method has the potential to support this process.
With the increasing availability of computational power, it is possible to focus on better image-matching approaches compared to object-based matching. The presented research proves that using an up-to-date reference solves the issue of outdated reference basemaps due to dynamically changing land cover. Since the preparation of a large mosaic for referencing only takes a few seconds in a cloud-based environment such as GEE, it is possible to use the most recent images from high-resolution operational satellites with high georeferencing accuracy as a reference, instead of large reference databases such as the Landsat Orthobase. In the processing, the reference masks were downloaded from GEE, and the process ran locally (which takes approximately 25 min per image), but if the approach were reversed and target images were uploaded to a cloud computing platform, the overall processing time might be strongly reduced.