1. Introduction
The Gulf of Mexico is an extremely complex, large, resilient marine ecosystem [
1] where important coral reef areas have been reported. These ecosystems are considered critical habitats, as they provide an important range of ecosystem services to Mexico, Cuba, and the United States [
2]. Recently, some authors have considered three regions as a general division of the Mexican Atlantic territory: the Reef Corridor of the Southwest Gulf of Mexico (RCSGM), the Yucatan and Campeche Bank, and the Mexican Caribbean [
3]. In the RCSGM, two natural, protected areas are found: the Sistema Arrecifal Lobos-Tuxpan Flora and Fauna Protection Zone (SALT) and the Veracruz Reef System National Park (Spanish: PNSAV) [
4]. There are records of fish species, stony corals, macroalgae, crustaceans, echinoderms, gastropods, native species, and also invasive species in the PNSAV; however, the health status of 50% of the coral reefs is poorly known [
5]. The Census of Marine Life conducted between 2000 and 2010 collected data indicating “that at least 50% and potentially
% of marine species remain undescribed by science”; however, Snelgrove [
6] points out that emerging technologies are accelerating the detection of new habitats and taxonomic categorization. The ecological monitoring of coral reefs is the collection of data and information, both physical and biological, from the natural environment to detect changes in the processes or attributes measured [
7,
8]. This process begins by sampling a representative area of the reef that can be resampled over time to detect changes in reef conditions; hence, it is important to use standardized methods with standard sample sizes [
8]. One of the most common monitoring methods to assess coral reef conditions is the Point Intercept Transect (PIT) method [
9], and, as a variation of this method for research purposes, the video-transect method is applied [
8]. Considering not every variable of the reef can be measured, indicators are used to report changes over time [
8]. Indicators are metrics used to explain the causal relationship between an ecosystem attribute or process and ecosystem degradation [
7]. The species–area relationship, cited in ecology as the power law, was observed in 1778 by Forster and mathematically modeled in 1921 by Arrhenius [
10], increasing the interest of researchers in complexity metrics. Through pioneering research that included underwater imaging and the monitoring of ecological indicators in the PNSAV [
11], a variation was found in abundance metric results, obtained from the comparison between data recorded in situ by diving and those recorded offline from videos filmed by a Remotely Operated Vehicle (ROV), and it was assumed that the image quality was affected by the visibility conditions in the water column. The quality of underwater images is affected by the presence of one or more factors such as water density, light attenuation, and the dispersion effect [
12]. When working with large numbers of data, their quality is subject to issues at three points: the data source, generation, and processing. These issues compromise reliability, veracity, consistency, completeness, structure, and transmission; therefore, they require the use of data quality indicators [
13] in the stage identified in our research as the underwater image preprocessing.
Underwater photomosaics are a resource of interest due to their applicability in different areas, such as inspection, detection of objects, control of underwater vehicles, studies in marine biology, and archaeology exploration [
14]. One of the earlier records of underwater photomosaics is the composition assembled from the series of images captured from the Thresher submarine between 1963 and 1964, which was used to determine the possible causes of its sinking [
15]. In 1987, the Woods Hole Oceanographic Institution (WHOI) built a Titanic photomosaic with approximately 100 images taken in 1985 by the Acoustically Navigated Geological Undersea Surveyor (ANGUS), requiring about 700 h to accomplish [
16]. Another implementation of this technique involved the wreck site of the lumber schooner Rouse Simmons, where, in 2006, a digital video was recorded with a video camera mounted on a Diver Propulsion Vehicle (DPV), capturing 242 images that were overlaid and hand-assembled in Adobe Photoshop 7.0 to obtain a plan and profile views of the boat [
17]. In marine biology, photomosaics have been used to detect changes in the abundance, cover, and size of benthic organisms, using an algorithm for image registration and the estimation of image motion and camera trajectory, achieving a two-dimensional high-resolution photomosaic of the reef benthos which is ideally useful in survey areas
[
18]. Image stitching, also known as image mosaicing, is a computer vision application where small-scale images are aligned to find overlaps with the aim of creating a single large-scale composition using an algorithm, known as the feature detector, to detect features or interest points (keypoints) [
19]. Ancuti et al. demonstrated that a previous improvement of an underwater image increased the number of detected keypoints [
14]. Many works have contributed to the improvement of a single underwater image, and can be mainly classified into two groups: image enhancement and image restoration [
20,
21,
22]. The enhancement of underwater images incorporates image data without considering underwater image formation models [
20,
21,
22]; meanwhile, the restoration of underwater images is governed by the criteria and prior knowledge of an underwater image formation model [
20,
21,
22]. According to Alsakar et al., image enhancement methods can be categorized as spatial domain-based, frequency domain-based, color constancy-based, and deep learning-based [
20]. Miao et al. suggest the following classifications of image restoration methods based on their underlying underwater imaging model: point spread function-based, Jaffe–McGlamery model-based, turbulence degradation model-based, and image dehazing-based [
22]. Dark Channel Prior (DCP) is considered the most studied single-image dehazing-based method in the literature [
23]. This approach, proposed in 2009 by He et al., recovers a scene radiance from the estimation of atmospheric light and a transmission map based on the observation “that at least one color channel has some pixels whose intensities are very low and close to zero” in haze-free outdoor images [
24]. Sea-Thru [
25] is an image restoration algorithm based on a revised underwater image formation model proposed by Akkaynak and Treibitz in 2018 [
26] and was developed to estimate the backscatter and attenuation coefficients of an RGBD (Red, Green, Blue, Depth) image with the aim of recovering lost colors and achieving water removal [
14]. Sea-Thru also works with monocular images. The Color Balance and Fusion Based on White Balancing (CBFWB) algorithm [
14], belonging to the group of image restoration algorithms, includes a fusion with color-correction methods [
22], stands out for “better exposedness of the dark regions, improved global contrast, and edges sharpness”, and is applicable to image processing with the particularity to work in a manner that is “reasonably independent of the camera settings” [
14]. For the three algorithms selected for this work, we can give the following information: (a) DCP is the basis of 22 underwater restoration methods developed from 2010 to 2019 [
22]; (b) Sea-Thru is the most recently proposed underwater image formation model [
26]; and (c) an image improvement method comparison conducted by Ancuti et al. in 2018 found that CBFWB outperforms other methods in terms of the means of the average values of the PCQI, UCIQE, and UIQM metrics [
14].
The image quality evaluation metrics applied for quantitative analysis are the general-purpose Patch-Based Contrast Quality Index (PCQI) [
27], the metric proposed by Yang and Sowmya in 2015 known as Underwater Color Image Quality Evaluation (UCIQE) [
28], and the metric presented in 2016 by Panneta et al. called the Underwater Image Quality Measure (UIQM) [
29], all set to perform the underwater imaging appraisal [
14,
22]. However, image quality evaluation metrics do not consider computational complexity in the assessment of improvement algorithms. To select an underwater image improvement algorithm in this evaluation, as a preprocessing stage in the composition of underwater photomosaics, two key factors are considered: the processing time and the keypoint increase with respect to the original image, which suggests a multicriteria decision analysis. Jadhav and Sonar carried out a comparative study of multicriteria decision-making methods applied to software selection, evaluating the Analytic Hierarchy Process (AHP), Weighted Scoring Method (WScM), and Hybrid Knowledge-Based Systems (HKBS) [
30]. The fact that an application similar to this project has not been documented heretofore, and taking as a reference the comparison of methods used to evaluate software and the low complexity of the Decision Matrix, the score of the well-known Weighted Sum Method (WSuM) of Peter C. Fishburn is used as a decision metric, and this model can be tuned to the needs of the problem. The techniques used in this work to perform feature detection using improved images are Scale-Invariant Feature Transform (SIFT), presented by Lowe in 2004; and Oriented FAST and Rotated BRIEF (ORB), proposed by Rublee et al. in 2011 [
19]. SIFT is suitable for object recognition and image registration implementations because of its robustness under conditions of changes in scale, rotation, and illumination, while ORB manages to be a very useful method in the context of computational efficiency, which is why it is used in augmented-reality applications [
31].
The objective of this evaluation is to select a suitable method to improve the monocular underwater image database provided by the Institute of Marine and Fishery Sciences Studies (Spanish: ICIMAP) of Veracruzana University [
32]. Since this work is focused on applying computer vision techniques such as photomosaics to monitor coral reefs, evaluation metrics are the techniques used to detect and describe local points of interest in an image and the processing time of the improvement algorithm. Therefore, the contributions of this research are the following:
Application of three enhancement and restoration algorithms to the PNSAV monocular underwater image database with minimum hardware requirements;
Introduction of a novel multicriteria decision framework to evaluate the performance of image improvement algorithms built on two factors: detection of keypoints and processing time;
Incorporation of processing time as a decision criterion to address computational efficiency challenges, which are relevant for coral reef exploration applications where fast image processing is critical.
In this paper, an overall introduction is included in
Section 1.
Section 2 describes the materials and methods, and the results and discussion are described in
Section 3. Finally,
Section 4 closes this work with some conclusions and further work recommendations.
2. Materials and Methods
To implement digital image processing, the computing equipment used is of the commercial type with an AMD Ryzen 5 5500U processor, Radeon Graphics 2.10 GHz, and 16.0 GB RAM. Regarding the material evaluated, the images and videos stem from GoPro HERO4 Black and Ambarella MK-1 cameras, and are provided by the ICIMAP [
32]. The programming language requirements and algorithm download links for reproducibility purposes are described in
Table 1.
An overview diagram of the method for obtaining the keypoints and processing times, which starts to improve original monocular images with Sea-Thru, is presented in
Figure 1. The available code for Sea-Thru reduces the original size of images to optimize computational resources. For comparative purposes, the original images are resized to the Sea-Thru outcome dimension prior to the implementation of the DCP and CBFWB algorithms. Every single image-processing time is recorded by finishing the image improvement, and the outcome image is saved to apply a feature detector. Keypoints are detected using SIFT and ORB techniques and the results are recorded.
A schematic diagram of the multicriteria evaluation method is sketched in
Figure 2, where the increase in keypoints (obtained by the difference between the keypoints of the improved image and the keypoints of the original image resized) is the first criterion and processing time is the second criterion. Afterward, these criteria are normalized using the Min–Max method and weighted according to the research application. In the next step, a Weighted Sum Method score is quantified for each alternative, and, finally, the best choice corresponds to the highest score.
In this research, the underwater images processed with the Sea-Thru, DCP, and CBFWB algorithms comprise three experimental cases: (a) AMBA0150.jpg, an image captured with the Ambarella MK-1 camera; (b) a frame extracted from GOPR0223.mp4, a video taken with the GoPro 4 Black 4K; and (c) a sample of ten consecutive frames of the AMBA0064.mp4 video, captured with the Ambarella MK-1 camera. In the end, six different scenarios are shown to visualize the impact of weighting the criteria on the decision.