Next Article in Journal
Ambient Sensors for Elderly Care and Independent Living: A Survey
Previous Article in Journal
The SDN Approach for the Aggregation/Disaggregation of Sensor Data
Previous Article in Special Issue
Aerial Mapping of Forests Affected by Pathogens Using UAVs, Hyperspectral Sensors, and Artificial Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UAVs, Hyperspectral Remote Sensing, and Machine Learning Revolutionizing Reef Monitoring

1
Queensland University of Technology, 2 George St, Brisbane, QLD 4000, Australia
2
Research Engineering Facility, Institute for Future Environments, Queensland University of Technology, 2 George St, Brisbane, QLD 4000, Australia
3
Environment and Sustainability Institute, University of Exeter, Penryn, Cornwall TR10 9FE, UK
4
Institute for Advanced Study, Wissenschaftskolleg zu Berlin, Wallotstrasse 19, 14193 Berlin, Germany
5
Institute for Future Environments, Robotics and Autonomous Systems, Queensland University of Technology, 2 George St, Brisbane, QLD 4000, Australia
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(7), 2026; https://doi.org/10.3390/s18072026
Submission received: 18 April 2018 / Revised: 17 May 2018 / Accepted: 13 June 2018 / Published: 25 June 2018
(This article belongs to the Special Issue UAV or Drones for Remote Sensing Applications)

Abstract

:
Recent advances in unmanned aerial system (UAS) sensed imagery, sensor quality/size, and geospatial image processing can enable UASs to rapidly and continually monitor coral reefs, to determine the type of coral and signs of coral bleaching. This paper describes an unmanned aerial vehicle (UAV) remote sensing methodology to increase the efficiency and accuracy of existing surveillance practices. The methodology uses a UAV integrated with advanced digital hyperspectral, ultra HD colour (RGB) sensors, and machine learning algorithms. This paper describes the combination of airborne RGB and hyperspectral imagery with in-water survey data of several types in-water survey of coral under diverse levels of bleaching. The paper also describes the technology used, the sensors, the UAS, the flight operations, the processing workflow of the datasets, the methods for combining multiple airborne and in-water datasets, and finally presents relevant results of material classification. The development of the methodology for the collection and analysis of airborne hyperspectral and RGB imagery would provide coral reef researchers, other scientists, and UAV practitioners with reliable data collection protocols and faster processing techniques to achieve remote sensing objectives.

1. Introduction

Coral reefs are under pressure, and as of 2016, 34.8% of the Great Barrier Reef has been affected by coral bleaching, and other reefs worldwide are also experiencing degradation due to industrial and environmental events (e.g., cyclones) [1]. Large scale monitoring of coral reefs is an expensive, time-consuming, and challenging task, which utilises various mediums (satellite, airplane, boat and both manned and unmanned in-water surveys). Current surveillance practice for reef monitoring is to use satellites, ship-based multibeam sonar (MBES), plane based airborne surveying, or visually inspecting for signs of bleaching damage [2,3]. Over the last decades, the commercial satellite-based remote sensing capability has been significantly improved [4]. Now, with the best ground sampling distance (GSD) of 30 cm for panchromatic, and 1.2 m for multispectral data, satellite imagery remains the best-suited option for large-scale monitoring of entire reef. The satellite surveys, however, have some significant limitations, due to cloud cover and to camera resolution hindering the ability to distinguish finer details of reef ecosystems. Indeed, the approach struggles to provide marine researchers and biosecurity managers with often required benthic conditions on the sub-centimetre level GSD.
Recent advances in remote sensed imagery and geospatial image processing using unmanned aerial systems (UASs) can enable rapid and ongoing monitoring tools for coral bleaching detection and surveillance. This paper describes a UAS-based remote sensing methodology to increase the efficiency of existing surveillance practices, to detect coral type and coral bleaching, by combining in-water spectral observations and automated classification techniques to determine levels of benthic cover and coral growth in shallow water reef regions. The methodology is evaluated on its ability to identify differences between coral genera, bleaching level, and water depth, with an in-depth accuracy assessment to give insights for further refinements. This shows that UAS-based hyperspectral remote sensing techniques can offer an efficient and cost-effective approach to mapping and monitoring reef habitats over large, remote, and inaccessible areas.
The technology used, the sensors, the UAV, and the flight operations are presented in Section 2 of the paper. Later, in Section 3, we discuss the processing workflow for each imagery dataset, and methods for combining multiple airborne with ground-based datasets. Section 4 presents results of correlation between the different processed datasets.

2. Materials and Methods

Our approach to generate a predictive reef monitoring system employs multiple stages (Figure 1). The first is data collection. This includes collection of airborne RGB and hyperspectral imagery, as well as in-water expert visual assessment of each coral specimen.
In the second stage, images are processed (using data in Table 1) to obtain the coral type, bleaching assessment, and reef orthomosaics. Hyperspectral scans are then orthorectified using multiple datacubes to produce a georeferenced hyperspectral snapshot of the reef. These orthorectified datacubes are processed to obtain radiance and reflectance, that are used in combination with expert visual coral type and bleaching assessment for signature extraction and calculation of indices of coral types and bleaching levels.
The third stage combines all the multiple sources of data into a single information system. To do this, individual coral genera are segmented, with the aim of creating a table (Table 2) that contains different attributes for each coral specimen within the reef.
  • Coral genus
  • Expert bleaching assessment (1 to 6 where 1 relates to severely bleached coral, 6—unbleached)
  • Depth
  • Notes about surroundings (e.g., coral size, proximity to other coral)
  • Latitude
  • Longitude
Once the table is populated with georeferenced data, the spectral signatures are then extracted to generate a correlation between coral genus and each bleaching level. Results obtained in the correlation analysis are the foundation for the development of a preliminary coral bleaching detection model, which is followed by an evaluation and accuracy assessment to obtain final reef belching detection model.

2.1. In-Water Surveys

Coral is categorised by the Australian Institute of Marine Science (AIMS) into one of six levels of bleaching through visual inspection, where level one is severely bleached and level six an unbleached healthy coral specimen. Due to water’s optical properties and algal growth associated with reef ecosystems, the spectral properties of a coral sample can vary drastically based on the current conditions, genus, and depth of the area.
Our approach was to use survey test sites where experts identified various genera of coral at varying levels of bleaching across a segment of reef, as a reference for approximation of optical variances due to water and algae. Figure 2 and Figure 3 show examples of coral bleaching data obtained at different in-water survey locations, and after each point is identified, they are then classified based on genus, location (longitude and latitude), depth, and bleaching level.

2.2. Airborne Methods, UAV, and Sensors

The UAV used was an S800 EVO Hexacopter (DJI-Innovations, Shenzhen, China) that weighs 6.0 kg with motors, propellers, electronic speed controllers (ESCs), control and navigation avionics, and 16,000 mAh battery. The frame is fitted with a retractable undercarriage, providing a sensor field of view clear of obstacles. The UAV has a recommended maximum take-off weight of 8 kg, thus allowing 2 kg for sensor payload. A WooKong-M flight controller provides the navigation and control systems of the UAV, and comes with a stabilization controller and GPS unit with inbuilt compass and inertial management unit (IMU). The flight controller has multiple autopilot modes to enable both remote control by operator and autonomous with position and altitude hold, and auto go home/landing with enhanced fail-safe.
In this work and methodology, we propose the use of high resolution RGB and hyperspectral UAV-based imagery, as they provide the required level of detail in capturing finer differences in coral bleaching signatures. A Canon EOS 5DS R (Canon, Tokyo, Japen) digital camera was used to capture high-resolution RGB images from the mission route, and assist in identification and monitoring of coral in the studied area using the imagery augmented by the camera’s GPS coordinates. The camera specifications include full frame (36 × 24 mm) CMOS 50.6-megapixel sensor, 28-mm Canon lens, and a single-frequency GPS receiver.
Hyperspectral imagery was acquired using a Headwall Nano-Hyperspec (Headwall Photonics Inc., Bolton, MA, USA) hyperspectral sensor. This sensor records datacubes of 274 spectral bands in the VNIR (Visible and Near-Infrared) range (400–1000 nm) with ~2.2 nm spectral interval and 5 nm spectral resolution (fwhm with 20 μm slit). This sensor is equipped with a calibrated f/1.8 4.8 mm Schneider lens, which results in 50.7 deg field of view over 640 pixels. The collected hyperspectral data cubes are synchronized with GPS/INS (Inertial Navigation System) positioning and orientation information to perform datacube orthorectification.
Image stabilization is also an important step, therefore, the hyperspectral sensor was integrated into a S800 UAV by a custom designed gimbal made by Queensland University of Technology (QUT) Research Engineering Facility (REF). This gimbal has 2-axis stabilization, that ensures seamless operation of the push-broom hyperspectral scanner in windy conditions. The gimbal design features carbon reinforcement over 3D printed structure, advanced dampening, brushless motors, and a BaseCam SimpleBGS 32 bit gimbal controller (BaseCam electronics, Riga, Latvia), with the total weight below 1 kg. Mounting the push-broom scanner on the gimbal proved to enhance camera performance in wind gust conditions by ensuring the minimal and consistent required overlaps between consecutive datacubes over the large study area. This leads to an increased overall flight efficiency in real life open environments. Figure 4a shows the S800 UAV with the hyperspectral sensor on the gimbal during one of the missions. The gimbal CAD design model is presented in Figure 4b.

2.3. Orthorectification of Hyperspectral Images

Data orthorectification (step 3) was conducted using Headwall SpectralView software, where the individual hyperspectral data cubes are geolocalised and stitched, with verification by the overlay of underwater contours of the reef. Figure 5 shows stitched images overlaid in Google Earth.
The hyperspectral data are then processed with spectral analysis tasks, such as MATLAB Scyllarus open source toolbox, Headwall SpectralView, and Scyven, to obtain radiance values which are then processed with a white reference image to generate the reflectance dataset (step 4). Once obtained, the reflectance dataset is corrected for water depth by running through a noise reduction algorithm (step 5), processed through Scyven using support vector machine automated material discovery, to cluster similar spectral signatures which are geo-referenced (step 6), before being extracted (step 7) for reef indices (step 8) and material classification (step 9).

2.4. Image Processing, Radiance, Reflectance, White Reference

Step 4 orthorectified images were processed with Headwall SpectralView to identify the corresponding per pixel illumination value. Once completed, the image then goes through a semi-automated process in which the image is imported into Scyven, matched with a corresponding white reference illumination pattern (generated from white reference file), then lastly processed to generate the reflectance value corresponding to the white reference.

2.5. Depth Correction

Estimating and correcting hyperspectral imagery for water depth (step 5) is a complex process. ENVI offers unique tools for processing characteristics of waterbody features, and this enables quick and easy depth approximation utilizing a bottom albedo-independent bathymetry algorithm that utilizes a log ratio transformer. The log ratio transformer calculates depth independent of the bottom material and brightness (bright sand or dark seabed vegetation). The use of the algorithm means that typical errors caused by varying floor composition [5] (Figure 6) will not affect the overall accuracy of the approximation.
The use of additional information from in-water surveys is required for refining and generating accurate depth results with a reasonable degree of error. Agisoft Photoscan gives an alternative approach to the water depth estimation by using the reconstruction of a DEM from multiple overlaying images, which allows depth approximation of larger areas with ease, but comes with a more limited result set where smaller depth changes can be missed.
To apply depth correction, the effects of the water column must be ascertained, and to do this, the coral samples at varying water levels are compared; Figure 7 shows the variances in spectral signatures between two samples of massive Porites with a bleaching level of four, with one at a depth of 1 m and the other at 2.3 m.
To derive the effect of the water column on the data, the log-transformed ratio between the green and blue wavelengths is used, otherwise, the effects of varying benthic cover are also taken into account [6]. Once the data are derived, it can be seen that this is tightly packed (Figure 8).
This data is log transformed, and used as input into the equation identified by Mishra, D. [7] (below), which outputs the percentage variance due to water column effects.
y = 18.353 x 2 + 10.805 x + 0.238
This identified that 6.4% max variation in the spectral signature was due to water column effects, which could be corrected using a noise reduction algorithm, and follows the trend which was identified by Zoffoli, M [8].

2.6. Coral Georeferencing

In-water survey data contain GPS coordinates that are then matched to imagery. This can be completed by using the orthorectified data to overlay GPS coordinates across the image, and then extrapolating the corresponding pixel coordinates; Figure 9 shows an example of extrapolated coral locations and signature extraction. This data is then stored in an attribute table (step 6), along with the depth data and other characteristics (see Table 2 for table extract).

2.7. Spectral Signature Extraction

Expert visual assessment and georeferencing (step 6) are used with the depth-corrected reflectance values (step 5) to extract the mean spectral signature for different corals with different levels of bleaching. Headwall SpectralView is used to identify pixels that contain corals of varying bleaching levels, for which the spectral values are then extracted with Scyven. Figure 9 demonstrates the survey points for spectral signature extraction once the images have been converted into reflectance (step 4), then preprocessed with an automated material discovery (part of the material classification phase, step 9) which groups objects of similar spectral signatures.
After each genus of coral has been identified using material discovery, they then need to be compared based on their spectral signatures to identify potential issues with material classification (e.g., spectral signatures of each bleaching level not discernible for other genus samples).
Figure 10 shows an example of the processed results where different regions are identified and grouped based on their spectral signatures. Brighter colours indicate regions with spectral patterns similar to that of the coral, sand, or other material identified in the region.

2.8. Reef Indices

The next step (step 8) is to calculate coral bleaching indices using the spectral signature extraction (step 7), as well as visual assessment (step 1), and georeferencing (step 6). The indices were selected to evaluate symptoms of bleaching such as higher spectral concentrations in the 400 nm and 750 nm regions. A list of the indices used, as well as their equations, is found in Table 1, where each genus of coral has been identified and then classified, based on differences in spectral samples comparing to the next closest spectral sample. Figure 11 shows an example of the spectral signature comparison used for coral bleaching index generation.

2.9. Classification

Classification is the next step, where the in-water survey data along with the spectral signature extraction are combined using ENVI to develop pixel-by-pixel analysis algorithms, and also, subpixel (spectral mixture analysis) approaches using support vector machine (SVM). SVM was chosen due to the high classification accuracy, ability to handle multiple predictor variables, in-built handling of cases where spectral samples cannot be completely separated, and its ability to handle large datasets. It has also been found that the accuracy of SVM could be further increased if the gamma of images is increased up to a maximum of 75% (variations between captured data brightness and ideal gamma level are expected). This change causes SVMs to be equivalent to naive Bayes classifiers; Zanaty [9] proposed the SVM kernel of Gaussian radial basis polynomials function, which was found to be nearly 10% more accurate than the standard radial bias SVM classification method. Even without this, SVM is equivalent to the multilayer perception neural network in terms of accuracy of result. SVM using radial bias algorithm (shown below in Equation (2)) is therefore the method used in this paper, within the ENVI system, for material classifications.
R a d i a l   B a s i s   F u n c t i o n :   K ( x i , x j ) = e x p   ( γ | | x i x j | | 2 ) , γ > 0
Scyven is also used to apply various noise reduction algorithms and data refinement through targeted unsupervised classification, to identify regions requiring additional accuracy refinements.

3. Field Experiments

3.1. Site and In-Water Surveys

The site surveyed on 15 March 2017 is at Pandora Reef located in Queensland, Australia (Figure 12). The Australian Institute of Marine Science (AIMS) monitors this site regularly, and advised its selection due to the diverse range of coral genera, bleaching levels, and varying water depths. This paper demonstrates the method for an area of Pandora Reef in which 64 in-water survey points were collected to act as ground truths for spectral signature extraction (step 7) and training for the material classification (Table 2).

3.2. In-Water Data Collection Survey

Table 2 summarizes the data which were collected from the in-water surveys on the 15 March 2017 at Pandora Reef. The following attributes were collected: date, reference image id, coral genus, bleaching level, depth, and additional notes about surroundings. These points were then mapped onto the orthorectified data, where its respective location is stored as a set of pixel coordinates. Figure 13 shows all the in-water survey locations mapped in Google Earth.

4. Results

4.1. Water Depth and Water Depth Extraction Results

Figure 6 shows the result from ENVI’s relative water depth tool. Application of the bathymetry method is able to determine relative water depth by utilizing a log ratio transform. Visual assessment shows that the results in Figure 6 are consistent with the researcher’s knowledge of the study area and the in-water surveys (Figure 3).

4.1.1. Orthomosaics

The orthomosaic of the entire Pandora Reef (Figure 14) assisted in water depth extraction and generic reef structure. The orthomosaic was generated from Phantom 3 Pro images and is a geospatial RGB aerial view of the reef, with 5.71 cm/px resolution in a format of 234 megapixel image over 62.3 Ha area. The area was covered within a 20 min flight. Approximate areas covered by the underwater surveys are highlighted by red (3 m depth, 0.1 Ha) and white (6 m depth, 0.1 Ha) polygons.
Processing Canon 5DS R images gives significantly better 9.91 mm/px resolution of the orthomosaic/3D mode/DEMl, also bringing larger dynamic range to imagery (Figure 15b), and fully customised lenses and shooting options.
Figure 15a shows an orthomosaic generated out of 395 50,6-megapixel Canon 5DS R images processed on medium quality settings, which resulted in a 1.33 gigapixel image (47,904 × 29,080) over an 8 hectare area of primary AIMS interest. These orthomosaics give researchers a method of monitoring benthic cover and geomorphological variances in benthic algae (Murfitt et al. [10]). Orthomosaics assist in identifying regions that are at greatest risk of degradation, and ascertain what the cause of this degradation is and what systems need to be employed to try and reduce/prevent further degradation in that region. The orthomosaics also give researchers a method for identifying in-water survey data points, and giving approximations of what coral is typically found in this region. For example, Palma et al. [11] found that small rocky outcrops with a regular shape were typically indicative of soft corals whereas large colonies in rocky regions typically indicated Acropora coral.
Airborne ultra-HD DSLR cameras provide researchers with a complex and fully adjustable tool resulting in very detailed snapshots of the marine environment. Larger dynamic range, higher resolution, more accurate focusing and white-balancing comes at the cost of a heavier and larger UAV to carry the sensor. We note also that other factors, such as correct camera settings and camera orientation relative to the sun and the object of interest, position of the sun, tide heights, sea state, and water torpidity, define the success, and sometimes are the most important factors for reef monitoring by remote sensing.

4.1.2. Photogrammetry Ocean Floor Digital Elevation Models

Figure 16 shows the overlaid example of the DEM generated of the ocean floor surface of a fragment of Pandora Reef reconstructed out of Phantom 3 Pro (22 cm/px resolution) and Canon 5DS R (3.96 cm/px resolution) imagery. Despite the presence of minor artefacts and anomalies due to the presence of the moving border between two different mediums (sea and air), we observed adequate level of detail and accuracy of reconstruction for shallow waters (up to 5–8 m deep).

4.2. Spectral Signature Extraction Results

Figure 10, Figure 17 and Figure 18 show an example of spectral signature extraction for one coral type. Figure 18 shows the results with Porites massive bleaching level one and six, Figure 18 with soft corals level 5, Porites massive level 4, Goniopora level 3, and Acropora level 2. These spectral similarities will be one of the greatest potential causes of error in the classification step (step 9) and require processing of each identified point to confirm that the classification is correct. This will be done by solely comparing the classifications in the bands where differences are discernible.

4.3. Classification Results

Material classification through Scyven, although very similar to ENVI, is done by taking the spectral fingerprint data (Figure 9, Figure 17 and Figure 18) and then running it through an automated process, which matches and groups sets of specific spectral data based on their similarity. Figure 19 shows the results of material classification applied to Porites massive coral with the other object classification omitted for clarity.
Figure 20 shows output of the automatic unsupervised material classification from ENVI, to highlight and show variations which are typically associated with Porites massive coral indices (Table 1). These data are utilized to show regions that require exclusion for reduction of false positives.
Once the data has been refined, it can then be classified using the extracted spectral signatures (step 7 Figure 1) and the reef indices generated (step 8 Figure 1), whilst also being careful and taking note of the regions that were highlighted by the unsupervised classifications. For example, in the case of Figure 20, the benthos is shown as light green, spume as light yellow, and the numerous additional misclassifications caused by the ill-classified sandbar. Figure 21 shows the results of this classification after refining the extracted spectral signatures using additional noise reduction algorithms in the form of spectral and spatial filtering, additional ground truth discrimination from unsupervised classification results, and removal of false positives highlighted by coral bleaching indices (Table 1) developed from Figure 20 and Table 2.
Table 3 Results of the accuracy assessment with labelled data material classification has been completed.
The classification uses all the information in the image for classifying the different coral types. The accuracy results in Table 3 show the effect of including images with spume during the image processing phase. Classifications which have been greatly affected have been marked with EA (excessive area) in the table. Figure 22 shows the improved results of refining the data by excluding those regions of spume (marked in Figure 6), and this is reflected in Table 4.

4.4. Index Results

Different indices were computed in Section 2.8 in order to highlight the symptoms of bleaching across various genera of coral. Figure 23 shows examples of two NDVI images in the green and yellow wavelengths (identified by indices as bands with greatest variances over multiple bleaching levels), as well as two coral bleaching indices calculated from the hyperspectral data for the Acropora and Turbinaria genera of coral. This comparison shows the overall difference between the index output and the NDVI, where the NDVI images gives a clear representation, primarily focusing on the healthiest coral. The indices are capable of showing the overall health of the reef by not only showing the healthiest parts, but also all the various genera of coral at different levels of bleaching, and give health approximations based on benthic cover.
The results of the mean spectral signature extraction show a trend between spectral reflectance and severity of bleaching, which is consistent across various genera of coral. Coral bleaching indices developed from Figure 12 allow for analysis of stress caused by bleaching, however, these indices are not always able to distinguish coral genus from bleaching level, and may only be useful in monitoring temporal changes. NDVI images give a more easily interpreted representation, but cannot distinguish between variances of different genera and only show limited reef health data.

4.5. Limitations

The methodology shown in this manuscript has limitations, and requires careful and detailed planning of in-water and airborne surveys that can affect the overall feasibility and accuracy of the classification. The ability to classify objects can be greatly distorted or made near impossible, depending on water depth and the methodology used. Lesser and Mobley [12] found that it is possible to conduct coral classification with hyperspectral imagery to depths of around 10–12 m; this, however, requires the use of the Lambert–Beer law, which requires additional data of measured photoabsorption from testing areas. In this work, we use a band ratio equation which drops the maximum classification depth to around 6–8 m, which is satisfactory for large-scale remote sensing in either inaccessible or near-inaccessible areas, and offers options for health monitoring, however, this is a limitation for remote sensing of reef sections that do not fit these constraints.
Water turbidity, tidal conditions, and weather conditions all play a part in affecting the overall quality of the captured data. Water turbidity can affect the ability to identify coral in regions and, depending on its level, can prevent the ability to identify any coral. Nonetheless, data with moderate levels of turbidity still have some value in showing the current level of benthos/symbiotic algae in the region. Tidal conditions can cause significant amounts of spume to be created, methods can be employed to remove sections affected, but their effectiveness is limited by the spumes’ location and, where classification is required, can cause serious issues with the overall accuracy of results.

5. Conclusions

The results show that airborne UAV-based hyperspectral imagery has the potential to detect coral bleaching over large areas across multiple coral genera, where accurate visual inspection may not be possible. The methods and workflow developed during this study enabled material classification using both supervised and unsupervised methods to gain results with an acceptable percentage of error. The results presented and coral indices generated will contribute to the generation of valuable information for reef protection by offering methods that highlight possible symptoms of bleaching in a timely manner. It could also be extrapolated to other areas of research in remote sensing, such as mineral exploration, biodiversity, and ecological assessment.

Author Contributions

F.G. and D.B. conceived and designed the field and airborne experiments, guided airborne surveys and quality assessed the collection of the UAV imagery. D.B. undertook the relevant airborne surveys and monitored in-water surveys. M.P., F.G., D.B. conceptualized the methodology in Figure 1. M.P. developed and implemented the methodology, imagery workflow and analysis. K.J.G. guided result conceptualization and project limitation identification. M.P. and D.B processed the data. M.P was the principal author whilst all authors contributed significantly to the whole paper.

Acknowledgments

This work was funded by the Institute for Future Environments at Queensland University of Technology (QUT), with logistics support by The Australian Institute of Marine Science (AIMS). We would also like to acknowledge the High Performance Computing and Research Support Group at QUT for the computational resources and services used in this work, Sam Noolan for his assistance with collecting in-water survey data, Research Engineering Facility (Gavin Broadbent, Dmitry Bratanov) at QUT for operation of the DJI S800 UAV and sensors in the data collection and assistance in the data processing. With special thanks been given to Kaveh Deilami, Kyle Moyle and Juan Sandino Mora for their assistance during this work.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

References

  1. Hughes, T.; Kerry, J.; Baird, A.; Connolly, S.; Dietzel, A.; Eakin, M.; Heron, S.; Hoey, A.; Hoogenboom, M.; Liu, G.; et al. Global warming transforms coral reef assemblages. Nature 2018, 556, 492–496. [Google Scholar] [CrossRef] [PubMed]
  2. AIMS.gov Resources: Coral Bleaching Events. 2016. Available online: http://www.aims.gov.au/docs/research/climate-change/coral-bleaching/bleaching-events.html (accessed on 10 April 2018).
  3. Anderson, D.; Armstrong, R.; Weil, E. Hyperspectral Sensing of Disease Stress in the Caribbean Reef-Building Coral, Orbicella faveolata—Perspectives for the Field of Coral Disease Monitoring. PLoS ONE 2013, 8, e81478. [Google Scholar] [CrossRef] [PubMed]
  4. Goodman, J.A.; Purkis, S.J.; Phinn, S.R. Coral Reef Remote Sensing. A Guide for Mapping, Monitoring, and Management; Springer: Dordrecht, The Netherlands, 2013. [Google Scholar]
  5. Hamylton, S.; Hedley, J.; Beaman, R. Derivation of High-Resolution Bathymetry from Multispectral Satellite Imagery: A Comparison of Empirical and Optimisation Methods through Geographical Error Analysis. Remote Sens. 2015, 7, 16257–16273. [Google Scholar] [CrossRef]
  6. Gordon, H.R.; Brown, O.B. Influence of bottom albedo on the diffuse reflectance of a flat homogeneous ocean. Appl. Opt. 1974, 13, 2153–2159. [Google Scholar] [CrossRef] [PubMed]
  7. Mishra R, D.; Narumalani, S.; Rundquist, D.; Lawson, M.; Perk, R. Enhancing the detection and classification of coral reef and associated benthic habitats: A hyperspectral remote sensing approach. J. Geophys. Res. Oceans 2007, 112. [Google Scholar] [CrossRef] [Green Version]
  8. Zoffoli, M.; Frouin, R.; Kampel, M. Water Column Correction for Coral Reef Studies by Remote Sensing. Sensors 2004, 14, 16881–16931. [Google Scholar] [CrossRef] [PubMed]
  9. Zanaty, E. Support Vector Machines versus Multilayer Perception in data classification. Egypt. Inform. J. 2012, 13, 177–183. [Google Scholar] [CrossRef]
  10. Grasmuek, M.; Eberli, G.; Viggano, D.; Correa, T.; Rathwel, G.; Lou, J. Autonomous underwater vehicle (AUV) mapping reveals coral mound distribution, morphology, and oceanography in deep water of the Straits of Florida. Oceans 2006, 33. [Google Scholar] [CrossRef] [Green Version]
  11. Murfitt, S.; Allan, B.; Bellgrove, A.; Rattray, A.; Young, M.; Lerodiaconou, D. Applications of unmanned aerial vehicles in intertidal reef monitoring. Sci. Rep. 2017, 7, 10259. [Google Scholar] [CrossRef] [PubMed]
  12. Lesser, M.; Mobley, C. Bathymetry, water optical properties, and benthic classification of coral reefs using hyperspectral remote sensing imagery. Coral Reefs 2007, 26, 819–829. [Google Scholar] [CrossRef]
Figure 1. Predictive detection model workflow.
Figure 1. Predictive detection model workflow.
Sensors 18 02026 g001
Figure 2. Example of in-water reference data with relevant bleaching level: 1 (severely bleached) to 6 (unbleached). Bleaching levels and genus from top left to bottom right: 5 Acropora, 4 Porites massive, 3 Acropora, 3 Goniopora, 2 Acropora, 1 Acropora. Courtesy of Sam Noonan (Australian Institute of Marine Science (AIMS)).
Figure 2. Example of in-water reference data with relevant bleaching level: 1 (severely bleached) to 6 (unbleached). Bleaching levels and genus from top left to bottom right: 5 Acropora, 4 Porites massive, 3 Acropora, 3 Goniopora, 2 Acropora, 1 Acropora. Courtesy of Sam Noonan (Australian Institute of Marine Science (AIMS)).
Sensors 18 02026 g002
Figure 3. Example of in-water reference data of Acropora with relevant bleaching level: 1 (severely bleached) to 6 (unbleached). Top left—6, bottom right—1. Courtesy of Sam Noonan (AIMS).
Figure 3. Example of in-water reference data of Acropora with relevant bleaching level: 1 (severely bleached) to 6 (unbleached). Top left—6, bottom right—1. Courtesy of Sam Noonan (AIMS).
Sensors 18 02026 g003
Figure 4. (a) Headwall Nano hyperspectral sensor onboard S800 UAV; (b) Custom-made 2-axis gimbal hosting the hyperspectral camera (SolidWorks 3D model).
Figure 4. (a) Headwall Nano hyperspectral sensor onboard S800 UAV; (b) Custom-made 2-axis gimbal hosting the hyperspectral camera (SolidWorks 3D model).
Sensors 18 02026 g004
Figure 5. Multi-orthorectification by Headwall SpectralView overlaid in Google Earth, polygonal regions (in red) indicate visible spume regions in close proximity to coral.
Figure 5. Multi-orthorectification by Headwall SpectralView overlaid in Google Earth, polygonal regions (in red) indicate visible spume regions in close proximity to coral.
Sensors 18 02026 g005
Figure 6. Depth approximation non-homogeneous composition (left—ENVI depth approximation; centreFigure 5 extract; right—Agisoft Photoscan DEM (Digital Elevation Module) depth approximation).
Figure 6. Depth approximation non-homogeneous composition (left—ENVI depth approximation; centreFigure 5 extract; right—Agisoft Photoscan DEM (Digital Elevation Module) depth approximation).
Sensors 18 02026 g006
Figure 7. Spectral variance between massive Porites coral samples with level 4 bleaching at depths of 1 m and 2.3 m.
Figure 7. Spectral variance between massive Porites coral samples with level 4 bleaching at depths of 1 m and 2.3 m.
Sensors 18 02026 g007
Figure 8. Spectral signature comparison between reflectance in blue and green bands of coral in different locations. Dots represent the ratio values between the green and blue spectral values, the orange line represents the trendline between all the data points.
Figure 8. Spectral signature comparison between reflectance in blue and green bands of coral in different locations. Dots represent the ratio values between the green and blue spectral values, the orange line represents the trendline between all the data points.
Sensors 18 02026 g008
Figure 9. Examples of spectral responses “fingerprints” of bleaching levels 1 to 6 for Acropora coral.
Figure 9. Examples of spectral responses “fingerprints” of bleaching levels 1 to 6 for Acropora coral.
Sensors 18 02026 g009
Figure 10. Coral identified with various levels of bleaching.
Figure 10. Coral identified with various levels of bleaching.
Sensors 18 02026 g010
Figure 11. Spectral signature comparison for reef index generation Alv1.
Figure 11. Spectral signature comparison for reef index generation Alv1.
Sensors 18 02026 g011
Figure 12. Pandora Reef site, Great Barrier Reef, QLD, Australia.
Figure 12. Pandora Reef site, Great Barrier Reef, QLD, Australia.
Sensors 18 02026 g012
Figure 13. Numbered in-water survey GPS coordinates at Pandora Reef shown through Google Earth.
Figure 13. Numbered in-water survey GPS coordinates at Pandora Reef shown through Google Earth.
Sensors 18 02026 g013
Figure 14. Photogrammetry orthomosaic of Pandora Reef (63 Ha) shows areas of day long in-water surveys performed by AIMS at the depth of 3 m (orange area) and 6 m (white area) 400% zoomed fragment on the bottom right (tide height 2.6 m).
Figure 14. Photogrammetry orthomosaic of Pandora Reef (63 Ha) shows areas of day long in-water surveys performed by AIMS at the depth of 3 m (orange area) and 6 m (white area) 400% zoomed fragment on the bottom right (tide height 2.6 m).
Sensors 18 02026 g014
Figure 15. (a) Photogrammetry orthomosaic of western part of Pandora Reef (8 Ha) shows 400% zoomed fragments of bleaching events and the AIMS’s Antares research boat on the bottom right (tide height 1.6 m). (b) Fragments of the orthomosaics from Canon 5DS R (left) and DJI Phantom 3 Pro imagery (right).
Figure 15. (a) Photogrammetry orthomosaic of western part of Pandora Reef (8 Ha) shows 400% zoomed fragments of bleaching events and the AIMS’s Antares research boat on the bottom right (tide height 1.6 m). (b) Fragments of the orthomosaics from Canon 5DS R (left) and DJI Phantom 3 Pro imagery (right).
Sensors 18 02026 g015aSensors 18 02026 g015b
Figure 16. Digital elevation model (DEM) of Pandora Reef generated from Phantom 3 Pro and Canon 5DS R (purple area) and in-water survey depths (black dots).
Figure 16. Digital elevation model (DEM) of Pandora Reef generated from Phantom 3 Pro and Canon 5DS R (purple area) and in-water survey depths (black dots).
Sensors 18 02026 g016
Figure 17. Example of spectral response “fingerprints” of bleaching levels 1 to 4 for Porites massive coral.
Figure 17. Example of spectral response “fingerprints” of bleaching levels 1 to 4 for Porites massive coral.
Sensors 18 02026 g017
Figure 18. Example of spectral response “fingerprints” of bleaching levels 2, 3, 4 and 5 for Acropora, Gonipora, Porites massive and soft corals, respectively, for spectral comparison.
Figure 18. Example of spectral response “fingerprints” of bleaching levels 2, 3, 4 and 5 for Acropora, Gonipora, Porites massive and soft corals, respectively, for spectral comparison.
Sensors 18 02026 g018
Figure 19. Porites massive material classification, with water classification omitted.
Figure 19. Porites massive material classification, with water classification omitted.
Sensors 18 02026 g019
Figure 20. Unsupervised material classification for Pandora Reef.
Figure 20. Unsupervised material classification for Pandora Reef.
Sensors 18 02026 g020
Figure 21. Support vector machine (SVM) material classification for Pandora Reef.
Figure 21. Support vector machine (SVM) material classification for Pandora Reef.
Sensors 18 02026 g021
Figure 22. Support vector machine (SVM) material classification for Pandora Reef with spume regions marked in Figure 5 excluded.
Figure 22. Support vector machine (SVM) material classification for Pandora Reef with spume regions marked in Figure 5 excluded.
Sensors 18 02026 g022
Figure 23. Full colour spectrum images blue to red intensity (red high), (a) NDVI yellow; (b) NDVI green; (c) Turbinaria index Tlv5.3; (d) Acropora index Alv1.2.
Figure 23. Full colour spectrum images blue to red intensity (red high), (a) NDVI yellow; (b) NDVI green; (c) Turbinaria index Tlv5.3; (d) Acropora index Alv1.2.
Sensors 18 02026 g023
Table 1. Coral bleaching indices used in the analysis of reef type and bleaching.
Table 1. Coral bleaching indices used in the analysis of reef type and bleaching.
Coral IndexEquation
Genus Classification (R540 − R575/(R450 + R586)
Bleaching Classification
Acropora
Bleaching Level 1 (Alv1)Alv1.1 = (R395 − R404)/(R395 + R404)
Alv1.2 = (R575 − R604)/(R575 + R604)
Alv1.3 = (R711 − R732)/(R711 + R732)
Bleaching Level 2 (Alv2)Alv2.1 = (R404 − R489)/(R404 + R489)
Alv2.2 = (R595 − R662)/(R595 + R662)
Bleaching Level 3 (Alv3)Alv3.1 = (R446 − R473)/(R446 + R473)
Alv3.2 = (R531 − R555)/(R531 + R555)
Alv3.3 = (R586 − R622)/(R586 + R622)
Bleaching Level 4 (Alv4)Alv4.1 = (R446 − R489)/(R446 + R489)
Alv4.2 = (R569 − R600)/(R569 + R600)
Alv4.3 = (R611 − R671)/(R611 + R671)
Bleaching Level 5 (Alv5)Alv5.1 = (R484 − R522)/(R484 + R522)
Alv5.2 = (R695 − R720)/(R695 + R720)
Bleaching Level 6 (Alv6)Alv6.1 = (R400 − R418)/(R400 + R418)
Alv6.2 = (R460 − R484)/(R460 + R484)
Alv6.3 = (R724 − R768)/(R724 + R768)
Porites Massive
Bleaching Level 1 (PLv1)PLv1.1 = (R437 − R473)/(R437 + R473)
PLv1.2 = (R680 − R737)/(R680 + R737)
Bleaching Level 2 (PLv2)PLv2.1 = (R411 − R473)/(R411 + R473)
PLv2.2 = (R640 − R671)/(R640 + R671)
Bleaching Level 3 (PLv3)PLv3.1 = (R429 − R473)/(R429 + R473)
PLv3.2 = (R576 − R640)/(R576 + R640)
Bleaching Level 4 (PLv4)PLv4.1 = (R406 − R418)/(R406 + R418)
PLv4.2 = (R533 − R582)/(R533 + R582)
Gonipora
Bleaching Level 3 (GLv3)GLv3.1 = (R409 − R477)/(R409 + R477)
GLv3.2 = (R640 − R722)/(R640 + R722)
Turbinaria
Bleaching Level 5 (TLv5)TLv5.1 = (R415 − R442)/(R415 + R442)
TLv5.2 = (R471 − R486)/(R471 + R486)
TLv5.3 = (R500 − R544)/(R500 + R544)
TLv5.4 = (R675 − R717)/(R675 + R717)
Soft Coral
Bleaching Level 5 (SLv5)SLv5.1 = (R429 − R444)/(R429 + R444)
SLv5.2 = (R506 − R544)/(R506 + R544)
SLv5.3 = (R577 − R604)/(R577 + R604)
SLv5.4 = (R662 − R708)/(R662 + R708)
Table 2. Fragment of the attribute table at Pandora Reef (14 of 64 in-water data collected shown).
Table 2. Fragment of the attribute table at Pandora Reef (14 of 64 in-water data collected shown).
Photo IDCoral TypeLv BleachedBleachedDepthNotesLatitudeLongitudePixel xPixel y
76Porites massive1Yes 1.5>18.8129146.426714891412
77Porites massive1Yes 18.8129146.426715071421
78Porites massive4No2.3 18.8130146.426815441567
79Porites massive4No2.3 18.8130146.426815291548
80Porites massive4No2.3 18.8130146.426815411551
81Goniopora sp.3No1.7 18.8132146.426815981733
82Goniopora sp.3No1.7 18.8132146.426816041729
83Goniopora sp.3No1.7 18.8132146.426816071727
84Acropora sp.2Yes Acropora plate. 18.8131146.426816041669
85Acropora sp.2Yes Acropora plate. 18.8131146.426816041666
86Acropora sp.3No1.4Acropora plate.18.8131146.426916591684
87Acropora sp.3No1.4 18.8131146.426916581675
88Porites massive3No 1.5>18.8131146.426816101666
89Porites massive3No 18.8131146.426816151675
90Porites massive3No 1.5>18.8131146.426916331636
Table 3. Material classification accuracy assessment for Figure 21 (EA = excessive area).
Table 3. Material classification accuracy assessment for Figure 21 (EA = excessive area).
Coral TypeBleaching LevelSignature Accuracy (%)Points FoundAccuracy (%)Found PixelsArea (%)Overall Accuracy (%)
Porites massive179.172/366.66791,1572.93266.67
Porites massive294.441/250.0030,0070.96550.00
Porites massive368.752/728.57155,5291.78628.57
Porites massive4100.004/580.00339,30510.9150 (EA)
Acropora sp.1100.007/1353.846173,8105.59153.85
Acropora sp.246.672/450.0012480.04046.67
Acropora sp.3100.002/2100.0038,7291.246100.00
Acropora sp.4100.002/2100.0041,6181.339100.00
Acropora sp.5100.003/1421.42832,1191.03321.43
Acropora sp.6100.002/2100.00115,8793.7280 (EA)
Soft coral544.002/2100.00392,56912.6280 (EA)
Turbinaria sp.573.912/2100.0027630.08973.91
Table 4. Material classification accuracy assessment for Figure 22 at Pandora Reef where regions with spume region marker in Figure 5 are excluded.
Table 4. Material classification accuracy assessment for Figure 22 at Pandora Reef where regions with spume region marker in Figure 5 are excluded.
Coral TypeBleaching LevelSignature Accuracy (%)Points FoundAccuracy (%)Found PixelsArea (%)Overall Accuracy (%)
Porites massive188.792/2100.00238,9986.63688.79
Porites massive289.322/2100.0057,3231.59289.32
Porites massive396.321/1100.00221,2036.14296.32
Porites massive488.133/3100.0045,3861.26088.13
Acropora sp.196.544/4100.001,130,66331.3960 (EA)
Acropora sp.493.542/2100.0055,0851.53093.54
Acropora sp.590.753/3100.00477,66313.2640 (EA)
Acropora sp.690.272/2100.00214,0465.94490.27

Share and Cite

MDPI and ACS Style

Parsons, M.; Bratanov, D.; Gaston, K.J.; Gonzalez, F. UAVs, Hyperspectral Remote Sensing, and Machine Learning Revolutionizing Reef Monitoring. Sensors 2018, 18, 2026. https://doi.org/10.3390/s18072026

AMA Style

Parsons M, Bratanov D, Gaston KJ, Gonzalez F. UAVs, Hyperspectral Remote Sensing, and Machine Learning Revolutionizing Reef Monitoring. Sensors. 2018; 18(7):2026. https://doi.org/10.3390/s18072026

Chicago/Turabian Style

Parsons, Mark, Dmitry Bratanov, Kevin J. Gaston, and Felipe Gonzalez. 2018. "UAVs, Hyperspectral Remote Sensing, and Machine Learning Revolutionizing Reef Monitoring" Sensors 18, no. 7: 2026. https://doi.org/10.3390/s18072026

APA Style

Parsons, M., Bratanov, D., Gaston, K. J., & Gonzalez, F. (2018). UAVs, Hyperspectral Remote Sensing, and Machine Learning Revolutionizing Reef Monitoring. Sensors, 18(7), 2026. https://doi.org/10.3390/s18072026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop