Next Article in Journal
Comparative Study of Marine Ranching Recognition in Multi-Temporal High-Resolution Remote Sensing Images Based on DeepLab-v3+ and U-Net
Next Article in Special Issue
Quantitative Short-Term Precipitation Model Using Multimodal Data Fusion Based on a Cross-Attention Mechanism
Previous Article in Journal
Assessment of L-Band SAOCOM InSAR Coherence and Its Comparison with C-Band: A Case Study over Managed Forests in Argentina
Previous Article in Special Issue
Coupled Tensor Block Term Decomposition with Superpixel-Based Graph Laplacian Regularization for Hyperspectral Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SealNet 2.0: Human-Level Fully-Automated Pack-Ice Seal Detection in Very-High-Resolution Satellite Imagery with CNN Model Ensembles

by
Bento C. Gonçalves
1,2,*,
Michael Wethington
1 and
Heather J. Lynch
1,2
1
Department of Ecology and Evolution, Stony Brook University, Stony Brook, NY 11794-5245, USA
2
Institute for Advanced Computational Sciences, Stony Brook University, Stony Brook, NY 11794-3365, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(22), 5655; https://doi.org/10.3390/rs14225655
Submission received: 20 September 2022 / Revised: 1 November 2022 / Accepted: 4 November 2022 / Published: 9 November 2022

Abstract

:
Pack-ice seals are key indicator species in the Southern Ocean. Their large size (2–4 m) and continent-wide distribution make them ideal candidates for monitoring programs via very-high-resolution satellite imagery. The sheer volume of imagery required, however, hampers our ability to rely on manual annotation alone. Here, we present SealNet 2.0, a fully automated approach to seal detection that couples a sea ice segmentation model to find potential seal habitats with an ensemble of semantic segmentation convolutional neural network models for seal detection. Our best ensemble attains 0.806 precision and 0.640 recall on an out-of-sample test dataset, surpassing two trained human observers. Built upon the original SealNet, it outperforms its predecessor by using annotation datasets focused on sea ice only, a comprehensive hyperparameter study leveraging substantial high-performance computing resources, and post-processing through regression head outputs and segmentation head logits at predicted seal locations. Even with a simplified version of our ensemble model, using AI predictions as a guide dramatically boosted the precision and recall of two human experts, showing potential as a training device for novice seal annotators. Like human observers, the performance of our automated approach deteriorates with terrain ruggedness, highlighting the need for statistical treatment to draw global population estimates from AI output.

1. Introduction

Pack-ice seals (crabeater seals (Lobodon carcinophaga), Weddell seals (Leptonychotes weddelli), leopard seals (Hydrurga leptonyx), and Ross seals (Omnatophoca rossi)) play a prominent role as key Antarctic marine predators in the Southern Ocean (SO) ecosystem [1,2]. The importance of crabeater seals as a major consumer of Antarctic krill is well documented and the focus of considerable attention vis a vis seal conservation and management [3]. Antarctic krill, in turn, is not only a key component in biogeochemical cycles in the SO through its consumption of phytoplankton [4], but also sits at the base of the SO foodweb [5], attaining upwards of 300 million tonnes of biomass during its peak abundance in the austral summer [6]. These massive concentrations of krill sustain large populations of a wide range of marine predators, from marine mammals and seabirds to fish and squid. Understanding fluctuations in krill populations through the SO, and how they track sea ice conditions [7], is fundamental to gauge the health of the SO ecosystem—especially under anthropogenic climate change [7]. Through their trophic interaction with krill, and due to the inherent difficulties of estimating krill populations directly [6,8], surveying pack-ice seal populations provides a window to krill population dynamics. Among krill predators in the SO, pack-ice seals haul out on sea ice for breeding, molting, and predator avoidance [9] and are large enough to be individually counted using high-resolution satellite imagery [10]. For these reasons, pack-ice seals may serve as useful proxies for krill abundance if they can be surveyed at large spatial scales cost-effectively.
The role of pack-ice seals as an ecosystem health barometer in the SO has fueled several population estimation efforts. The earliest efforts to do so are reviewed in [11] but a more recent and highly structured attempt at estimating pan-Antarctic pack-ice seal population abundance was organized by the Scientific Committee for Antarctic Research under the umbrella of the International Antarctic Pack Ice Seals (APIS) program [12]. APIS was a large collaboration between six countries from 1994 to 2000 that employed vessel- and helicopter-based line transects to sample seal densities at several locations and times. Though such studies brought invaluable insights on seal biology (e.g., [13,14,15]), the spatial and temporal coverages are limited to draw continent-wide population estimates. Aerial transect-based monitoring programs such as APIS, especially in remote locations such as the SO, are not only resource-intensive [12] but dangerous for field biologists [16], making them unfeasible as means to repeatedly survey pack-ice seal populations.
Remote sensing imagery from very-high-resolution (VHR) sensors provides a cheaper and safer alternative to survey several polar animal populations [17], and has been demonstrated to work for pack-ice seals [10] and several other large marine mammal populations (e.g., southern elephant seals [18], walrus [19], and whales [20]). Manually annotating pack-ice seals in VHR imagery, however, is not only laborious but also extremely difficult given variations in lighting, substrate conditions, and imagery post-processing artifacts. These difficulties, allied with a scarcity of trained observers and the non-negligible cost of commercial VHR imagery, hamper our efforts to survey imagery manually at scale. Citizen science efforts mitigate these limitations by crowd-sourcing large areas across several volunteers (e.g., [21]). In order to draw useful ecological insights, however, robust sampling designs with overlapping predictions are needed to offset annotator inexperience [22,23]. In an effort by LaRue and collaborators [24], the Satellite Over Seals project (SOS), a 3-year citizen science campaign with >300,000 volunteers, enabled a Weddell seal survey covering the entirety of Antarctic fast-ice VHR imagery for the year of 2011, and the most comprehensive global estimate for the species to date [25]. Without a massive scale-up, even such large-scale campaigns will fail to keep pace with the volume of imagery collected over the Southern Ocean [26]. For these reasons, computer automation seems a necessary component of any long-term monitoring program.
The field of computer vision (CV), particularly deep learning approaches powered by modern GPUs and large annotated datasets (e.g., [27]), has demonstrated the potential for assisting in several laborious visual tasks across different application areas (e.g., agriculture [28], construction [29], and medicine [30]). With the popularization of commercial satellites [31], extremely large amounts of remote sensing imagery are amassed on a daily basis. In order to convert these imagery into actionable insights, CV solutions for remote sensing imagery have become commonplace (e.g., sea ice segmentation and classification [32,33,34]). In niche applications such as detecting wildlife, however, the lack of expert annotators to create large training datasets imposes a barrier to successfully applying such methods [35]. Nonetheless, there have been several proof-of-concept studies on the applicability of CV on VHR to detect/measure whales [20], pack-ice seals [10], and penguin colonies [36]. Among the difficulties of moving from proof-of-concept studies on wildlife detection in VHR to fully automated survey programs are ensuring that conditions at which detection models are trained and validated capture the full range of scenarios encountered when making predictions.
Here, we present a human-level, fully automated solution to detect pack-ice seals in VHR imagery, built upon our previous proof-of-concept study [10]. Our pipeline works by pre-processing Antarctic coastline VHR imagery through a sea ice segmentation model [32] that narrows down candidate imagery to scenes with relevant sea ice substrate. These selected scenes are scanned for seals using an ensemble of CNN models and ultimately converted to a database of geolocated predicted seals. Models in our pipeline were trained on a large annotated dataset obtained and curated over the course of five years. This extensive dataset and a massive deployment of GPU resources allowed us to train, validate, and test a wide range of different model solutions. Using annotations from an experienced observer as the basis for comparison, our automated approach outperforms two human observers with >100 h of experience when faced with novel VHR imagery randomly sampled from the existing imagery collection.

2. Materials and Methods

Our seal detection pipeline employs an ensemble of CNN models to derive georeferenced seal locations from very-high-resolution satellite imagery (Figure 1). Detailed information on our imagery datasets, CNN model training, and ensembling techniques can be found in the sections below.

2.1. Imagery and Data Annotation

For training, validation, model selection, model ensembling, and out-of-sample performance estimation, we employed three different annotated datasets (Figure 2): a training/validation set, an expert-selected test set, and a randomly selected test set. All three datasets comprised panchromatic Worldview-3 high-resolution satellite images (each one referred to as a ‘scene’) from the Antarctic coastline with an on-nadir resolution of 0.34 m/pixel. The annotation process consisted of manually browsing through Antarctic coastline imagery at a scale at which individual seals are detectable; at the centroid of each putative seal found, a geolocated point was added to a GIS (geographic information system) spatial point database (i.e., ESRI ® shapefile). We used a double-observer approach to create a consensus test dataset for out-of-sample performance estimation and model selection. The training/validation and test datasets used in this study were extracted from a set of 38 panchromatic scenes covering 8719.82 km2. We selected training and test scenes that would represent a comprehensive range of environmental and sensor conditions, including images captured over a range of off-nadir angles, lighting conditions, and cloud covers. Our training/validation and test sets represent a significant expansion and update of the datasets used in [10]. In addition to eliminating any putative seals that were re-classified on further consideration, we annotated several new scenes for training and validation. We also changed our hard-negative sampling strategy from extracting crops at polygons that marked locations without seals to extracting random crops that did not overlap with any seal annotations in scenes for which we had seal annotations. This new hard-negative sampling approach reflects the introduction of a sea ice detection step prior to detecting seals, allowing us to focus on areas with sea ice conditions amenable to seals being hauled out and thus available for detection. We revised all hard-negative samples to remove potential false positives. To avoid a potential selection bias, we created an additional test set (heretofore, the ‘random crops test set’) composed of 300 non-overlapping 1 km2 crops from 100 randomly selected WV03 scenes from the Antarctic coastline identified as having at least 5% sea ice cover by a sea ice segmentation CNN [32]. These crops were annotated by three different observers (BG, MW, HJL) with varying levels of experience classifying seals in WV03 imagery.

2.1.1. Training/Validation Set

We employed non-overlapping training and validation sets to train seal detection CNNs via gradient descent and run a comprehensive hyperparameter search optimized for out-of-sample performance. Using non-overlapping groups of seals from scenes in the training/validation set, we randomly assigned 80% of those groups of seal annotations to model training and 20% to model validation. Centered on each seal location, we extracted a 768 × 768 cropped image (i.e., patch) and saved a binary mask for that crop with ‘1’ on pixels at seal centroids and ‘0’ elsewhere. In addition, we recorded the total number of seals present in each patch. Since several seals may overlap in a single training patch, sampling training images at random would result in a bias towards seals situated within groups of seals (which are easier to detect [10]). To address that, we designed a weighted sampler that ensures that every seal is equally likely to be represented during training by down-weighting the probability of sampling individual seals based on the number of seals found within a radius of 50 m, making solitary seals and seals in larger groups just as likely to be represented during training. For training models that require bounding box annotations (i.e., instance segmentation and object detection models), we generated 11 × 11 bounding boxes centered on each seal. For each scene in training/validation sets, we extracted 300 non-overlapping ‘hard-negative’ 768 × 768 patches by randomly drawing crops from regions without any seal annotation. To ensure that no false positives existed in negative patches, we manually reviewed each negative patch and excluded those potentially containing seals. The final training/validation set was composed of a total of 8735 patches with seal annotations and 7750 hard-negative patches.

2.1.2. Expert-Selected Test Set

Our expert-selection test set builds on the approach from our previous study [10] by adding annotation revisions and replacing the original negative scenes with scenes that better reflect our new pipeline design with its sea-ice detection pre-processing step. We employed scene-wide annotations here to provide realistic out-of-sample metrics of model predictive performance in production (i.e., when predicting, a detection model has to go through entire scenes). This test set was used for both CNN model selection and model ensemble training. Moreover, we used predictions from the 10 top-performing models in the test scenes along with consensus annotations to train model ensembles that reclassify each point as a seal or a false positive based on the prediction from each model for that specific point (see below).

2.1.3. Random Crops Test Set

We used a set of 300 non-overlapping 1 km2 crops from 100 randomly selected WV03 scenes to validate models, optimize model ensembles, and remove selection bias from out-of-sample performance estimates. The 100 WV03 scenes were sampled at random from a set of 1948 scenes obtained by classifying a suite of 14,872 WV03 scenes through our sea ice segmentation model [32] and eliminating from consideration those with <5% (predicted) sea ice cover. Because the vast majority of imagery contains neither seals nor features such as rocks and/or shadows that could be confounded with seals, we used our top-10 model ensemble along with a stratified sampling approach to select the 300 crops for the test set; specifically, we selected 100 crops where all 10 models predicted one or more seals (seals very likely present), 100 crops where none of the models predicted any seals (seals very likely absent), and 100 crops where there was disagreement among the models as to the presence of seals in the scene (seal presence ambiguous). To create a consensus dataset representing manual (human) annotation, three observers (BG, MW, and HJL) independently annotated these 300 crops. While all three observers had considerable experience annotating seals in WV03 imagery, there was a gradient in the amount of experience (BG > MW > HJL); so, the consensus annotation used to represent ‘truth’ (in the absence of true ground validation, which is impossible in this scenario) was constructed by having the most experienced observer (BG) review and edit (as appropriate) the union of all manual annotations and high-probability CNN predictions. To explore the benefits of AI-guided annotation, each observer had a 50% chance of having access to AI help in the form of ensemble model predictions with their associated probability. Finally, we used the random crops test set as a tool to calibrate and evaluate model ensembles on out-of-sample annotations.

2.2. CNN Training and Validation

CNN-model architecture choice focused on established rather than state-of-the-art methods to favor explainability, comparison with other studies, and ease of implementation instead of pure predictive performance. Seal detection CNNs tested in this study were designed for three different tasks: object detection (i.e., drawing bounding boxes around each object of interest and giving appropriate labels to each bounding box), instance-segmentation (i.e., object detection with segmentation masks inside bounding boxes), and semantic segmentation (i.e., labeling every pixel in the image). For object detection and instance segmentation, we tested Fast R-CNN [37] and MaskRCNN [38], respectively, as implemented in the native torchvision package [39]. In both cases, we modify the default anchor box sizes for predicted objects to a smaller size that better matches our ‘truth’ bounding boxes, and, in the case of MaskRCNN, we swapped the original binary cross-entropy (BCE) criterion from the segmentation loss for a region-based dice loss (see the section on loss functions below). For semantic segmentation models, we tested both the U-Net architecture [40], as implemented in the segmentation-models-pytorch package [41] but with an added regression head, and TransUnet, a transformer-based U-Net-like architecture, as implemented in the original paper [42], with the exception of an added regression head. To ensure a fair comparison between widely different CNN modalities, we used a unified validation metric across all models: f1-score between predicted and ‘truth’ seals in the validation set. Training epochs consisted of going through every training image exactly once, starting a validation round whenever 1/3 of the training images were processed, with a total of three validation rounds per training epoch. CNNs were trained with an AdamW optimizer [43] with a policy that reduces the learning rate by a factor of two whenever there is no improvement in terms of validation metrics for N consecutive validation rounds—where ‘N’ is a hyperparameter—and stops training whenever there is no improvement for 3 training epochs. Across all settings, we used 512 × 512 input images, either sampled at random from the training pool or processed sequentially from the validation pool, grouped into mini-batches with the greatest number of images allowed given constraints from the model architecture and GPU memory availability. All experiments were performed using pytorch [44] with mixed-precision training [45] on NVIDIA V100 GPUs with 32 GB of memory from the Bridges-2 supercomputer [46].

2.2.1. Data Augmentation

To make our models more robust to scale, positioning, illumination, and other potential confounding factors, we employed a data augmentation pipeline during training, tailored to take full advantage of random crops and rotations given the nature of our imagery and dimensions of our objects of interest. We intentionally extracted larger patches (768 × 768) than our model input size (512 × 512) to keep training images diverse and non-obvious, and to reduce a potential bias for detecting seals at the center of input images. Whenever a seal was no longer present in the original patch after applying a random crop, the count for that patch (based on the ‘truth’ dataset) was adjusted to reflect that. For training, we used two data augmentation strategies: (1) a simple approach with random crops, vertical and horizontal flips, random shifts in position, random re-scaling, random 90-degree rotations (i.e., 90, 180, or 270 degrees), and brightness and contrast shifts; (2) a more complex approach using the transforms listed above plus noise reduction and intensity shifts. Our training augmentations are integrated into the training loop using implementations from the Albumentations package [47]. As an additional step to make predictions more robust to orientation, we apply horizontal flips and 90, 180, and 270 degree rotations to each input image, and average out predicted masks and counts from all possible combinations.

2.2.2. Loss Functions

To train regression heads on semantic segmentation models (i.e., U-Net and TransUnet), we used Huber loss [48], whereas segmentation heads used either Focal loss [49], Dice loss [50], or a combination of both. For instance segmentation and object detection models (i.e., Fast R-CNN and Mask R-CNN), we used the default torchvision losses for the Region Proposal Network classifier, the Bounding box classifier, and the Bounding box coordinates regression. We swapped the original Mask R-CNN BCE loss for predicted masks with Dice loss since BCE is not suitable when few if any pixels fall into the positive class.

2.3. Hyperparameter Search and Model Selection

We tested a wide range of scenarios to find optimal combinations of hyperparameters for seal detection models according to f1-score [50] for the expert-selected test set (i.e., test f1-score), running a total of 1056 full-length experiments. For all models, we tested the impact of learning rate, the number of epochs without improvement that would trigger learning rate reduction, and the ratio of negative to positive images in the data loader. For semantic segmentation models, we tested the impact of the segmentation loss function (Dice loss, Focal loss, or mixed Dice and Focal losses), backbone architecture (Resnet34 [51] and EfficentNet [52] variants), relative weight for regression and segmentation losses, and rate of dropout [53] applied to regression heads. When measuring test f1-score for semantic segmentation models, we tested the potential of using a threshold on predicted counts to remove false positives. To analyze the relative impact of each hyperparameter on test f1-score, we fit a CatBoost regressor model using hyperparameter values from each trial as dependent variables and f1-score as the outcome variable and calculated the relative importance of each hyperparameter using Shapley [54] scores. To save processing time, experiments that underperformed in terms of the maximum validation f1-score (>0.7 for instance segmentation and >0.5 for object detection and instance segmentation) were not carried into the testing stage. After an initial set of 372 experiments with semantic segmentation models, we narrowed down our hyperparameter pool to speed up convergence. For the latter 251 experiments, we also used a range of thresholds to explore the impact of using predicted count as a post-processing step to remove false positives, i.e., for each threshold, predicted points on patches where the predicted count was smaller than the threshold were discarded before comparison with ground truth annotations.

2.4. Model Ensembling

Our examination of a large suite of models allowed us to deploy a model ensemble post-processing step. The first step to creating model ensembles was gathering the correspondent predicted counts and logits for predicted seal locations at the expert-selected test set and training/validation set using the 10 top-performing models in terms of test f1-score. Whenever an individual model did not predict a seal at a location where other model(s) did predict a seal, cells with model logits and counts for that location were left as missing values. Predicted counts and logits from each model were then used as dependent variables to predict whether each point was in fact a true seal according to the ‘truth’ annotations. We split our random crops test set between validation and testing to run a hyperparameter search for ensemble models, trained at binary classification for true-positive vs. false-positive seals, ranging from simpler linear models (i.e., logistic regression and ElasticNet [55]) to more intricate tree-based models (i.e., random forest, CatBoost [56], and XGBoost [57]). Though the training and validation sets are already captured by individual models, we added the potential usage of these annotations to train model ensembles as a hyperparameter. We ran 50 independent hyperparameter search studies with 1500 experiments each. An experiment in our hyperparameter search consisted of sampling a combination of hyperparameter values from posteriors, training an ensemble model using those hyperparameter values, and updating posteriors according to f1-score in the validation portion of the random crops test set. To measure the contribution of each individual model to ensemble predictions, we used relative feature importance for logits and predicted counts from each model, in the form of feature weights for linear models and Shapley scores for tree-based models. We used a Bayesian optimization routine [58], implemented in the optuna package [59], with multivariate normal priors for hyperparameters to find the best-performing ensemble model, using the f1-score of the validation half of the random crops test set as our metric. The full range of hyperparameter choices for ensemble models can be found in the project’s code repository.

2.5. Evaluation

During all training experiments, our validation metric is the instance-level (i.e., individual seal centroids) f1-score. We measured this metric directly for model ensemble predictions; however, for CNN outputs, we need to match each predicted seal with consensus annotations. Because different annotators may identify the seal centroid in slightly different locations, and given expected seal dimensions of roughly 2 m, we used a tolerance of 1.5 m to declare two seals a match. For semantic segmentation models, we did so by applying a sigmoid transform followed by a binary thresholding step, leaving us with seal mask polygons. We then extracted the centroid of each polygon and looked for a match with the centroids of the consensus dataset. For instance segmentation and object detection models, however, we simply extracted the centroid from each predicted bounding box for comparison with consensus centroids. We evaluated the out-of-sample performance of model ensemble predictions, the best individual performing models, the original SealNet model [10], and human observers against our random crops test set consensus annotations. Since model ensembles use the validation portion of the random crops test set for model selection, all models and model ensembles were evaluated on the test portion of the random crops dataset to ensure a fair comparison. For all evaluation steps, model predictions on land were masked out using a sea ice mask derived from the Antarctic Digital Database (ADD) high-resolution coastline polygons available on the Quantarctica project [60]. To evaluate the consistency of output model probabilities, we measured the correlation between the sum of logits around predicted seal centroids and their corresponding ‘truth’ label. Similarly, we measured the same correlation for ensemble models using model-derived logits and their corresponding ‘truth’ labels.

3. Results

Semantic segmentation models largely outperformed object detection and instance segmentation models (Figure 3) in terms of test f1-score (0.39 ± 0.08, 0.04 ± 0.02, and 0.04 ± 0.02, respectively), attaining a top test f1-score of 0.58. Our initial set of experiments with semantic segmentation models (Figure 4, marked in orange) showed that test-time-augmentation is beneficial in terms of test f1-score, and that mid-range backbone architectures in terms of complexity (i.e., EfficientNet-b0, b1, and b2) had a slight edge over the extremes (i.e., ResNet34 and EfficientNet-b3); thus, we turned test-time-augmentation on by default and focused on mid-range backbone architectures for the later part of hyperparameter search experiments. First phase results also hinted that our ranges for the ratio of negative to positive images in training batches, learning rate, regression head weight, regression head dropout, and learning rate scheduler patience (Figure 4, marked in orange) could be adjusted to speed up convergence on better-performing models. Though there was a slight edge for simple augmentations over complex augmentations and random sampling over weighted group sampling, we opted to keep both options for later experiments. Even with fewer iterations, final experiments (Figure 4, marked in teal) largely outperformed initial ones in terms of test f1-score (0.35 ± 0.09 vs. 0.26 ± 0.13). Using a threshold on predicted counts as post-processing to remove false positives showed an average increase of 0.04 ± 0.05 in terms of test f1-score. Applying a threshold based on predicted counts as a post-processing step dramatically changed the distribution of test f1-scores (Figure 5); moreover, virtually all best-performing models had increases in f1-score by applying post-processing via the predicted count threshold.
The best-performing study for ensemble models achieved an f1-score of 0.69 in the validation portion of the random crops tests set (Table 1). The correlation between model logits and ‘truth’ labels was consistently higher for ensemble models when compared with individual CNN models (Table 1). The vast majority of hyperparameter search studies (42 out of 50, binomial p-value: <0.001) converged on XGBoost tree ensembles as their model of choice, with varying internal settings. The single best-performing study and 2 out of the 10 best-performing studies, however, converged on CatBoost tree ensemble as their model of choice. All independent studies kept CNN model predictions from the validation set as training data and dropped those for the training set. Shapley values for feature importance on the best-performing models in independent studies (Figure 6) heavily favored features from the best-performing CNN in terms of f1-score in the random crops test set (CNN 3) and features regarding patch-level predicted counts, followed closely by model logits.

4. Discussion

Our best-fitting CNN ensembles (Figure 1) attain an f1-score of 0.71 on a randomly-sampled dataset, with double-observer coverage and no exposure during training or validation, outperforming two human observers with >100 h of experience and access to annotations from both a simpler ensemble (Table 1) and the previous SealNet CNN pipeline [10]. The improvement in predictive performance stems from three primary factors: (1) a larger and more carefully curated training dataset that focused on scenes with sea ice (Figure 2); (2) a comprehensive hyperparameter search study (Figure 2 and Figure 4), only feasible in a multi-GPU setting; and (3) a new methodology using binary thresholding and regression counts followed by a model ensemble post-processing step. Our SealNet 1.0 classifier used the regression output to dictate how many logit hotspots would be extracted from predicted segmentation masks [10]. Here, we use regression outputs as a post-processing step for segmentation masks, which removes many false positives and leads to an improved f1-score (Figure 5). This SealNet 2.0 approach is also preferable because it relies solely on pixel-level, centroid mask annotations for prediction, which hinge upon stronger supervision signals during training when compared with patch-level ‘true’ counts.
Surprisingly, though the problem at hand theoretically aligns better with instance segmentation/object detection frameworks, our experiments with MaskRCNN [38] and Fast R-CNN [37] showed lackluster results (Figure 3) when compared with U-Nets [40]—a considerably simpler semantic segmentation approach. The extremely poor precision scores obtained with these methods could derive from limitations for training without foreground objects, creating a bias for over-predicting seals. With our U-Net-based approach, we are not only able to train using background-only patches, but we can also find optimal ratios of patches with and without foreground objects to maximize the balance between precision and recall through a hyperparameter search (Figure 4). This capability could give an edge to U-Net-based and other semantic segmentation approaches in ‘needle-in-a-haystack’ problems, which are ubiquitous in object detection applications for remote sensing imagery (e.g., [20,61,62]). The importance of showing negative examples during training in this kind of setting is supported by the relatively high negative-to-positive ratio found in our best-performing models (Figure 4, panel d).
Our top-10 individual CNN models, surprisingly, have a slightly lower out-of-sample recall than the global average for phase 2 experiments (0.54 vs. 0.55); however, on average, they are able to attain dramatically higher precision (0.52 vs. 0.33). This emphasis on avoiding false positives is also present when we look at the extremely high correlation between out-of-sample precision and f1-score (r = 0.93) and the strong negative correlation between out-of-sample recall and f1-score (r = −0.43). The relatively low correlation between f1-scores on the expert-selected test set performance and the random crops test set (r = 0.49)—and the dramatic performance decrease from SealNet 1.0 [10] on our more diverse test set Table 1—illustrates the importance of designing comprehensive test suites and cautions against over-relying on performance estimates on limited test sets.
Having access to AI help in the form of output from a simple ensemble model leads to a substantial improvement in the f1-score of human observers, improving precision without sacrificing recall. Though we are not able to draw statistical insights given our limited observer pool, our results suggest that human supervision could be used as quality control for AI output, as in most human-in-the-loop AI approaches (HITL [63,64]), and may also be used to guide inexperienced observers on challenging detection/classification tasks such as ours.
Though our ensemble models consistently outperformed individual CNN models (Table 1), we found the performance boost to be too small to justify the added computational cost of running imagery through ensembles when compared with the best-performing individual CNN (CNN 3, f1-score 0.69 vs. ensemble 1, f1-score 0.71). This similarity in performance is not surprising given the pronounced impact of features coming from CNN 3 on the best-performing ensembles (Figure 6), with minor contributions from a few other CNN models. Moreover, though we had a diverse set of hyperparameters within our 10 best-performing models, they hinged on the same datasets and model architecture (U-Net), which may have contributed to the high redundancy in including features from multiple CNNs. In contrast, several successful cases of applying ensemble models to CV rely on merging widely different individual components (e.g., [65,66]). On the other hand, ensemble models consistently outperformed individual CNNs in terms of the correlation between model logits and true labels (Table 1), which makes them more desirable as an AI-guided annotation tool and could translate to a lower bias on novel input imagery.
While our approach performs extremely well in simple terrain, with large groups of seals, it does encounter difficulties with rough terrain and single seals (Figure 7). Notably, these settings also tend to be the most challenging for human observers, shown by the high correlation between AI and observer errors. These difficulties are unlikely to be surmounted directly by improvements in AI because we cannot reliable annotate ground-truth datasets in these settings. Although identifying the portions of VHR scenes where seals, if present, could be detected is a tractable problem for modern CV models, estimating seal densities in areas where they cannot be detected is a non-trivial problem and merits further investigation.
In addition to high out-of-sample performance (Table 1), when compared with other candidate sampling methods for surveying pack-ice seals (i.e., fixed-wing airplanes, helicopters, UAVs, and human expert-based VHR surveys), AI-based approaches showed an operational cost comparable to that of the cheapest option available (helicopter-based flight transect surveys) [26], even taking into account the considerable cost of purchasing commercial VHR imagery. Cost-efficiency aside, GPU-accelerated AI-based approaches generate orders of magnitude less emissions than any other surveying method mentioned above [26]. The success of citizen science campaigns such as SOS [24], however, show the potential to utilize regular citizen science surveys as a validation method for fully automated pack-ice seal detection pipelines, especially given the difficulty of covering every potential real-world scenario during model evaluation.

5. Conclusions

Our results show compelling evidence for the immediate applicability of CNN-based, fully automated approaches for pack-ice seal surveys in VHR imagery. Moreover, they highlight the importance of comprehensive hyperparameter search studies and diverse training and evaluation datasets when employing AI methods to address complex tasks such as Antarctic pack-seal annotation in VHR imagery. With the addition of a pre-processing step to select VHR scenes where seals, if present, could be found and regular random checks by human observers for quality control, our approach with CNN ensembles is capable of delivering reliable, continental-scale putative Antarctic pack-ice seal locations.

Author Contributions

Conceptualization, B.C.G. and H.J.L.; methodology, B.C.G.; software, B.C.G.; validation, B.C.G., H.J.L. and M.W.; data curation, B.C.G., H.J.L. and M.W.; writing—original draft preparation, B.C.G.; writing—review and editing, B.C.G., M.W. and H.J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the U.S. National Science Foundation (Award 1740595) and the U.S. National Aeronautics and Space Administration (Award 80NSSC21K1150). Geospatial support for this work provided by the Polar Geospatial Center under NSF-OPP awards 1043681 and 1559691.

Data Availability Statement

Annotation datasets with geolocated seals for training, validation, and test sets, along with a document code repository for model training, validation, and prediction, are available at the project’s GitHub page, accessed on 20 September 2022.

Acknowledgments

We acknowledge support and computational resources provided by the Stony Brook University Institute for Advanced Computational Sciences, the Rutgers RADICAL group, and staff of the Bridges-2 computing program.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
VHRvery-high-resolution (satellite imagery)
SOSouthern Ocean
CVcomputer vision
GISgeographic information system
ADDAntarctic Digital Database
CNNconvolutional neural network

References

  1. Forcada, J.; Trathan, P.N.; Boveng, P.L.; Boyd, I.L.; Burns, J.M.; Costa, D.P.; Fedak, M.; Rogers, T.L.; Southwell, C.J. Responses of Antarctic pack-ice seals to environmental change and increasing krill fishing. Biol. Conserv. 2012, 149, 40–50. [Google Scholar] [CrossRef] [Green Version]
  2. Southwell, C.; Bengtson, J.; Bester, M.; Blix, A.; Bornemann, H.; Boveng, P.; Cameron, M.; Forcada, J.; Laake, J.; Nordøy, E.; et al. A review of data on abundance, trends in abundance, habitat use and diet of ice-breeding seals in the Southern Ocean. CCAMLR Sci. 2012, 19, 49–74. [Google Scholar]
  3. Southwell, C.; Paxton, C.G.; Borchers, D.; Boveng, P.; De La Mare, W. Taking account of dependent species in management of the Southern Ocean krill fishery: Estimating crabeater seal abundance off east Antarctica. J. Appl. Ecol. 2008, 45, 622–631. [Google Scholar] [CrossRef] [Green Version]
  4. Cavan, E.; Belcher, A.; Atkinson, A.; Hill, S.L.; Kawaguchi, S.; McCormack, S.; Meyer, B.; Nicol, S.; Ratnarajah, L.; Schmidt, K.; et al. The importance of Antarctic krill in biogeochemical cycles. Nat. Commun. 2019, 10, 4742. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Trathan, P.N.; Hill, S.L. The importance of krill predation in the Southern Ocean. In Biology and Ecology of Antarctic Krill; Springer: Berlin/Heidelberg, Germany, 2016; pp. 321–350. [Google Scholar]
  6. Atkinson, A.; Siegel, V.; Pakhomov, E.; Jessopp, M.; Loeb, V. A re-appraisal of the total biomass and annual production of Antarctic krill. Deep. Sea Res. Part I Oceanogr. Res. Pap. 2009, 56, 727–740. [Google Scholar] [CrossRef]
  7. Reiss, C.S.; Cossio, A.; Santora, J.A.; Dietrich, K.S.; Murray, A.; Mitchell, B.G.; Walsh, J.; Weiss, E.L.; Gimpel, C.; Jones, C.D.; et al. Overwinter habitat selection by Antarctic krill under varying sea-ice conditions: Implications for top predators and fishery management. Mar. Ecol. Prog. Ser. 2017, 568, 1–16. [Google Scholar] [CrossRef] [Green Version]
  8. Atkinson, A.; Hill, S.L.; Pakhomov, E.A.; Siegel, V.; Anadon, R.; Chiba, S.; Daly, K.L.; Downie, R.; Fielding, S.; Fretwell, P.; et al. KRILLBASE: A circumpolar database of Antarctic krill and salp numerical densities, 1926–2016. Earth Syst. Sci. Data 2017, 9, 193–210. [Google Scholar] [CrossRef] [Green Version]
  9. Holdgate, M. Antarctic Ecology; Academic Press: Cambridge, MA, USA, 1970; Volume 1. [Google Scholar]
  10. Gonçalves, B.; Spitzbart, B.; Lynch, H. SealNet: A fully-automated pack-ice seal detection pipeline for sub-meter satellite imagery. Remote Sens. Environ. 2020, 239, 111617. [Google Scholar] [CrossRef]
  11. Øritsland, T. Biology and Population Dynamics of Antarctic Seals; Academic Press: Cambridge, MA, USA, 1970; pp. 361–366. [Google Scholar]
  12. Ackley, S.; Bengtson, J.; Bester, M.; Blix, A.; Bornemann, H.; Boveng, P.; Boyd, I.; Cameron, M.; Nordøy, E.; Plötz, J.; et al. The International Antarctic Pack Ice Seals (APIS) Program. Multi-disciplinary research into the ecology and behavior of Antarctic pack ice seals. Summary Update. In The Expert Group on Seals (EGS); Marthan, N.B., Brent, S.S., Eds.; Scientific Committee on Antarctic Research (SCAR): Cambridge, UK, 2006. [Google Scholar]
  13. Southwell, C. Satellite-linked dive recorders provide insights into the reproductive strategies of crabeater seals (Lobodon carcinophagus). J. Zool. 2004, 264, 399–402. [Google Scholar] [CrossRef]
  14. Rogers, T.L. Age-related differences in the acoustic characteristics of male leopard seals, Hydrurga leptonyx. J. Acoust. Soc. Am. 2007, 122, 596–605. [Google Scholar] [CrossRef]
  15. Gurarie, E.; Bengtson, J.L.; Bester, M.N.; Blix, A.S.; Cameron, M.; Bornemann, H.; Nordøy, E.S.; Plötz, J.; Steinhage, D.; Boveng, P. Distribution, density and abundance of Antarctic ice seals off Queen Maud Land and the eastern Weddell Sea. Polar Biol. 2017, 40, 1149–1165. [Google Scholar] [CrossRef]
  16. Sasse, D.B. Job-related mortality of wildlife workers in the United States, 1937–2000. Wildl. Soc. Bull. 2003, 31, 1015–1020. [Google Scholar]
  17. LaRue, M.; Knight, J. Applications of Very High-Resolution Imagery in the Study and Conservation of Large Predators in the Southern Ocean. Conserv. Biol. 2014, 28, 1731–1735. [Google Scholar] [CrossRef]
  18. McMahon, C.R.; Howe, H.; Van Den Hoff, J.; Alderman, R.; Brolsma, H.; Hindell, M.A. Satellites, the all-seeing eyes in the sky: Counting elephant seals from space. PLoS ONE 2014, 9, e92613. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Matthews, C.J.; Dispas, A.; Mosnier, A. Evaluating Satellite Imagery for Atlantic Walrus Odobenus Rosmarus Rosmarus Stock Assessment—A Pilot Study; Canadian Technical Report of Fisheries and Aquatic Sciences 3492; Fisheries and Oceans Canada: Winnipeg, MB, Canada, 2022.
  20. Borowicz, A.; Le, H.; Humphries, G.; Nehls, G.; Höschle, C.; Kosarev, V.; Lynch, H.J. Aerial-trained deep learning networks for surveying cetaceans from satellite imagery. PLoS ONE 2019, 14, e0212532. [Google Scholar] [CrossRef] [Green Version]
  21. Welvaert, M.; Caley, P. Citizen surveillance for environmental monitoring: Combining the efforts of citizen science and crowdsourcing in a quantitative data framework. SpringerPlus 2016, 5, 1–14. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Brown, E.D.; Williams, B.K. The potential for citizen science to produce reliable and useful information in ecology. Conserv. Biol. 2019, 33, 561–569. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Salas, L.A.; LaRue, M.; Nur, N.; Ainley, D.G.; Stammerjohn, S.E.; Pennycook, J.; Rotella, J.; Paterson, J.T.; Siniff, D.; Stamatiou, K.; et al. Reducing error and increasing reliability of wildlife counts from citizen science surveys: Counting Weddell Seals in the Ross Sea from satellite images. bioRxiv 2020. [Google Scholar] [CrossRef]
  24. LaRue, M.A.; Ainley, D.G.; Pennycook, J.; Stamatiou, K.; Salas, L.; Nur, N.; Stammerjohn, S.; Barrington, L. Engaging ‘the crowd’in remote sensing to learn about habitat affinity of the Weddell seal in Antarctica. Remote Sens. Ecol. Conserv. 2020, 6, 70–78. [Google Scholar] [CrossRef] [Green Version]
  25. LaRue, M.; Salas, L.; Nur, N.; Ainley, D.; Stammerjohn, S.; Pennycook, J.; Dozier, M.; Saints, J.; Stamatiou, K.; Barrington, L.; et al. Insights from the first global population estimate of Weddell seals in Antarctica. Sci. Adv. 2021, 7, eabh3674. [Google Scholar] [CrossRef]
  26. Gonçalves, B.C.; Wethington, M.; Lynch, H.J. Roadmap to a fully-automated, pan-Antarctic pack-ice seal monitoring program. 2022; in preparation. [Google Scholar]
  27. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  28. Tian, H.; Wang, T.; Liu, Y.; Qiao, X.; Li, Y. Computer vision technology in agricultural automation—A review. Inf. Process. Agric. 2020, 7, 1–19. [Google Scholar] [CrossRef]
  29. Xu, S.; Wang, J.; Shou, W.; Ngo, T.; Sadick, A.M.; Wang, X. Computer vision techniques in construction: A critical review. Arch. Comput. Methods Eng. 2021, 28, 3383–3397. [Google Scholar] [CrossRef]
  30. Esteva, A.; Chou, K.; Yeung, S.; Naik, N.; Madani, A.; Mottaghi, A.; Liu, Y.; Topol, E.; Dean, J.; Socher, R. Deep learning-enabled medical computer vision. NPJ Digit. Med. 2021, 4, 1–9. [Google Scholar] [CrossRef] [PubMed]
  31. Belward, A.S.; Skøien, J.O. Who launched what, when and why; trends in global land-cover observation capacity from civilian earth observation satellites. ISPRS J. Photogramm. Remote Sens. 2015, 103, 115–128. [Google Scholar] [CrossRef]
  32. Gonçalves, B.C.; Lynch, H.J. Fine-Scale Sea Ice Segmentation for High-Resolution Satellite Imagery with Weakly-Supervised CNNs. Remote Sens. 2021, 13, 3562. [Google Scholar] [CrossRef]
  33. Boulze, H.; Korosov, A.; Brajard, J. Classification of Sea Ice Types in Sentinel-1 SAR Data Using Convolutional Neural Networks. Remote Sens. 2020, 12, 2165. [Google Scholar] [CrossRef]
  34. Han, Y.; Liu, Y.; Hong, Z.; Zhang, Y.; Yang, S.; Wang, J. Sea ice image classification based on heterogeneous data fusion and deep learning. Remote Sens. 2021, 13, 592. [Google Scholar] [CrossRef]
  35. Weinstein, B.G. A computer vision for animal ecology. J. Anim. Ecol. 2018, 87, 533–545. [Google Scholar] [CrossRef]
  36. M Le, H.; Goncalves, B.; Samaras, D.; Lynch, H. Weakly labeling the antarctic: The penguin colony case. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 18–25. [Google Scholar]
  37. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  38. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  39. Marcel, S.; Rodriguez, Y. Torchvision the Machine-Vision Package of Torch. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; Association for Computing Machinery: New York, NY, USA, 2010; pp. 1485–1488. [Google Scholar] [CrossRef]
  40. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015. [Google Scholar] [CrossRef]
  41. Yakubovskiy, P. Segmentation Models Pytorch. 2020. Available online: https://github.com/qubvel/segmentation_models.pytorch (accessed on 7 November 2022).
  42. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021. [Google Scholar] [CrossRef]
  43. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  44. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
  45. Micikevicius, P.; Narang, S.; Alben, J.; Diamos, G.; Elsen, E.; Garcia, D.; Ginsburg, B.; Houston, M.; Kuchaiev, O.; Venkatesh, G.; et al. Mixed Precision Training. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, CA, 30 April–3 May 2018. [Google Scholar]
  46. Brown, S.T.; Buitrago, P.; Hanna, E.; Sanielevici, S.; Scibek, R.; Nystrom, N.A. Bridges-2: A Platform for Rapidly-Evolving and Data Intensive Research. In Practice and Experience in Advanced Research Computing; Association for Computing Machinery: New York, NY, USA, 2021. [Google Scholar]
  47. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and flexible image augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef] [Green Version]
  48. Huber, P.J. Robust Estimation of a Location Parameter. Ann. Math. Stat. 1964, 35, 73–101. [Google Scholar] [CrossRef]
  49. Lin, T.; Goyal, P.; Girshick, R.B.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. arXiv 2017. [Google Scholar] [CrossRef]
  50. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015. [Google Scholar] [CrossRef]
  52. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning; Chaudhuri, K., Salakhutdinov, R., Eds.; Proceedings of Machine Learning Research; PMLR: Cambridge, MA, USA, 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  53. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  54. Shapley, L.S. A Value for N-Person Games; RAND Corporation: Santa Monica, CA, USA, 1952. [Google Scholar] [CrossRef]
  55. Zou, H.; Hastie, T. Regularization and Variable Selection via the Elastic Net. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  56. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
  57. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef]
  58. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian Optimization of Machine Learning Algorithms. In Proceedings of the Advances in Neural Information Processing Systems; Pereira, F., Burges, C., Bottou, L., Weinberger, K., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
  59. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-Generation Hyperparameter Optimization Framework. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 2623–2631. [Google Scholar] [CrossRef]
  60. Matsuoka, K.; Skoglund, A.; Roth, G.; de Pomereu, J.; Griffiths, H.; Headland, R.; Herried, B.; Katsumata, K.; Le Brocq, A.; Licht, K.; et al. Quantarctica, an integrated mapping environment for Antarctica, the Southern Ocean, and sub-Antarctic islands. Environ. Model. Softw. 2021, 140, 105015. [Google Scholar] [CrossRef]
  61. Xue, Y.; Wang, T.; Skidmore, A.K. Automatic Counting of Large Mammals from Very High Resolution Panchromatic Satellite Imagery. Remote Sens. 2017, 9, 878. [Google Scholar] [CrossRef] [Green Version]
  62. Kellenberger, B.; Marcos, D.; Tuia, D. Detecting mammals in UAV images: Best practices to address a substantially imbalanced dataset with deep learning. Remote Sens. Environ. 2018, 216, 139–153. [Google Scholar] [CrossRef] [Green Version]
  63. Zanzotto, F.M. Human-in-the-loop artificial intelligence. J. Artif. Intell. Res. 2019, 64, 243–252. [Google Scholar] [CrossRef] [Green Version]
  64. Branson, S.; Wah, C.; Schroff, F.; Babenko, B.; Welinder, P.; Perona, P.; Belongie, S. Visual recognition with humans in the loop. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 438–451. [Google Scholar]
  65. Paladini, E.; Vantaggiato, E.; Bougourzi, F.; Distante, C.; Hadid, A.; Taleb-Ahmed, A. Two Ensemble-CNN Approaches for Colorectal Cancer Tissue Type Classification. J. Imaging 2021, 7, 51. [Google Scholar] [CrossRef]
  66. Pramanik, R.; Biswas, M.; Sen, S.; de Souza Júnior, L.A.; Papa, J.P.; Sarkar, R. A fuzzy distance-based ensemble of deep models for cervical cancer detection. Comput. Methods Programs Biomed. 2022, 219, 106776. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Simplified diagram for SealNet 2.0 showing the training of individual CNN models, model hyperparameter search, and model ensembling. Boxes colored in light-blue denote models, orange boxes denote datasets, and gray boxes denote model output. Thick black lines from datasets to models indicate model training. Dashed vertical lines indicate model selection steps. The best individual CNN models are trained on seal detection, including centroid segmentation and seal count regression, using a random search with training and validation, and the f1-score at the expert-selected test set for model selection. The best ensembling models are selected via Bayesian optimization, using top-10 CNN model predictions for the training set, validation set, and the expert-selected test set as dependent variables for training; true positive vs. false positive as the response variable; and the f1-score at the validation split from the random crops test set as a validation metric. Finally, we use the test portion of the random crops test set to estimate the out-of-sample performance of the best-performing model ensemble.
Figure 1. Simplified diagram for SealNet 2.0 showing the training of individual CNN models, model hyperparameter search, and model ensembling. Boxes colored in light-blue denote models, orange boxes denote datasets, and gray boxes denote model output. Thick black lines from datasets to models indicate model training. Dashed vertical lines indicate model selection steps. The best individual CNN models are trained on seal detection, including centroid segmentation and seal count regression, using a random search with training and validation, and the f1-score at the expert-selected test set for model selection. The best ensembling models are selected via Bayesian optimization, using top-10 CNN model predictions for the training set, validation set, and the expert-selected test set as dependent variables for training; true positive vs. false positive as the response variable; and the f1-score at the validation split from the random crops test set as a validation metric. Finally, we use the test portion of the random crops test set to estimate the out-of-sample performance of the best-performing model ensemble.
Remotesensing 14 05655 g001
Figure 2. Training/validation set (light-blue), expert-selected test set (magenta), and random-crop test set (orange). Polygons denote entire Worldview-3 panchromatic scenes for the training/validation set and the expert-selected test set, and 1 km2 crops within panchromatic Worldview-3 scenes for the random-crop test set.
Figure 2. Training/validation set (light-blue), expert-selected test set (magenta), and random-crop test set (orange). Polygons denote entire Worldview-3 panchromatic scenes for the training/validation set and the expert-selected test set, and 1 km2 crops within panchromatic Worldview-3 scenes for the random-crop test set.
Remotesensing 14 05655 g002
Figure 3. Expert-selected test set f1-score for hyperparameter search experiments from different computer vision domains. To ensure a fair comparison of models from different domains, semantic segmentation output masks are passed through a sigmoid transform and thresholded to extract mask centroids. Similarly, instance segmentation and object detection output bounding boxes are converted to centroids to evaluate matches with ‘truth’ centroids. To avoid unnecessary expenditure of GPU credits, experiments that did not perform well on the validation set (validation f1-score > 0.7 for semantic segmentation models and >0.5 for instance segmentation and object detection models) were not carried into testing.
Figure 3. Expert-selected test set f1-score for hyperparameter search experiments from different computer vision domains. To ensure a fair comparison of models from different domains, semantic segmentation output masks are passed through a sigmoid transform and thresholded to extract mask centroids. Similarly, instance segmentation and object detection output bounding boxes are converted to centroids to evaluate matches with ‘truth’ centroids. To avoid unnecessary expenditure of GPU credits, experiments that did not perform well on the validation set (validation f1-score > 0.7 for semantic segmentation models and >0.5 for instance segmentation and object detection models) were not carried into testing.
Remotesensing 14 05655 g003
Figure 4. Results from random search hyperparameter study for semantic segmentation models, with phases one (orange) and two (teal). For continuous hyperparameters—namely, regression head weight, learning rate, regression head dropout, and negative-to-positive ratio—each circle corresponds to an independent random search experiment. For the continuous hyperparameters—regression head weight, learning rate, negative-to-positive ratio, and the discrete parameters—backbone architecture, patience, and test time augmentation, we narrowed down the range of options to speed up convergence on a best-performing model. Apart from narrowing down hyperparameter priors, we also used phase 2 to test the potential benefit of using regression output as a post-processing step to remove potential false positives, and thus, the test f1-score for those experiments is calculated after post-processing. The width of violin shapes for categorical variables (panels ej) indicate the proportion of observations that take on those values, whereas X-axis values indicate all discrete states tested for that categorical variable. Phase 2 experiments tested a narrower range of discrete states for test time augmentation, patience, and backbone architecture; thus, there were no phase 2 results for some discrete states of those variables. Experiments for which test f1-score was below 0.01 are excluded from this plot.
Figure 4. Results from random search hyperparameter study for semantic segmentation models, with phases one (orange) and two (teal). For continuous hyperparameters—namely, regression head weight, learning rate, regression head dropout, and negative-to-positive ratio—each circle corresponds to an independent random search experiment. For the continuous hyperparameters—regression head weight, learning rate, negative-to-positive ratio, and the discrete parameters—backbone architecture, patience, and test time augmentation, we narrowed down the range of options to speed up convergence on a best-performing model. Apart from narrowing down hyperparameter priors, we also used phase 2 to test the potential benefit of using regression output as a post-processing step to remove potential false positives, and thus, the test f1-score for those experiments is calculated after post-processing. The width of violin shapes for categorical variables (panels ej) indicate the proportion of observations that take on those values, whereas X-axis values indicate all discrete states tested for that categorical variable. Phase 2 experiments tested a narrower range of discrete states for test time augmentation, patience, and backbone architecture; thus, there were no phase 2 results for some discrete states of those variables. Experiments for which test f1-score was below 0.01 are excluded from this plot.
Remotesensing 14 05655 g004
Figure 5. Side-by-side comparison between phase 2 experiment results with (1) and without (2) regression post-processing. Post-processing consisted of discarding predicted points within patches where regression output (i.e., predicted number of seals in patch) is smaller than a specified threshold. For each model, we explore a range of thresholds to obtain the maximum possible test f1-score obtained after post-processing, using the same optimal threshold across the entire expert-selected test set.
Figure 5. Side-by-side comparison between phase 2 experiment results with (1) and without (2) regression post-processing. Post-processing consisted of discarding predicted points within patches where regression output (i.e., predicted number of seals in patch) is smaller than a specified threshold. For each model, we explore a range of thresholds to obtain the maximum possible test f1-score obtained after post-processing, using the same optimal threshold across the entire expert-selected test set.
Remotesensing 14 05655 g005
Figure 6. Feature importance for ensemble model features grouped by CNN index (a) and feature type (b). Ensemble models were either CatBoost or XGBoost tree-based ensembles trained for classifying false-positive and true-positive seal detections. Models were derived from a hyperparameter search using a training set with the logits, predicted seal counts, and distances from crop centers from the output of 10 U-Net CNNs at the validation and expert-selected test sets. Feature importances were obtained via Shapley scores at the validation portion of the random crops test set.
Figure 6. Feature importance for ensemble model features grouped by CNN index (a) and feature type (b). Ensemble models were either CatBoost or XGBoost tree-based ensembles trained for classifying false-positive and true-positive seal detections. Models were derived from a hyperparameter search using a training set with the logits, predicted seal counts, and distances from crop centers from the output of 10 U-Net CNNs at the validation and expert-selected test sets. Feature importances were obtained via Shapley scores at the validation portion of the random crops test set.
Remotesensing 14 05655 g006
Figure 7. Prediction samples from our best ensemble model on eight scenes from the random crops test set. Samples were chosen to represent scenarios where the model predicts seal locations correctly (panels ad), fails to find existing seals (panels df), and annotates background objects as seals (panels g,h). Seals marked with teal circles indicate true positives (i.e., predicted seal present in consensus dataset), whereas purple and orange circles indicate false-negative and false-positive seals, respectively. Numbers next to circle annotations indicate the number of human observers that agreed with that particular model annotation. Number ‘1’ annotations unaccompanied by circles in panel e indicate edge cases where a human observer annotated a seal that was not present in the consensus dataset or model predictions. Imagery copyright Maxar Technologies Inc. 2022.
Figure 7. Prediction samples from our best ensemble model on eight scenes from the random crops test set. Samples were chosen to represent scenarios where the model predicts seal locations correctly (panels ad), fails to find existing seals (panels df), and annotates background objects as seals (panels g,h). Seals marked with teal circles indicate true positives (i.e., predicted seal present in consensus dataset), whereas purple and orange circles indicate false-negative and false-positive seals, respectively. Numbers next to circle annotations indicate the number of human observers that agreed with that particular model annotation. Number ‘1’ annotations unaccompanied by circles in panel e indicate edge cases where a human observer annotated a seal that was not present in the consensus dataset or model predictions. Imagery copyright Maxar Technologies Inc. 2022.
Remotesensing 14 05655 g007
Table 1. Out of sample performance for human observers (with and without the help of AI output), individual CNN models, and model ensembles measured at the random crops tests set. AI help is provided through the output of a simple ensemble model (i.e., an ElasticNet classifier, ‘ensemble naive’), with a color gradient based on model certainty. Because whether an observer will have access to AI help is assigned independently at random, human observers had different sets of imagery processed with the aid of AI output. U-Nets 1–5 are ordered according to their ranking based on f1-score in the expert-selected test set. SealNet 1.0 predictions were obtained with the original SealNet. Similarly, ensemble models 1–5 are numbered in descending order of f1-score on the validation portion of the random crops test set. We include the correlation between model logits and ‘truth’ labels as a measurement of consistency.
Table 1. Out of sample performance for human observers (with and without the help of AI output), individual CNN models, and model ensembles measured at the random crops tests set. AI help is provided through the output of a simple ensemble model (i.e., an ElasticNet classifier, ‘ensemble naive’), with a color gradient based on model certainty. Because whether an observer will have access to AI help is assigned independently at random, human observers had different sets of imagery processed with the aid of AI output. U-Nets 1–5 are ordered according to their ranking based on f1-score in the expert-selected test set. SealNet 1.0 predictions were obtained with the original SealNet. Similarly, ensemble models 1–5 are numbered in descending order of f1-score on the validation portion of the random crops test set. We include the correlation between model logits and ‘truth’ labels as a measurement of consistency.
Observer/ModelPrecisionRecallf1AI HelpArchitectureLogit Correlation
HJL0.350.560.43No--
HJL0.580.690.63Yes--
MW0.500.630.56No--
MW0.550.690.61Yes--
CNN 10.600.630.62-UnetEfficientNet-b10.54
CNN 20.450.670.54-UnetEfficientNet-b10.33
CNN 30.710.670.69-UnetEfficientNet-b10.60
CNN 40.440.670.53-UnetEfficientNet-b10.36
CNN 50.680.530.60-UnetEfficientNet-b00.53
SealNet 1.00.070.020.03-SealNet0.07
ensemble 10.800.640.71-CatBoost0.69
ensemble 20.740.670.70-XGBoost0.67
ensemble 30.640.700.67-CatBoost0.67
ensemble 40.730.670.70-XGBoost0.68
ensemble 50.730.660.70-XGBoost0.67
ensemble naive0.590.690.64-ElasticNet0.60
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gonçalves, B.C.; Wethington, M.; Lynch, H.J. SealNet 2.0: Human-Level Fully-Automated Pack-Ice Seal Detection in Very-High-Resolution Satellite Imagery with CNN Model Ensembles. Remote Sens. 2022, 14, 5655. https://doi.org/10.3390/rs14225655

AMA Style

Gonçalves BC, Wethington M, Lynch HJ. SealNet 2.0: Human-Level Fully-Automated Pack-Ice Seal Detection in Very-High-Resolution Satellite Imagery with CNN Model Ensembles. Remote Sensing. 2022; 14(22):5655. https://doi.org/10.3390/rs14225655

Chicago/Turabian Style

Gonçalves, Bento C., Michael Wethington, and Heather J. Lynch. 2022. "SealNet 2.0: Human-Level Fully-Automated Pack-Ice Seal Detection in Very-High-Resolution Satellite Imagery with CNN Model Ensembles" Remote Sensing 14, no. 22: 5655. https://doi.org/10.3390/rs14225655

APA Style

Gonçalves, B. C., Wethington, M., & Lynch, H. J. (2022). SealNet 2.0: Human-Level Fully-Automated Pack-Ice Seal Detection in Very-High-Resolution Satellite Imagery with CNN Model Ensembles. Remote Sensing, 14(22), 5655. https://doi.org/10.3390/rs14225655

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop