Next Article in Journal
Split-Band Interferometry-Assisted Phase Unwrapping for the Phase Ambiguities Correction
Next Article in Special Issue
Comparative Analysis of Responses of Land Surface Temperature to Long-Term Land Use/Cover Changes between a Coastal and Inland City: A Case of Freetown and Bo Town in Sierra Leone
Previous Article in Journal
Multiscale Union Regions Adaptive Sparse Representation for Hyperspectral Image Classification
Previous Article in Special Issue
Topic Modelling for Object-Based Unsupervised Classification of VHR Panchromatic Satellite Images Based on Multiscale Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Counting of Large Mammals from Very High Resolution Panchromatic Satellite Imagery

1
Faculty of Geo-Information Science and Earth Observation (ITC), University of Twente, P.O. Box 217, 7500 AE Enschede, The Netherlands
2
Department of Environmental Science, Macquarie University, Sydney, NSW 2109, Australia
*
Authors to whom correspondence should be addressed.
Remote Sens. 2017, 9(9), 878; https://doi.org/10.3390/rs9090878
Submission received: 22 June 2017 / Revised: 18 August 2017 / Accepted: 21 August 2017 / Published: 23 August 2017
(This article belongs to the Special Issue Learning to Understand Remote Sensing Images)

Abstract

:
Estimating animal populations by direct counting is an essential component of wildlife conservation and management. However, conventional approaches (i.e., ground survey and aerial survey) have intrinsic constraints. Advances in image data capture and processing provide new opportunities for using applied remote sensing to count animals. Previous studies have demonstrated the feasibility of using very high resolution multispectral satellite images for animal detection, but to date, the practicality of detecting animals from space using panchromatic imagery has not been proven. This study demonstrates that it is possible to detect and count large mammals (e.g., wildebeests and zebras) from a single, very high resolution GeoEye-1 panchromatic image in open savanna. A novel semi-supervised object-based method that combines a wavelet algorithm and a fuzzy neural network was developed. To discern large mammals from their surroundings and discriminate between animals and non-targets, we used the wavelet technique to highlight potential objects. To make full use of geometric attributes, we carefully trained the classifier, using the adaptive-network-based fuzzy inference system. Our proposed method (with an accuracy index of 0.79) significantly outperformed the traditional threshold-based method (with an accuracy index of 0.58) detecting large mammals in open savanna.

1. Introduction

Global biodiversity loss is a pressing environmental issue [1]. Populations of a number of wild animals have been reduced by half over the past four decades [2,3]. Counting wild animals to determine population size is an essential element of wildlife conservation and environmental management [4]. However, accurate population estimation using ground-based methods remains challenging, requiring considerable investment in resources and time [5]. Aerial surveys have been used as an alternative approach to detect large mammal populations and generate statistical estimates of their abundance in open areas [6]. In developed countries, wildlife such as caribou, elk, deer and moose have been monitored using aerial surveys [7,8,9]. For developing nations, where scores of endangered and threatened fauna are found, such an alternative is not always feasible due to limitations in access, technology, aircraft availability and skilled human resources [10,11]. It is therefore desirable to develop alternative approaches for conducting wildlife population counts in such regions.
Advances in satellite technology have provided new avenues in remote sensing for environmental applications, including the remote counting and mapping of animal populations. Lower spatial resolution satellite images have proven inadequate to detect and count individual animals [12], but the availability of commercial satellite images with a spatial resolution of one meter or less (e.g., IKONOS, QuickBird, GeoEye and WorldView) has made such an undertaking more feasible [13]. As a result, studies have been undertaken utilizing satellite remote sensing data to detect animals. For example, Fretwell et al. [14] successfully estimated the abundance of penguins from fecal staining of ice by using a combination of medium resolution (15–30 m) Landsat-7 ETM+ and very high resolution (0.6–2.5 m) QuickBird satellite images, but they did not attempt to count individual birds. Stapleton et al. [15] used different very high resolution (VHR) satellite images (i.e., QuickBird, WorldView-1 and WorldView-2) to track the distribution and abundance of polar bears. Although their findings demonstrated the potential of remote sensing applications for wildlife detection and monitoring, they also revealed the need for more automated detection processes to expedite analysis. Yang et al. [16] explored mammal detection in open savanna country from VHR (0.5–2 m) GeoEye-1 satellite images, using a hybrid image classification approach. Through a two-step process of pixel-based and object-based image classification, they were able to demonstrate the feasibility of automated detection and counting of large wild animals in vast open spaces. However, the method they proposed requires the input by an expert of a number of parameters, and therefore this method remains subjective and labor-intensive. Fretwell et al. [17] compared a number of classification techniques endeavoring to automatically detect whale-like objects. They found that a simple thresholding technique of the panchromatic and coastal band delivered the best results. Neither Stapleton et al. [15] nor Fretwell et al. [17] made full use of the multispectral band, while the panchromatic band played an important role in their research. To our knowledge, there has been no substantial exploration of the feasibility of using a single panchromatic (black and white) band for wildlife detection. The typical panchromatic band data obtained from airborne platforms have a much wider spectral range than is utilized by multispectral bands (red, green, blue) [18], and also have a higher radiometric resolution (number of bits per pixel). Moreover, panchromatic satellite images have a higher spatial resolution than multispectral images [19].
Object counting can also be achieved with computer vision techniques, such as local feature-based subspace clustering algorithms [20,21] and global feature-based saliency detection approaches [22,23,24,25]. The conventional clustering method, such as the K-means clustering algorithm, has been used to extract local features, but its performance relies on finding “similar” records in the training data and could therefore be highly influenced by noise [21]. Data in a specific category can also be well-represented by low-dimensional subspace where noise can be reduced [26]. To achieve a good result by eliminating the influence of errors (e.g., noise, outliers), Peng et al. [20] proposed a graph-oriented learning method, which applied the L2-Graph for subspace learning and subspace clustering, for facial recognition and moving-vehicle detection [26]. However, studies on subspace clustering mainly concentrate on high-dimensional data clustering, such as facial recognition and motion image segmentation. Saliency detection is a well-researched problem in computer vision. It aims at indicating the saliency likelihood of each pixel by generating bounding boxes, binary foreground and background segmentation, or saliency maps [27]. The aforementioned methods have proven to be useful for multi-level features with multi-band images, but are difficult to apply to a single-band image where the object consists of few pixels.
Aerial photographs have been used for bird censuses since the 1980s, counting image points falling below an established threshold [28,29]. Bajzak and Piatt [29] studied the greater snow goose, contrasting its white plumage against the surrounding mud flats by size and tonal class. Similarly, a panchromatic image can use thresholding as a simple image segmentation method that divides an image into objects and background [30,31,32]. It works well when targets contrast sharply with their background. However, thresholding methods have their limitations: (1) targets cannot be separated from ground elements with similar brightness values; (2) gray value thresholding does not make full use of geometric information; and (3) threshold values are defined manually and depend heavily on the user’s expertise.
Animal detection using remote sensing then predominantly switched to a two-step process [33]: (1) highlighting suspected targets; and then (2) classifying them, using geometric information. Groom et al. [33] proposed a scheme using geometric feature (object-size) filters to count birds against a monochromatic background. As targets were visually small and dim, they were not easily discerned against their background [34]. Using filters and image processing techniques, targets embedded in the scene could be visualized and detected [35,36,37,38]. However, the performance of such filters remains dependent on the brightness contrast between the target and background [34]. Several studies have employed wavelet-based techniques to address this concern [39,40,41]. The discernibility of targets from the background may vary at different scales, which can be problematic for object detection [19,42]. Wavelet analysis can transform signals into multiple resolutions, using an adaptive window [43], and thereby latently detect targets in cluttered backgrounds.
After highlighting the targets, the major challenge becomes how to make full use of geometric features to help separate a target from its surroundings. Spectral characteristics, cluster size, shape and other spatial features have been used in rule sets for image segmentation [44]. McNeill et al. [45] analyzed potential regions using shapes, by rejecting those with a compactness greater than a specified threshold value. Descamps et al. [46] counted large birds by fitting suspected objects (birds) into bright ellipses surrounded by a darker background. Expert knowledge can also play a critical role in image classification [16,47,48]. For example, Yang et al. [16] developed a specific rule set using expert knowledge to remove misclassified objects generated by object-based analysis. In another study, Wang et al. [47] proposed a hybrid neural network and expert system to quantify understory bamboo from satellite imagery, and they concluded that integration of a neural network and expert system appeared to be more efficient than when using either a neural network or an expert system alone. However, these methods rely on experts’ subjective experience and knowledge, which can be challenging for practical applications.
An alternative approach to using an expert system is machine learning: a data analysis technique that automates model building through algorithms that iteratively learn from a given dataset. Though different classifiers based on machine learning generate varying levels of accuracy for different datasets [49], the most recent machine-learning techniques have a proven ability to solve complex problems [50]. For example, convolutional neural networks (CNNs) [51] have emerged as state-of-the-art models for image classification and object detection [52,53,54,55,56,57]. Local connections, shared weight, pooling and multiple layers are four architectural factors that make CNNs excel in processing natural signals [58]. However, the human involvement level is high when tailoring the CNN algorithm to a specific task [59], and large data sets are required for training purposes to ensure a high quality output [60]. Another major limitation of CNNs is their intrinsic black-box nature: their internal workings are hidden and not easily understood [61], so the models they generate are unexplainable [62]. The fuzzy neural network (FNN) is an alternative model that incorporates both the explicit knowledge representation of an fuzzy inference system (FIS) and the learning ability of an artificial neural network [63,64].The McCulloch–Pitts model [65] was one of the earliest applications to use fuzzy sets with a neural network concept. Since the 1990s, Takagi and others have developed a solid foundation for the fuzzy neural network [66]. In 1993, Jang proposed the adaptive-network-based fuzzy inference system (ANFIS) [67]. This algorithm has been widely employed in applied mathematics [68,69,70,71], and, unlike traditional expert systems, does not require a high level of expert knowledge when developing decision rules.
This study aims to detect and count large mammals in open spaces from a single, VHR GeoEye-1 panchromatic image, using a novel semi-supervised object-based scheme that combines a wavelet algorithm and a fuzzy neural network.

2. Materials and Methods

2.1. Study Area and Animal Species

The study area is located in the Maasai Mara National Reserve (also known as Maasai Mara or the Mara), a large game reserve in the Great Rift Valley in the southern part of Kenya (Figure 1).
The reserve’s topography is mainly open savanna (grassland) with clusters of acacia trees along the southeastern area of the park [72]. The reserve not only protects the habitat of resident species, but also preserves a critical part of the route used by wildebeests and zebras during the great migration that traverses the Maasai Mara via the Serengeti National Park. The wildebeest is the dominant species of the Maasai Mara, and herd sizes can range from a few individuals to many thousands [73]. Serengeti wildebeests migrate seasonally, and are seen intermittently in the Mara between August and November [74]. The sheer numbers of animals that congregate during migration make the wildebeest an ideal candidate species to map through the use of satellite technology.

2.2. Satellite Images

We acquired two GeoEye-1 satellite images of part of the Maasai Mara National Reserve through the DigitalGlobe Foundation (www.digitalglobefoundation.org/), each covering an area of 25 km2. Both images are cloud free, and include one panchromatic (0.5 m) and four multispectral (2 m) bands. The image captured on 11 August 2009 depicts large numbers of animals. The other image, without any large animals present, was captured on 10 August 2013. To address our research objective, we carefully selected three small pilot study areas from the first image, each covering an area of 120 × 120 m (Figure 2). These pilot study areas were chosen to represent different levels of complexity regarding three criteria: (a) complexity of the landscape; (b) abundance of animals; and (c) feasibility and reliability of the visual interpretation of target animals. Pilot area No. 1 represents low complexity, with a few dozen animals viewed against a uniform background; Pilot area No. 2 represents moderate complexity, with more than one hundred animals viewed against a slightly less uniform background; and Pilot area No. 3 represents high complexity, with several hundred animals viewed against a non-uniform background.

2.3. Visual Interpretation to Establish Ground Truth for Large Animals Discerned on GeoEye-1 Imagery

Ground truth is required to calibrate the model, as well as validate the classification result. Using the panchromatic band of the GeoEye-1 image, large mammals (e.g., wildebeests and zebras) are visualized as 3–4 pixels long and 1–2 pixels wide [16]. Due to their similarity in size, large animals can be confused with small ground features such as bushes and termite mounds [75]. To facilitate the visual interpretation of target animals and avoid the problem of subjectivity, we used one pan-sharpened GeoEye-1 image with, and one without, the presence of large animals (Figure 3). We invited two experienced wildlife researchers from Africa as independent visual interpreters. Together we visually compared the two separate temporal images of the three pilot study locations at multiple scales under the ArcGIS 10.3.1 environment (ESRI Inc., Redlands, CA, USA). After the observers had discussed their interpretation results, especially regarding uncertain objects, and had agreed which identified objects were indeed large mammals, their knowledge was recorded as confirmed animal ground truth points. In total, we identified 50, 128 and 426 large mammals in the pilot study areas 1, 2 and 3, respectively.

2.4. Semi-Automatic Animal Detection Algorithm

Large mammals were identified by a series of multistage, semiautomatic techniques in VHR panchromatic satellite images. Our proposed scheme includes four principal steps (Figure 4): image preprocessing, preclassification, reclassification and accuracy assessment. Visual interpretation was incorporated for the purpose of reclassification and accuracy assessment.

2.4.1. Image Preprocessing

To highlight large mammals in the panchromatic imagery, we applied a histogram stretch in ENVI 5.2 (Exelis Visual Information Solutions, Inc., Boulder, CO, USA). Due to the limited resolution of the panchromatic band of VHR satellite images, an individual animal is represented as a cluster of pixels consisting of no more than eight pixels. To fully use their geometric information, we resampled the original image. Bicubic interpolation, which uses weighted arithmetic means, was chosen, as it maintains the quality of detailed information through antialiasing [76]. The image was carefully resized to eight times the original size, taking the wavelet decomposition performance into account, as well as memory and computation time, using
I = [ a i , j ] m × n
where the original image [ a i , j ] m × n is a matrix with m rows and n columns. We describe the resampled image as
I = f ( λ [ a i , j ] m × n )
where I is the new image, λ represents the diagonal matrix of the resized scale and f is the bicubic interpolation function.

2.4.2. Wavelet-Based Preclassification

Based on the generally accepted methodology of image decomposition and reconstruction, we used the wavelet-based method when highlighting suspected large mammals, to enhance their contrast against the immediate surroundings and to suppress irrelevant background [77,78]. Wavelet transform (WT) is based on the theory of Short-Time Fourier Transform (STFT) [79]. The WT differs from STFT in that it replaces infinite triangle function bases with finite decay wavelet bases. The finite decay wavelet bases, which are stretched (or squeezed) and translated from the mother wavelet, have an average value of 0 [80]. The WT of a continuous signal is defined as
T ( a , b ) = w ( a ) x ( t ) ψ * ( t b a ) dt
where a is scale, b is the position parameter, w ( a ) is a weighting function and ψ * ( t b a ) is the wavelet base [81]. If the wavelet base sufficiently corresponds to an input signal, the WT coefficient at this position is high [82]. The optimal mother wavelet and parameters were selected by comparing the performance of mainstream wavelet families regarding maintaining geometry features of suspected targets in our experimental imagery. A Haar wavelet (or db1 wavelet) was selected as it is not continuous and is therefore able to detect signals containing a sudden transition [83].
The image was transformed into a series of sub-images: A1 (low-frequency image), H1 (high-frequency image in the horizontal direction) and V1 (high-frequency image in the vertical direction); and then the same procedure was applied to the low frequency image (A1). Such a method permits multiresolution processing in both directions. After three transformation iterations, nine sub-images were generated, containing details as well as background. To highlight suspected targets and suppress background information, a weighted fusion algorithm was used. We then calculated the mean-square error (MSE) [84] between sub-images (resized to the original) and the original image. Sub-images containing more high-frequency information yielded higher MSE values. The weight of each sub-image should be
ω i = σ i 2 j = 1 n σ j 2 ( i , j = 1 , 2 , , n )
where i , j are the serial numbers of the current image, σ i ( j ) is the MSE of the current sub-image, and n is the total number of calculated sub-images. The weighted fusion algorithm creates a high signal-to-noise ratio (SNR) image. We then used Ostu’s method [85] in MATLAB (The Mathworks Inc., Natick, MA, USA), to discriminate between each suspected animal blob and the background.

2.4.3. Selecting Geometric Features

The next concern was how to identify which suspected large mammals were true large mammals. This entailed deciding which geometric features to use, typically length and area. We also considered gray value (hue) pixels. We used cross-validation (a model assessment technique) to verify the performance of classifiers [86]. This basically involves grouping raw data: one group is used as training set and the other for validation. K-fold cross-validation (K-CV) is a commonly used validation technique in object detection [86,87]. We divided the data into ten groups, and used each group once as the training dataset while the other nine groups acted as the validation dataset. We determined the most suitable combination for this experiment by calculating the average value of the training errors and checking errors using the dataset mentioned above at situations of different feature combinations. After employing the K-fold cross-validation multiple times, we decided a combination of feature area, major axis length, minor axis length and bounding box area was most suitable for this experiment.

2.4.4. ANFIS-Based Reclassification

A total of 100 blobs (or unknown objects) were randomly selected from the database to train the final model. The distribution of training data was comparable to the distribution of the whole dataset. Before we trained these data using ANFIS, a number of rules was decided upon. The Fuzzy C-Mean (FCM, or Fuzzy ISODATA), which was originally designed by Dunn [88], is a well-accepted clustering algorithm ideally suited to solving a natural problem [89,90]. As shown in Figure 5, this algorithm generated 10 cluster centers (corresponding to 10 membership functions for each variable). To limit the number of feature fields, we used expert knowledge to eliminate redundant classes. Finally, we input the 100 randomly selected blobs to train ANFIS in MATLAB. With the function genfis 2 , we built an initial fuzzy inference system (FIS) structure. We then loaded the initial FIS structure into the function anfis to train the ANFIS and develop the model. A hybrid method, including least-squares and backpropagation gradient descent, was applied to optimise the model. ANFIS model evaluation was conducted according to the ‘evalfis’ function. Required parameters for the ‘anfis’ function, including training error goal, initial training step size, step size decrease rate and step size increase rate, were set to default values (0, 0.01, 0.9, 1.1), which were proven to be adequate for most situations [91]. In order to avoid overfitting, we set the epoch number to 75 by considering both training error and checking error (see Appendix A). The adaptive tuning stops when the least-squares error is less than the training error goal, or has reached the epoch number. By loading all the datasets containing feature values into the model, all suspected blobs were classified by the inference system into targets and non-targets.

2.5. Accuracy Assessment

We assessed the accuracy of the classification results by comparing the number of large mammals detected by the computer model with the ground truthing, and then calculated the omission error and commission error [92]. Detection accuracy (DA), which is the most commonly used metric, is highly inversely correlated ( D A + o m i s s i o n   e r r o r = 1 ) [93]. The values for both the omission error and the commission error are always between 0 and 1. The closer their values are to 0, the better the result.
The accuracy index (AI), which was devised by Pouliot et al. [94], was computed as:
AI = N TP FN N
where TP (true positive) denotes the number of targets occurring in both the ground truth and our processing result; FN (false negative) denotes the number of targets that do appear in the ground truth, but not in our processing result; FP (false positive) denotes the number of targets occurring in our processing result, but not in the ground truth data; and N is the number of ground truth targets in the study area. The higher the value of the accuracy index, the better the result.

3. Results

In Figure 6, the visual results of our semi-automated ANFIS-wavelet approach to detecting large mammals are compared with the results gained with the thresholding method.
The accuracy index regarding the proposed method for the low complexity study area (No. 1) was as high as 0.86 (Table 1). For the higher-complexity sites, the results also yielded acceptable accuracy indices: 0.79 and 0.72, respectively, for the moderately (No. 2) and highly (No. 3) complex sites. As shown in Table 2, the thresholding method produced accuracy indices of 0.64, 0.56 and 0.54, respectively, for the low, moderate and high complexity areas, with an average accuracy index of 0.58. The average accuracy index of our proposed method, depicted in Table 1, is 0.79, which is 0.21 higher than that of the thresholding method. Also, the calculated omission and commission errors of our approach (0.09 and 0.12, respectively) are lower than those of the thresholding method (0.15 and 0.24, respectively). It should also be noted that, if the study area is more complex, this does not necessarily mean that the detection is less accurate. As shown in Figure 6, specific ground features can introduce inaccuracies, such as the errors appearing in this study close to roads and edges of forests. In absolute terms of detected targets, the thresholding technique and our semi-automated ANFIS-wavelet approach showed different accuracies for each pilot study area. The statistical results regarding this study area illustrate that a higher detection accuracy is obtained with the ANFIS-wavelet method than with the threshold-based method.

4. Discussion

The results from this study demonstrate that it is feasible to use VHR panchromatic satellite imagery to detect and count large mammals in extensive open areas. In comparison with the traditional thresholding technique, our ANFIS-wavelet method produced a higher accuracy index and less commission/omission errors.
Although the thresholding method performs adequately when the targets share similar gray values and are dissimilar to their background, it is less accurate in more complex areas. There are two main reasons for the higher commission error found when using the thresholding method. Firstly, when the gray values of suspected objects (animals) are similar to those of the surroundings, they may be ignored by the threshold-based segmentation. In the ANFIS-wavelet method, the representation of the target is considered at different spatial scales. Suspected animals that do contrast with their immediate background, once different spatial scales are considered, will contribute to a higher weighted value in the preclassification results. Secondly, when animal objects and terrain have similar gray values, they cannot be altered simply by using thresholds: more information is required before further processing can be undertaken [32]. We statistically selected four geometric features to distinguish non-target objects from large mammals in the feature space. This approach proved more accurate than merely using a simple threshold value.
The commission error derived from our method was found to be three percentage points greater than the omission error, resulting in more non-target objects being incorrectly classified as large mammals than large mammals being incorrectly omitted. Further analysis revealed that commission errors always appeared near roads and vegetation. Bushes were confused with large mammals because of similarities in geometric features. Rough road surfaces or vehicles may result in discontinuous blobs and may thus also be recognized as large mammals by our method. Two reasons for omission include targets that are not clearly distinguishable from the background and targets that are too close to each other.
The geometric features chosen to distinguish an animal from its background were area, major axis length, minor axis length and bounding area. These features differ between target animals and non-targets such as shrubs or boulders. Even though some features were highly correlated, they can also help us in detecting animals. For example, defining both major and minor axis length can help to eliminate objects that do not have a correct length–width ratio.
The ANFIS-wavelet method has proved to be a feasible method for detecting animals in open savanna landscapes. This method is based on wavelet preclassification followed by ANFIS reclassification. The wavelet-based classification is able to highlight objects and maintain their geometric features. This is critical because the targets are dim and small, and as much useful information as possible needs to be retained. By using multiscale analysis, targets can be precisely located in poorer quality (i.e., low SNR) imagery without information loss. The ANFIS, which combines the advantages of machine learning and a fuzzy system, makes it possible to learn from data and concomitantly use existing expert knowledge, resulting in a method that is both efficient and stable.

5. Conclusions

We developed a novel semi-supervised object-based method that combines a wavelet algorithm and a fuzzy neural network for detecting and counting large mammals (e.g., wildebeests and zebras) from a single, very high resolution GeoEye-1 panchromatic image in open savanna. To discern large mammals from their surroundings and discriminate between animals and non-targets, we used the wavelet technique to highlight potential objects. To make full use of geometric attributes, we carefully trained the classifier, using the adaptive-network-based fuzzy inference system. We then compared our method with the traditional threshold-based method. The results showed that our proposed method (with an accuracy index of 0.79) significantly outperformed the traditional threshold-based method (with an accuracy index of 0.58) in detecting large mammals in open savanna. The greater availability of VHR images, and the advances in image segmentation techniques, mean that animal detection by means of remote sensing technology is a pragmatic alternative to direct animal counting. Further developments in image processing should eventually make it feasible to detect and monitor medium-sized and small animals remotely from space as well.

Acknowledgments

Yifei Xue was supported by the China Scholarship Council (CSC) and co-funded by the ITC Research Fund from the Faculty of Geo-Information Science and Earth Observation (ITC), University of Twente, the Netherlands. We acknowledge the satellite imagery support received from the DigitalGlobe Foundation. We also thank Festus Ihwagi and Tawanda Gara for their assistance with the visual interpretation of the GeoEye-1 satellite images.

Author Contributions

Yifei Xue, Tiejun Wang and Andrew K. Skidmore conceived and designed the experiment. Yifei Xue analyzed the data and wrote the paper. All authors contributed to the editing of manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

ANFIS is a hybrid method which combines both least-square and backpropagation algorithms. The training datasets were used to construct the initial model, and the validation datasets were used for tuning. The training algorithm stops when either the training error goal value is satisfied, or the number of training epochs is reached. The training error goal was always being used as default value 0 when solving an unknown problem [91]. After a certain epoch number, the model will overfit the training data. To avoid overfitting, an optimal epoch number is required, but it is also difficult to determine. We evaluated the training error and the checking error (also known as validation error) with increasing the epoch number (Figure A1). Root-mean-square error (RMSE) is one of the most used indexes for performance indication [95]. The RMSE of the training data decreases along with the epoch number, but the tendency is slowed down after around 120 epochs and does not seem to have an obvious descent after 200 epochs. The RMSE of the checking data decreases along with the epoch number until around 75 epochs, and increases rapidly before around 120 epochs. According to this quantitative analysis, we found that it is proper to set the epoch number to around 75.
Figure A1. Indentification of optimum epoch number based on the root-mean-square error of both training error and checking error.
Figure A1. Indentification of optimum epoch number based on the root-mean-square error of both training error and checking error.
Remotesensing 09 00878 g007

References

  1. Skidmore, A.K.; Pettorelli, N.; Coops, N.C.; Geller, G.N.; Hansen, M.; Lucas, R.; Mücher, C.A.; O’Connor, B.; Paganini, M.; Pereira, H.M.; et al. Environmental science: Agree on biodiversity metrics to track from space. Nature 2015, 523, 403–405. [Google Scholar] [CrossRef] [PubMed]
  2. Cuttelod, A.; Garcia, N.; Malak, D.A.; Temple, H.; Katariya, V. The Mediterranean: a biodiversity hotspot under threat. In Wildlife in a Changing World: An Analysis of the 2008 IUCN Red List of Threatened Species; Vié, J.-C., Hilton-Taylor, C., Stuart, S.N., Eds.; IUCN: Gland, Switzerland, 2009. [Google Scholar]
  3. Carrington, D. Earth has lost half of its wildlife in the past 40 years, says WWF. Available online: https://www.theguardian.com/environment/2014/sep/29/earth-lost-50-wildlife-in-40-years-wwf (accessed on 30 August 2016).
  4. Ramono, W.; Rubianto, A.; Herdiana, Y. Spatial distributions of Sumatran rhino calf at Way Kambas National Park based on its footprint and forest fire in one decade (2006 to 2015). In Proceedings of the Scientific Program of the 15th International Elephant & Rhino Conservation and Research Symposium, Singapore, 14–18 November 2016; p. 63. [Google Scholar]
  5. Witmer, G.W. Wildlife population monitoring: Some practical considerations. Wildl. Res. 2005, 32, 259–263. [Google Scholar] [CrossRef]
  6. Jones, G.P. The Feasibility of Using Small Unmanned Aerial Vehicles for Wildlife Research; University of Florida: Gainesville, FL, USA, 2003. [Google Scholar]
  7. Gasaway, W.C.; DuBios, S.D.; Reed, D.J.; Harbo, S.J. Estimating Moose Population Parameters from Aerial Surveys; University of Alaska: Fairbanks, AK, USA, 1986. [Google Scholar]
  8. Couturier, S.; Courtois, R.; Crépeau, H.; Rivest, L.-P.; Luttich, S.N. Calving photocensus of the Rivière George Caribou Herd and comparison with an independent census. Rangifer 1996, 16, 283–296. [Google Scholar] [CrossRef]
  9. Pettorelli, N.; Côté, S.D.S.; Gingras, A.; Potvin, F.; Huot, J. Aerial surveys vs hunting statistics to monitor deer density: The example of Anticosti Island, Quebec, Canada. Wildl. Biol. 2007, 3, 321–327. [Google Scholar] [CrossRef]
  10. Barnes, R.F.W. The problem of precision and trend detection posed by small elephant populations in West Africa. Afr. J. Ecol. 2002, 40, 179–185. [Google Scholar] [CrossRef]
  11. Ransom, J.I.; Kaczensky, P.; Lubow, B.C.; Ganbaatar, O.; Altansukh, N. A collaborative approach for estimating terrestrial wildlife abundance. Biol. Conserv. 2012, 153, 219–226. [Google Scholar] [CrossRef]
  12. Löffler, E.; Margules, C. Wombats detected from space. Remote Sens. Environ. 1980, 9, 47–56. [Google Scholar] [CrossRef]
  13. Maglione, P. Very high resolution optical satellites: An overview of the most commonly used. Am. J. Appl. Sci. 2016, 13, 91–99. [Google Scholar] [CrossRef]
  14. Fretwell, P.T.; LaRue, M.A.; Morin, P.; Kooyman, G.L.; Wienecke, B.; Ratcliffe, N.; Fox, A.J.; Fleming, A.H.; Porter, C.; Trathan, P.N. An emperor penguin population estimate: The first global, synoptic survey of a species from space. PLoS ONE 2012, 7. [Google Scholar] [CrossRef]
  15. Stapleton, S.; LaRue, M.; Lecomte, N.; Atkinson, S.; Garshelis, D.; Porter, C.; Atwood, T. Polar bears from space: Assessing satellite imagery as a tool to track arctic wildlife. PLoS ONE 2014, 9. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, Z.; Wang, T.; Skidmore, A.K.; de Leeuw, J.; Said, M.Y.; Freer, J. Spotting East African Mammals in Open Savannah from Space. PLoS ONE 2014, 9, 1–16. [Google Scholar] [CrossRef] [PubMed]
  17. Fretwell, P.T.; Staniland, I.J.; Forcada, J. Whales from space: Counting southern right whales by satellite. PLoS ONE 2014, 9, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Liu, J.G.; Mason, P.J. Essential Image Processing and GIS for Remote Sensing; John Wiley & Sons Ltd.: London, UK, 2009; ISBN 9780470510322. [Google Scholar]
  19. Zhang, K.; Wang, M.; Yang, S.; Member, S.; Xing, Y.; Qu, R. Fusion of panchromatic and multispectral images via coupled sparse non-negative matrix factorization. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 5740–5747. [Google Scholar] [CrossRef]
  20. Peng, X.; Member, S.; Zhang, L.; Yi, Z.; Member, S. Constructing the L2-Graph for Subspace Learning and Subspace Clustering. IEEE Trans. Cybern. 2016, 47, 1053–1066. [Google Scholar] [CrossRef] [PubMed]
  21. Otto, C.; Wang, D.; Jain, A. Clustering Millions of Faces by Identity. IEEE Trans. Pattern Anal. Mach. Intell. 2016. [Google Scholar] [CrossRef] [PubMed]
  22. Li, Z.; Itti, L. Saliency and gist features for target detection in satellite images. IEEE Trans. Image Process. 2011, 20, 2017–2029. [Google Scholar] [PubMed]
  23. Wang, Q.; Yuan, Y.; Yan, P.; Li, X. Saliency detection by multiple-instance learning. IEEE Trans. Cybern. 2013, 43, 660–672. [Google Scholar] [CrossRef] [PubMed]
  24. Wang, Z.; Du, L.; Wang, F.; Su, H.; Zhou, Y. Multi-Scale Target Detection in SAR Image Based on Visual Attention Model. In Proceedings of the 2015 IEEE 5th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Singapore, 1–4 September 2015; pp. 704–709. [Google Scholar]
  25. Wang, Q.; Lin, J.; Yuan, Y. Salient Band Selection for Hyperspectral Image Classification via Manifold Ranking. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1279–1289. [Google Scholar] [CrossRef] [PubMed]
  26. Elhamifar, E.; Rene, V. Sparse Subspace Clustering: Algorithm, Theory, and Applications Ehsan. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1–19. [Google Scholar] [CrossRef] [PubMed]
  27. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M.H. Saliency detection via graph-based manifold ranking. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2013, 3166–3173. [Google Scholar]
  28. Gilmer, D.S.; Brass, J.A.; Strong, L.L.; Card, D.H. Goose counts from aerial photographs using an optical digitizer. Wildl. Soc. Bull. 1988, 16, 204–206. [Google Scholar]
  29. Bajzak, D.; Piatt, J.F. Computer-aided procedure for counting waterfowl on aerial photographs. Wildl. Soc. Bull. 1990, 18, 125–129. [Google Scholar]
  30. Glasbey, C.A.; Horgan, G.W.; Darbyshire, J.F. Image analysis and three-dimensional modelling of pores in soil aggregates. J. Soil Sci. 1991, 42, 479–486. [Google Scholar] [CrossRef]
  31. Cunningham, D.J.; Anderson, W.H.; Anthony, R.M. An image-processing program for automated counting. Wildl. Soc. Bull. 1996, 24, 345–346. [Google Scholar]
  32. Laliberte, A.S.; Ripple, W.J. Automated wildlife counts from remotely sensed imagery. Wildl. Soc. Bull. 2003, 31, 362–371. [Google Scholar]
  33. Groom, G.; Krag Petersen, I.; Anderson, M.D.; Fox, A.D. Using object-based analysis of image data to count birds: Mapping of Lesser Flamingos at Kamfers Dam, Northern Cape, South Africa. Int. J. Remote Sens. 2011, 32, 4611–4639. [Google Scholar] [CrossRef]
  34. Bai, X.; Zhang, S.; Du, B.; Liu, Z.; Jin, T.; Xue, B.; Zhou, F. Survey on dim small target detection in clutter background: Wavelet, inter-frame and filter based algorithms. Procedia Eng. 2011, 15, 479–483. [Google Scholar] [CrossRef]
  35. Soni, T.; Zeidler, J.R.; Ku, W.H. Performance evaluation of 2-D adaptive prediction filters for detection of small objects in image data. IEEE Trans. Image Process. 1993, 2, 327–340. [Google Scholar] [CrossRef] [PubMed]
  36. Shirvaikar, M.V. A neural network filter to detect small targets in high clutter backgrounds. IEEE Trans. Neural Netw. 1995, 6, 252–257. [Google Scholar] [CrossRef] [PubMed]
  37. Casasent, D.; Ye, A. Detection filters and algorithm fusion for ATR. IEEE Trans. Image Process. 1997, 6, 114–125. [Google Scholar] [CrossRef] [PubMed]
  38. Trathan, P.N.; Ratcliffe, N.; Masden, E.A. Ecological drivers of change at South Georgia: The krill surplus, or climate variability. Ecography 2012, 35, 983–993. [Google Scholar] [CrossRef]
  39. Boccignone, G.; Chianese, A.; Picariello, A. Small target detection using wavelets. Proc. Fourteenth Int. Conf. Pattern Recognit. 1998, 2, 1776–1778. [Google Scholar]
  40. Davidson, G.; Griffiths, H.D. Wavelet detection scheme for small targets in sea clutter. Electron. Lett. 2002, 38, 1128–1130. [Google Scholar] [CrossRef]
  41. Kim, S. High-speed incoming infrared target detection by fusion of spatial and temporal detectors. Sensors 2015, 15, 7267–7293. [Google Scholar] [CrossRef] [PubMed]
  42. Zhao, J.; Liu, F.; Mo, B. An algorithm of dim and small target detection based on wavelet transform and image fusion. In Proceedings of the 2012 Fifth International Symposium on Computational Intelligence and Design, Hangzhou, China, 28–29 October 2012; pp. 43–45. [Google Scholar]
  43. Duk, V.; Ng, B.; Rosenberg, L. The potential of 2D wavelet transforms for target detection in sea-clutter. In Proceedings of the 2015 IEEE Radar Conference (RadarCon), Arlington, VA, USA, 10–15 May 2015; pp. 901–906. [Google Scholar]
  44. Groom, G.; Stjernholm, M.; Nielsen, R.D.; Fleetwood, A.; Petersen, I.K. Remote sensing image data and automated analysis to describe marine bird distributions and abundances. Ecol. Inform. 2013, 14, 2–8. [Google Scholar] [CrossRef]
  45. McNeill, S.; Barton, K.; Lyver, P.; Pairman, D. Semi-automated penguin counting from digital aerial photographs. In Proceedings of the 2011 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Vancouver, BC, Canada, 24–29 July 2011; pp. 4312–4315. [Google Scholar]
  46. Descamps, S.; Béchet, A.; Descombes, X.; Arnaud, A.; Zerubia, J. An automatic counter for aerial images of aggregations of large birds. Bird Study 2011, 58, 302–308. [Google Scholar] [CrossRef]
  47. Wang, T.J.; Skidmore, A.K.; Toxopeus, A.G. Improved understorey bamboo cover mapping using a novel hybrid neural network and expert system. Int. J. Remote Sens. 2009, 30, 965–981. [Google Scholar] [CrossRef]
  48. Dagnino, A.; Allen, J.I.; Moore, M.N.; Broeg, K.; Canesi, L.; Viarengo, A. Development of an expert system for the integration of biomarker responses in mussels into an animal health index. Biomarkers 2007, 12, 155–172. [Google Scholar] [CrossRef] [PubMed]
  49. Fernández-Delgado, M.; Cernadas, E.; Barro, S.; Amorim, D. Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 2014, 15, 3133–3181. [Google Scholar]
  50. Schmidhuber, J. Deep Learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed]
  51. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2323. [Google Scholar] [CrossRef]
  52. Johnson, J.; Karpathy, A.; Fei-Fei, L. DenseCap: Fully Convolutional Localization Networks for Dense Captioning. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Washington, WA, USA, 27–30 June 2016; pp. 4565–4574. [Google Scholar]
  53. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
  54. Papandreou, G.; Kokkinos, I.; Savalle, P.A. Modeling local and global deformations in Deep Learning: Epitomic convolution, Multiple Instance Learning, and sliding window detection. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2015, 390–399. [Google Scholar] [CrossRef]
  55. Wei, Y.; Xia, W.; Lin, M.; Huang, J.; Ni, B.; Dong, J.; Zhao, Y.; Yan, S. HCP: A Flexible CNN Framework for Multi-Label Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1901–1907. [Google Scholar] [CrossRef] [PubMed]
  56. Wu, H.; Zhang, H.; Zhang, J.; Xu, F. Typical target detection in satellite images based on convolutional neural networks. In Proceedings of the IEEE International Conference on System, Man and Cybernetics, Hong Kong, China, 9–12 October 2015; pp. 2956–2961. [Google Scholar]
  57. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  58. LeCun, Y.; Yoshua, B.; Geoffrey, H. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  59. Bengio, Y.; LeCun, Y. Scaling Learning Algorithms towards AI. In Large-Scale Kernel Machines; MIT Press: Cambridge, MA, USA, 2007; pp. 1–41. ISBN 1002620262. [Google Scholar]
  60. Kala, R.; Shulkla, A.; Tiwari, R. Fuzzy Neuro Systems for Machine Learning for Large Data Sets. In Proceedings of the 2009 IEEE International Advance Computing Conference, Patiala, India, 6–7 March 2009; pp. 6–7. [Google Scholar]
  61. Yager, R.R.; Zadeh, L.A. (Eds.) An Introduction to Fuzzy Logic Applications in Intelligent Systems; Springer Science & Business Media: Berlin, Germany, 2012. [Google Scholar]
  62. Ma, H.; Ma, X.; Liu, W.; Huang, Z.; Gao, D.; Jia, C. Control flow obfuscation using neural network to fight concolic testing. Lect. Notes Inst. Comput. Sci. Soc. Telecommun. Eng. LNICST 2015, 152, 287–304. [Google Scholar]
  63. Buckley, J.J.; Hayashi, Y. Fuzzy neural networks: A survey. Fuzzy Sets Syst. 1994, 66, 1–13. [Google Scholar] [CrossRef]
  64. Hosseini, M.S.; Zekri, M. Review of Medical Image Classification using the Adaptive Neuro-Fuzzy Inference System. J. Med. Signals Sens. 2012, 2, 49–60. [Google Scholar] [PubMed]
  65. McCulloch, W.S.; Pitts, W.H. A logical calculus of the idea immanent in nervous activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  66. Takagi, H.; Suzuki, N.; Koda, T.; Kojima, Y. Neural networks designed on approximate reasoning architecture and their applications. IEEE Trans. Neural Netw. 1992, 3, 752–760. [Google Scholar] [CrossRef] [PubMed]
  67. Jang, J.-S.R. ANFIS: Adaptive-Network-Based Fuzzy Inference System. IEEE Trans. Syst. Man Cybern. 1993, 23, 665–685. [Google Scholar] [CrossRef]
  68. Kurian, C.P.; George, V.I.; Bhat, J.; Aithal, R.S. Anfis model for the time series prediction of interior daylight illuminance. ICGST Int. J. Artif. Intell. Mach. Learn. 2006, 6, 35–40. [Google Scholar]
  69. Yun, Z.; Quan, Z.; Caixin, S.; Shaolan, L.; Yuming, L.; Yang, S. RBF neural network and ANFIS-based short-term load forecasting approach in real-time price environment. IEEE Trans. Power Syst. 2008, 23, 853–858. [Google Scholar]
  70. Boyacioglu, M.A.; Avci, D. An adaptive network-based fuzzy inference system (ANFIS) for the prediction of stock market return: The case of the Istanbul stock exchange. Expert Syst. Appl. 2010, 37, 7908–7912. [Google Scholar] [CrossRef]
  71. Hiremath, S. Transmission rate prediction for cognitive radio using adaptive neural fuzzy inference system. Proceedings of 2010 5th International Conference on Industrial and Information Systems, ICIIS 2010, Mangalore, India, 29 July–1 August 2010; pp. 92–97. [Google Scholar]
  72. Ford, A.T.; Fryxell, J.M.; Sinclair, A.R.E. Conservation challenges facing African savanna ecosystems. In Antelope Conservation: From Diagnosis to Action; Bro-Jørgensen, J., Mallon, D.P., Eds.; John Wiley & Sons Ltd.: London, UK, 2016; pp. 11–31. ISBN 9781118409572. [Google Scholar]
  73. Hopcraft, J.G.C.; Sinclair, A.R.E.; Holdo, R.M.; Mwangomo, E.; Mduma, S.; Thirgood, S.; Borner, M.; Fryxell, J.M.; Olff, H. Why are wildebeest the most abundant herbivore in the Serengeti ecosystem? In Serengeti IV: Sustaining Biodiversity in a Coupled Human-Natural System; University of Chicago Press: Chicago, IL, USA, 2015; pp. 35–72. [Google Scholar]
  74. Boone, R.B.; Thirgood, S.J.; Hopcraft, J.G.C. Serengeti Wildebeest Migratory Patterns Modeled from Rainfall and New Vegetation Growth. Ecology 2006, 87, 1987–1994. [Google Scholar] [CrossRef]
  75. Pringle, R.M.; Doak, D.F.; Brody, A.K.; Jocque, R.; Palmer, T.M. Spatial pattern enhances ecosystem functioning in an african savanna. PLoS Biol. 2010, 8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  76. Han, D. Comparison of commonly used image interpolation methods. In Proceedings of the 2nd International Conference on Computer Science and Electronics Engineering (ICCSEE 2013), Hangzhou, China, 22–23 March 2013; pp. 1556–1559. [Google Scholar]
  77. Li, H.; Manjunath, B.S.; Mitra, S.K. Multisensor image fusion using the wavelet transform. Graph. Model. Image Process. 1995, 57, 235–245. [Google Scholar] [CrossRef]
  78. Zhang, Y.; Dong, Z.; Wang, S.; Ji, G.; Yang, J. Preclinical diagnosis of magnetic resonance (MR) brain images via discrete wavelet packet transform with tsallis entropy and generalized eigenvalue proximate support vector machine (GEPSVM). Entropy 2015, 17, 1795–1813. [Google Scholar] [CrossRef]
  79. Daubechies, I. The wavelet transform, time-frequency localization and signal analysis. Inf. Theory IEEE Trans. 1990, 36, 961–1005. [Google Scholar] [CrossRef]
  80. Young, R.K. Wavelet Theory and Its Applications; Springer Science & Business Media: Berlin, Germany, 2012. [Google Scholar]
  81. Addison, P.S. The Illustrated Wavelet Transform Handbook: Introductory Theory and Applications in Science, Engineering, Medicine and Finance; IOP Publishing: Bristol, UK, 2002. [Google Scholar]
  82. Ye, X.; Wang, T.; Skidmore, A.K.; Fortin, D.; Bastille-Rousseau, G.; Parrott, L. A wavelet-based approach to evaluate the roles of structural and functional landscape heterogeneity in animal space use at multiple scales. Ecography 2015, 38, 740–750. [Google Scholar] [CrossRef]
  83. Lee, B.Y.; Tarng, Y.S. Application of the discrete wavelet transform to the monitoring of tool failure in end milling using the spindle motor current. Int. J. Adv. Manuf. Technol. 1999, 15, 238–243. [Google Scholar] [CrossRef]
  84. Ephraim, Y.; Malah, D. Speech enhancement using a minimum mean-square error short-time spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1984, 32, 1109–1122. [Google Scholar] [CrossRef]
  85. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  86. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 1995; pp. 1137–1143. [Google Scholar]
  87. Bengio, Y.; Grandvalet, Y. No unbiased estimator of the variance of k-fold cross-validation. J. Mach. Learn. Res. 2004, 5, 1089–1105. [Google Scholar]
  88. Dunn, J.C. A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. Cybern. Syst. 1973, 3, 32–57. [Google Scholar] [CrossRef]
  89. Güler, C.; Thyne, G.D. Delineation of hydrochemical facies distribution in a regional groundwater system by means of fuzzy c-means clustering. Water Resour. Res. 2004, 40, 1–11. [Google Scholar] [CrossRef]
  90. Shahi, A. An effective fuzzy C-Mean and Type-2 fuzzy logic for weather forecasting. J. Theor. Appl. Inf. Technol. 2009, 5, 556–567. [Google Scholar]
  91. Ozkan, C. Surface interpolation by adaptive neuro-fuzzy inference system based local ordinary kriging. Lect. Notes Comput. Sci. 2006, 3851, 196–205. [Google Scholar]
  92. Congalton, R.G. A review of assessing the accuracy of classification of remotely sensed data. Remote Sens. Environ. 1991, 37, 35–46. [Google Scholar] [CrossRef]
  93. Yin, D.; Wang, L. How to assess the accuracy of the individual tree-based forest inventory derived from remotely sensed data: A review. Int. J. Remote Sens. 2016, 37, 4521–4553. [Google Scholar] [CrossRef]
  94. Pouliot, D.A.; King, D.J.; Bell, F.W.; Pitt, D.G. Automated tree crown detection and delineation in high-resolution digital camera imagery of coniferous forest regeneration. Remote Sens. Environ. 2002, 82, 322–334. [Google Scholar] [CrossRef]
  95. Dragomir, O.E.; Dragomir, F.; Stefan, V.; Minca, E. Adaptive Neuro-Fuzzy Inference Systems as a Strategy for Predicting and Controling the Energy Produced from Renewable Sources. Energies 2015, 8, 13047–13061. [Google Scholar] [CrossRef]
Figure 1. Location of the Maasai Mara National Reserve in Kenya and the three pilot study areas on a natural color composite of a GeoEye-1 image, acquired on 11 August 2009.
Figure 1. Location of the Maasai Mara National Reserve in Kenya and the three pilot study areas on a natural color composite of a GeoEye-1 image, acquired on 11 August 2009.
Remotesensing 09 00878 g001
Figure 2. The panchromatic band of the GeoEye-1 image taken on 11 August 2009, showing large mammals in the Maasai Mara National Reserve. Pilot area No. 1 represents low complexity regarding animal numbers and uniformity of background; Pilot area No. 2 represents moderate complexity; and Pilot area No. 3 represents high complexity. The rectangle visible in the top-left corner of Pilot area No. 3 is a white vehicle.
Figure 2. The panchromatic band of the GeoEye-1 image taken on 11 August 2009, showing large mammals in the Maasai Mara National Reserve. Pilot area No. 1 represents low complexity regarding animal numbers and uniformity of background; Pilot area No. 2 represents moderate complexity; and Pilot area No. 3 represents high complexity. The rectangle visible in the top-left corner of Pilot area No. 3 is a white vehicle.
Remotesensing 09 00878 g002
Figure 3. Visual interpretation of target animals by comparing two pan-sharpened GeoEye-1 images (0.5 m): one acquired 10 August 2013, without large animals present (top), and one acquired 11 August 2009, with large animals (bottom). The three pilot study areas represent the complexity of the landscape and the abundance of animals appearing in these images, from left to right: low, moderate and high.
Figure 3. Visual interpretation of target animals by comparing two pan-sharpened GeoEye-1 images (0.5 m): one acquired 10 August 2013, without large animals present (top), and one acquired 11 August 2009, with large animals (bottom). The three pilot study areas represent the complexity of the landscape and the abundance of animals appearing in these images, from left to right: low, moderate and high.
Remotesensing 09 00878 g003
Figure 4. Workflow of the proposed method for counting large mammals from a single, very high resolution panchromatic GeoEye-1 satellite image.
Figure 4. Workflow of the proposed method for counting large mammals from a single, very high resolution panchromatic GeoEye-1 satellite image.
Remotesensing 09 00878 g004
Figure 5. Flow diagram of the adaptive-network-based fuzzy inference system (ANFIS) based reclassification system.
Figure 5. Flow diagram of the adaptive-network-based fuzzy inference system (ANFIS) based reclassification system.
Remotesensing 09 00878 g005
Figure 6. Results regarding large mammal detection in the three different pilot study areas. The columns show images of the three pilot areas: No. 1 is of low complexity, No. 2 of moderate complexity and No. 3 of high complexity. The first row contains the original panchromatic satellite images; the second row illustrates results based on the thresholding method; and the third row illustrates the results obtained using the method proposed in this study (i.e., ANFIS-wavelet). The green, red and yellow dots indicate true positive, false negative and false positive results, respectively.
Figure 6. Results regarding large mammal detection in the three different pilot study areas. The columns show images of the three pilot areas: No. 1 is of low complexity, No. 2 of moderate complexity and No. 3 of high complexity. The first row contains the original panchromatic satellite images; the second row illustrates results based on the thresholding method; and the third row illustrates the results obtained using the method proposed in this study (i.e., ANFIS-wavelet). The green, red and yellow dots indicate true positive, false negative and false positive results, respectively.
Remotesensing 09 00878 g006
Table 1. Accuracy assessment of the ANFIS-wavelet method for the three pilot study areas: No. 1, No. 2 and No. 3, with low, moderate and high complexity, respectively.
Table 1. Accuracy assessment of the ANFIS-wavelet method for the three pilot study areas: No. 1, No. 2 and No. 3, with low, moderate and high complexity, respectively.
Pilot Area No. 1Pilot Area No. 2Pilot Area No. 3Average
Ground truth50128416198
True positive47118370178
False positive4176428
False negative3105623
Omission error0.060.080.130.09
Commission error0.080.130.150.12
Accuracy index0.860.790.720.79
Table 2. Accuracy assessment of the threshold-based method for the three pilot study areas with low, moderate and high complexity, respectively.
Table 2. Accuracy assessment of the threshold-based method for the three pilot study areas with low, moderate and high complexity, respectively.
Pilot Area No. 1Pilot Area No. 2Pilot Area No. 3Average
Ground truth50128416198
True positive45105354168
False positive133312657
False negative5237233
Omission error0.100.180.170.15
Commission error0.220.240.260.24
Accuracy index0.640.560.540.58

Share and Cite

MDPI and ACS Style

Xue, Y.; Wang, T.; Skidmore, A.K. Automatic Counting of Large Mammals from Very High Resolution Panchromatic Satellite Imagery. Remote Sens. 2017, 9, 878. https://doi.org/10.3390/rs9090878

AMA Style

Xue Y, Wang T, Skidmore AK. Automatic Counting of Large Mammals from Very High Resolution Panchromatic Satellite Imagery. Remote Sensing. 2017; 9(9):878. https://doi.org/10.3390/rs9090878

Chicago/Turabian Style

Xue, Yifei, Tiejun Wang, and Andrew K. Skidmore. 2017. "Automatic Counting of Large Mammals from Very High Resolution Panchromatic Satellite Imagery" Remote Sensing 9, no. 9: 878. https://doi.org/10.3390/rs9090878

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop