Next Article in Journal
Mapping Urban Excavation Induced Deformation in 3D via Multiplatform InSAR Time-Series
Previous Article in Journal
Assessment of Multi-Frequency PPP Ambiguity Resolution Using Galileo and BeiDou-3 Signals
Previous Article in Special Issue
Classifying Major Explosions and Paroxysms at Stromboli Volcano (Italy) from Space
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Video Observation Data for Volcanic Activity Monitoring Using Computer Vision and Modern Neural NetWorks (on Klyuchevskoy Volcano Example)

1
Computing Center, Far Eastern Branch of the Russian Academy of Sciences, 680000 Khabarovsk, Russia
2
Institute of Volcanology and Seismology, Far Eastern Branch of the Russian Academy of Sciences, 683006 Petropavlovsk-Kamchatsky, Russia
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(23), 4747; https://doi.org/10.3390/rs13234747
Submission received: 10 October 2021 / Revised: 18 November 2021 / Accepted: 21 November 2021 / Published: 23 November 2021
(This article belongs to the Special Issue Application of Machine Learning in Volcano Monitoring)

Abstract

:
Currently, video observation systems are actively used for volcano activity monitoring. Video cameras allow us to remotely assess the state of a dangerous natural object and to detect thermal anomalies if technical capabilities are available. However, continuous use of visible band cameras instead of special tools (for example, thermal cameras), produces large number of images, that require the application of special algorithms both for preliminary filtering out the images with area of interest hidden due to weather or illumination conditions, and for volcano activity detection. Existing algorithms use preselected regions of interest in the frame for analysis. This region could be changed occasionally to observe events in a specific area of the volcano. It is a problem to set it in advance and keep it up to date, especially for an observation network with multiple cameras. The accumulated perennial archives of images with documented eruptions allow us to use modern deep learning technologies for whole frame analysis to solve the specified task. The article presents the development of algorithms to classify volcano images produced by video observation systems. The focus is on developing the algorithms to create a labelled dataset from an unstructured archive using existing and authors proposed techniques. The developed solution was tested using the archive of the video observation system for the volcanoes of Kamchatka, in particular the observation data for the Klyuchevskoy volcano. The tests show the high efficiency of the use of convolutional neural networks in volcano image classification, and the accuracy of classification achieved 91%. The resulting dataset consisting of 15,000 images and labelled in three classes of scenes is the first dataset of this kind of Kamchatka volcanoes. It can be used to develop systems for monitoring other stratovolcanoes that occupy most of the video frame.

Graphical Abstract

1. Introduction

There are 30 active volcanoes in Kamchatka. Every year explosive, effusive and extrusive eruptions take place in this region, during which tons of volcanic products in the form of lava, pyroclastics, volcanic gases and aerosols come to the surface of the earth. These natural phenomena have an impact on the environment and pose a threat to the population and air traffic in the Pacific Northwest [1]. Due to the geographic location of volcanoes and insufficient ground-based scientific infrastructure, at present, the main type of their instrumental observations is Earth remote sensing systems. The appearance in recent years of spacecraft with specialized instruments (Himawari-8 [2], GOES-R [3], Sentinel-1 SAR, etc. [4]) and modern methods of processing satellite data made it possible to significantly improve the continuous monitoring of volcanic activity in the considered region. Based on new technologies and data sources, the VolSatView information system was created [5]. It allows volcanologists to comprehensively work with various satellite data, as well as meteorological and ground information on the Kuril-Kamchatka region. Due to the open architecture [6], the system receives data from other thematic services: modelling data for the movement of ash clouds [7], reference information on volcanoes [8], etc.
At the same time, it is important to note that many key tasks, for example, determining the time of an eruption, assessing its basic characteristics, can most accurately be solved using ground-based equipment installed in the vicinity of the volcano. These means include video observation networks. In 2009, joint efforts of scientists from the Institute of Volcanology and Seismology of the Far East Branch of the Russian Academy of Sciences and the Computing Center of the Far East Branch of the Russian Academy of Sciences created a video observation system [9] for visual assessment of the state of the volcanoes Klyuchevskoy, Sheveluch, Avachinsky and Gorely. The continuous operational mode of the system leads to the generation of a large number of images, while many of them contain no valuable information, because the object of interest or the whole volcano is obscured by clouds, illumination effect, etc. The use of simple cameras instead of special tools (for example, thermal cameras [10,11,12], which overcomes the effects of environmental conditions such as fog or clouds [13]), does not allow us to filter out such information quickly and efficiently from future consideration by volcanologists. In this regard, the technologies required can provide analysis and filtering of incoming data for further assessment of the state of volcanoes, including the detection of signs of potential eruptions.
The study of natural hazards and processes is usually based on the use of large amounts of data from various observing systems (GNSS networks, seismic observation networks, etc.). This largely determines the use in this area of modern methods of machine learning and neural networks, which make it possible to effectively solve the problems of analysing the instrumental data and predicting the occurrence of various events. In [14] different supervised learning algorithms applied to Ecuador, Haiti, Nepal, and South Korean earthquakes data to classify damage grades to buildings. A ground motion prediction by ANN of MLP-architecture is considered in [15], performing training, validating, and testing on the Indian strong motion database including 659 records from 138 earthquakes. Paper [16] considers the landslide displacement prediction by GNSS time series analysis carried out with LSTM network. In volcano activity monitoring, the ANN application to remote sensing data were used for recognition of volcanic lava flow hot spots [17], for the estimation of columnar content and height of sulphur dioxide plumes from volcanic eruptions [18]. Image analysis is one of the vast applications of machine learning and deep learning technologies. The paper [19] proposed a new method of detecting Strombolian eruptions by a trained convolutional neural network, which automatically categorize eruptions in infrared images obtained from the rim of the Mount Erebus crater, with a correct classification rate of 84%. The visible band camera images are considered in paper [20], in which the ANN of MLP-architecture is used for multilabel classification of Villarrica volcano images, with result accuracy of 93%.
The key aspect for the application of machine learning and deep learning methods is the preparation of a high-quality reference data set [14], labelled in accordance with the semantics of the study. This task is often solved manually, but when the amount of initial data is large, it requires the development of additional algorithms. The indicated problem is also valid for the considered video observation network.
The article presents the results of comprehensive studies and development of algorithms and approaches to classify the Klyuchevskoy volcano images and detect images with possible activity. The convolutional neural network is used for classification, and the focus of the paper is on developing the algorithms to create a labelled dataset from an unstructured archive using authors techniques.

2. Study Object and Motivation

2.1. Video Observation Network

The considered video observation network capturing Kamchatka volcanoes consists of 9 cameras, installed on the Kamchatsky kray territory in two locations (Table 1): in the village of Klyuchi in the F.Yu. Levinson-Lessing Kamchatkan volcanological station (cameras 1–5), and in the city of Petropavlovsk-Kamchatsky on the building of the Institute of Volcanology and Seismology of the Far Eastern Branch of the Russian Academy of Sciences (cameras 7–9). The processes of data collection, transmission and storage are controlled by the automated information system Signal [21].
Each camera poll interval depends on the aviation colour code of the volcano set by the KVERT (Kamchatkan Volcanic Eruption Response Team) [1] or is fixed due to individual observation mode and possible technical limitations. For this reason, the number of archive images for the same volcano in different years may be different.
In our study, we used the data archive, produced with camera 1 in 2017–2018, with a total image count of 871,110. This is a StarDot NetCam XL [22] visible band camera, with Computar TV Zoom Lens [23]. The camera watches the Klyuchevskoy volcano, with a distance to the crater of about 30 km. The main camera specification is summarized in Table 2, and the conceptual dataflow described by Figure 1.
This camera was chosen for study due to the stable operation of video and transmitting network equipment, as well as the fact that during the observation period from 2009 to present, numerous eruptions of the Klyuchevskoy volcano were documented. It is important for our study, since it allows us to use the results of studies of eruptions stored in KVERT VONA (Volcano Observatory Notice for Aviation) records, which were obtained, in particular, with remote sensing data processing in the VolSatView system.

2.2. Motivation

Real-time video observation network generates huge amounts of data. Up to 1440 images per day are sent to the archive by one camera of the network in Kamchatka, with 1 frame/min rate. At the same time, for a detailed analysis of the volcanic events (for example, to determine exact eruption times, plume velocity, etc.), the specified frame rate must be reduced by 5–10 times, which will certainly lead to a multiple increase in the volume of data generated. In the case of the active phase of the eruption, the operative analysis of such amount of data by volcanologists seems to be extremely difficult. It should be borne in mind that a certain part of the images does not contain valuable information, since the area of interest on them is hidden behind clouds, fog, bright sunlight or is not visible due to the night-time. In addition, during the eruption, the WAN link load also increases, due to the large volume of transmitted data, which, with limited telecommunication resources at observation points, also leads to errors in data transmission and to damaged images consequently. Considering the above problems, the task of filtering non-informative and damaged images is one of the key ones, the solution of which became possible with computer technologies.
Processing the ground-based cameras data with computer methods and technologies is successfully applied to detect and analyse volcano activity. In [24] an automatic image processor based on the cellular neural networks is proposed to process analogical telecamera recordings. The accumulative difference image technique is adopted to provide real-time warning in case of volcanic activity events, with data framerate 0.05–0.1 s. The real-time thermo-graphic analysis of IR-camera data used to manage frame rate and automatically classify the volcano activity in [13]. Computer vision methods and principal component analysis were used to determine lava flow trajectories for volcanoes of Ecuador in [25]. In paper [20], denoted to study of Villarrica volcano activity during February-April 2015, the BLOB detection algorithm is applied to each image in pre-selected region of interest, then a feature vector is extracted for detected BLOB. The artificial neural network of MLP-architecture was trained on manually labelled image dataset, providing five-class classification of image scenes.
The considered video observation network for Kamchatka volcanoes is equipped with visible band and VNIR band (IR-cut filter triggered at night-time) cameras, so thermographic data analysis methods cannot be applied. The minimal possible image frame rate is 1 min, due to limited link bandwidth at observation site. This does not allow us to successfully apply the algorithms for detecting volcanic activity based on dynamic characteristics. The sufficient framerate to detect relatively slow ash emissions is 1 frame per second, while 10–25 frames/s should be used for events with fast dynamic [24]. The use of computer vision methods from [25] for the volcanic events analysis requires a preselected image set with a clearly visible area of interest. In [20], this is implemented on manually prepared dataset using histogram analysis of colour channels in a pre-selected region of interest. However, for the video observation network under consideration, there are situations when the camera lens moves (to observe events in a specific area of the volcano) or the zoom changes. The boundaries of the area of interest in the image change consequently (Figure 2), so it is a problem to set them in advance and keep them up to date.
Considering the described peculiarities, we proposed to use methods based on the convolutional neural networks (CNN) as a tool for classifying images, which allows to analyse the whole frame, without highlighting the area of interest and preliminary feature vector extraction. With the help of a trained neural network, the continuous image stream can be automatically distributed over a given set of classes [26,27,28], solving the problem of input control of data informativity. The most important condition for training a neural network is the presence of labelled image dataset. Considering the variety of scenes in the images of volcanoes, the training set size should be large enough, and the set itself should include images taken in various weather and seasonal conditions. Given the huge size of the unstructured archive available (see Table 1), it is time consuming to create such a dataset manually. There are no publicly available labelled datasets of volcanoes. In this regard, we divided the solution of the volcano images classification problem into two interrelated parts—the development of algorithms to create a marked observational data set and consecutive training the neural network for classifying images.

3. Methods and Algorithms for Volcano Image Analysis and the Results of Their Approbation

To create and adopt a labelled volcano image dataset from the unstructured archive we propose a multistage approach, including the use of author‘s methods adapted to the task (Figure 3).
  • The source image archive is divided into daytime and night-time image sets with the unsupervised clustering approach.
    For daytime images, the visibility of the volcano on the image is assessed, which makes it possible to rank the images according to the degree of their potential informativeness.
  • Based on the joint analysis of the estimates obtained for the images and data from the KVERT reports on the activity of Klyuchevskoy volcano, a class-labelled dataset is assembled.
  • The created image dataset is used to train and test the convolutional neural network.
Figure 3. The flowchart diagram of the proposed algorithm to produce labelled dataset for CNN classification training.
Figure 3. The flowchart diagram of the proposed algorithm to produce labelled dataset for CNN classification training.
Remotesensing 13 04747 g003
Below is the detailed description of the proposed methods and algorithms.

3.1. Day/Night Image Clustering

The considered video camera has no IR-cut filter, therefore images generated by that camera at night-time contain no useful information due to very low illumination. So those images should be determined and filtered. The simplest approach could be to calculate times of different twilight types (civil dawn (sunset begin), sunset end, sunrise begin, civil dusk (sunrise end), etc.) at given geographical point using the sunrise equation [29], and then to compare those times with the image timestamp. However, it was found that this approach doesn’t give stable criteria to consider image scene as night, because sometimes actual illumination remains enough to produce informative images for some period after sunset end (or some period before sunrise begin). In addition, sometimes the correct image timestamp is not available due to camera time synchronization errors. So, it is better to rely on the image itself. An image analysis approach is proposed in [30], applying a histogram statistics thresholding in HSV colour space to detect day and night scenes. It was improved in [31], resulting in scene classification precision of 96%. However, this method requires proper threshold parameters selection to meet considered image semantics.
Our approach is to adopt an unsupervised clustering, provided by DCEC architecture from paper [32]. It proposes a special clustering layer attached to embedding layer of convolutional autoencoder neural network. In the pretraining phase, latent features are extracted from input data, and autoencoder network weights adjusted to reconstruct input images using MSE loss function. Then a given number of clusters are initialized with k-means for features extracted, and further training updates feature clustering distribution, using Kullback-Leibler divergence as a loss function. The resulting neural network weights could be used to split input dataset into a given number of clusters. In [32], authors use DCEC to clusterize MNIST dataset consisting of 70,000 handwritten digits of size 28 × 28.
To train the DCEC model, we used Klyuchevskoy volcano archive for 2018 (351,318 images) as an input dataset. Because archive temporal resolution is 1 min, the adjacent images have minimal difference in day/night semantics, so only every third archive item was used for pretraining, resulting in a dataset of 117,106 images. Input images were resized to 256 × 256 and normalized. Pretraining was performed for 300 epochs with a batch size of 256. Label divergence has achieved 0.01 tolerance at the first train epoch. The trained model was tested over the 2018 archive. The results of clustering were verified, where possible, with sunrise equation output. The civil dusk and civil dawn times were used as approximate thresholds to consider image scene day or night. All the cases, where clustering result differs from verification, were checked manually. Figure 4 shows how clustering results differ from sunset and sunrise time periods and gives a more accurate day/night image dataset split. The only 266 of total 351,318 images (0.07%) were clustered incorrectly (Figure 5).

3.2. Parametric Methods for Volcano Visibility Estimate

A large number of day images are unusable for further analysis because the region of interest is obscured by clouds, fog, bright sunlight artifacts, etc. For preliminary quality ranking of images, we developed an algorithm, which estimates volcano visibility on the image. The basic idea is to represent volcano contours as polylines, with branching in vertexes. Defined by a set of parameters, these polylines are called volcano parametric contours. Comparing parametric contours for image with parametric contours for the preselected reference image, we can estimate the visibility of the observed object.
If compared contours are extracted with edge detection algorithm, then the distance between edge pixels on different images is measured. This distance could be the same for several pixels relative to some reference pixels, which gives false response in contours intersection areas and leads to multiple comparisons of one contour of the image with different contours of another image. With parametric contours, we analyse how one polyline falls on another polyline in shape, while all the segments of this polyline may not coincide at all, especially if they are small. The average distance between the two polylines plays the main role. This allows us to handle the camera possible shifts.

3.2.1. Parametric Contours Building

Parametric contouring building begins with a discrete edges map, extracted using the Canny edge detector [33]. Then, on the map, the sequences of pixels are selected between the branch points and the ends of the paths (Figure 6a). After that, the obtained pixel chains are used to construct polylines, by recursively dividing the chains into straight lines connecting the starting and ending pixels. The split point is selected in the pixel, the distance from which to the line is maximum, if this distance exceeds the d m a x threshold. An example of constructing a polyline along a chain of pixels is shown in Figure 6b. Result of constructing parametric contours is shown in Figure 7.
The procedure for comparing parametric contours obtained for two images is described in [34]. The result of the comparison is expressed by the estimate σ :
σ = j S ¯ i = 1 n j l s j i ,   s j i s S s ,
S ¯ —set of numbers of compared polylines,
n j —the number of compared segment pairs for polyline j ,
S and S —set of segments for first and second image respectively,
s j i S ,   s j i S —compared segments,
s S —segments of the second image,
l s j i ,   s j i —common part length of segments s and s ,
s —length of segment s .
This estimate shows the occurrence of the contours of the first image in the second (reference) image, which is equal to the ratio of the total length of the matched contours to the total length of the contours in the second image.

3.2.2. Finding the Shift between Contours

Due to the impact on the camera of environmental factors (wind load, temperature variations, etc.), as well as due to small manual turns of the camera to change the observation area, the contours in different images may have an offset relative to each other. The study of the image archive showed that the relative shift between the contours can reach 10% of the image width. Therefore, to align the parametric contours relative to each other, it is proposed to use the shift δ calculated using the discrete edges map. The use of the δ vector makes it possible to compensate for relative displacements of more than 20% of the image width.
Let the images received from the camera have dimensions w × h pixels, then each pixel of this image has the number y w + x , where x = 0 ,   1 , ,   w 1 , y = 0 ,   1 , ,   h 1 —coordinates of the pixel in the image. Using the Canny edge pixel detection algorithm for the image under study, we calculate the set G , consisting of the numbers of the edge points. The difference score between two sets of edge pixels numbers G 1 and G 2 , obtained for two images, is calculated as:
μ G 1 , G 2 = 1 G 1 n G 1 min min k G 2   d n , k d c u t , 1.0 ,
where d n , k is the distance between pixels with numbers n and k , and d c u t is a maximum distance between pixels which could be compared. The search for the minimum distance in (1) is carried out using the distance map to the edge points G 2 .
To align the parametric contours, consider the shifted edges points set G δ = n δ y w δ x | n G δ x n x < w + δ x δ y n y < h + δ y , where n x = n   m o d   w , n y = n / w , δ = δ x ,   δ y 2 . Using this set, we find the shift as:
δ G 1 , G 2 = arg   min δ   μ G 1 δ , G 2 .
Shift calculation in (2) is carried out on the image reduced by s times, where s = min w , h / 100 , and with d c u t 3 s . When shift for reduced image δ G 1 , G 2 is found, the shift for the original-size image is calculated with gradient descent method, starting from pint s δ G 1 , G 2 , d c u t max 5 , s . The function μ G 1 δ , G 2 may not have one clear minimum, therefore, to initialize the gradient descent method, it is necessary to take up to 5 minimums, which are worse than the global minimum by no more than 20%.

3.2.3. Building Reference Contours

To assess the volcano visibility on the image, the parametric contours extracted for it are compared with some reference contours, which are constructed for m preselected images taken in good weather conditions with good illumination and at different seasons. Such images are called reference images, their number 3 m 7 was determined during our experiments.
For reference image i, i = 1 ,   2 , ,   m , we find G i and calculate δ G i , G k , k = arg   max k G k . The shift is found as:
δ i = δ G i , G k 1 2 min l = 1 , 2 , , m δ G l , G k + max l = 1 , 2 , , m δ G l , G k ,
The parametric contours of reference image i are found as described in Section 3.2.1 and then shifted by δ i . The contours for all reference images are compared in pairs, and those contours that occur in at least γ · m images, are selected as reference ones. The γ is selected in the range from 0.5 to 1.0, it allows to control the contribution of the most characteristic contours of the considered volcano to obtain the most accurate estimates for the studied images (Figure 8).
The extracted reference contours are divided into internal and external ones. The external contours are composed of line segments that have a maximum y-coordinate for point of intersection with vertical lines drawn with one pixel distance. The rest of the contours are considered internal. This separation allows us to correctly analyse images taken, for example, in conditions of sunlight, when only the general outline of the volcano is visible. In this case, analysis on a common set of contours will result in an underestimated visibility, while the outer contours of the volcano will be fully visible.
The visibility estimate contours for the tested image is calculated by the formula:
σ = max min 1 , σ e x t min i = 1 , 2 ,   ,   m σ ¯ e x t , i , min 1 , σ i n t min i = 1 , 2 ,   ,   m σ ¯ i n t , i ,
where σ e x t and σ i n t —the occurrence of external and internal reference contours on the tested image, respectively, a σ ¯ e x t , i and σ ¯ i n t , i —the occurrence of external and internal reference contours in the reference image i.

3.2.4. Image Frequency Characteristic Estimate

Partial clouds in the frame can affect the contour visibility estimates in different ways. For example, it can be overestimated due to the partial coincidence of the cloud contours with the reference contours or underestimated when the lower part of the frame is hidden behind the clouds, but at the same time the upper part of the volcano with a potential region of interest is visible. To take this effect into account, we additionally estimated the contribution of the frequencies involved in the formation of the image.
Let us denote the image brightness component as I = I x y , x = 1 ,   2 , ,   w , y = 1 ,   2 ,   ,   h , where w и h are image width and height respectively. To determine which frequencies are involved in the formation of I , and what contribution they make, we divide the frequency spectrum into octaves. It is most convenient to do this if the width and height of the image are powers of two: w = 2 n w , h = 2 n h , n w , n h . Since the considered images have a size of 1024 × 768, they were scaled to 1024 × 1024, and n = n w = n h = 10 . The contribution vector f of each octave to the formation of the image is called the vector of the contribution of frequencies or the frequency characteristic of the image. The vector f = f i , i = 0 ,   1 ,   ,   n 1 , is calculated using the following formulas:
f i = 1 w i h i x = 1 w i y = 1 h i I x y i I x + 1 / 2 , y + 1 / 2 i + 1 ,
w k =   1 2 w k 1 ,   k > 0 w ,             k = 0   ,       h k =   1 2 h k 1 ,   k > 0 h ,             k = 0 ,
I x y k = I 2 x 1 , 2 y 1 k 1 + I 2 x , 2 y 1 k 1 + I 2 x 1 , 2 y k 1 + I 2 x , 2 y k 1 4 ,     k > 0 I x y ,     k = 0 .
Like in the approach for contours, the frequency characteristics of the analysed images are compared with the reference characteristic f ¯ calculated for the reference images:
f ¯ = 1 m i = 1 m f I i ,
where I i i-th reference image, f I i —frequency characteristic of i-th reference image. To compare the frequency characteristics f with the reference f ¯ , the following estimate is used:
ρ = max min b i = 0 n 1 f ¯ i f i j = 1 3 w j ω j i ,   1 , 0 ,
where ω j i are base function, which determines the influence of the parameters w j on individual groups of frequencies (Figure 9), b and w j , j = 1 , 2 , 3 are constant parameters that are calculated experimentally. To find them, we choose L images with different scenes, and manually define for them a range of estimates: [ ρ ˜ m i n f l ,   ρ ˜ m a x f l ], where f l —frequency characteristics of l-th image, l = 1 ,   2 , ,   L . The values of b and w j are determined by solving the following minimization problem:
b , w j = arg   min b , w j ε b , w j ,
ε b , w j = l = 1 L d ρ f l , b , w j , ρ ˜ m i n f l ,   ρ ˜ m a x f l 2 .
where d is a distance function, defined as:
d a , b , c = 0 ,             b a c b a ,             a < b a c ,             a > c .
Values of b and w j are constant for each camera. Table 3 contains the parameters values for considered camera.
The Figure 10 shows the different Klyuchevskoy volcano scenes with their corresponding ρ -estimates and frequency characteristics comparing to reference ones, calculated by Equations (3)–(6).
The result volcano visibility score α for the image is determined as follows:
α = σ f σ + ρ 1 f σ ,
f σ = min 1 , 1 Δ 2 σ τ 2 .
Contours visibility estimate σ makes the main contribution to α . If σ is in the Δ -vicinity of defined threshold τ , then frequency characteristics ρ -estimate is used for result score correction.

3.3. Emerging the Labelled Dataset and CNN Training

The proposed algorithms, described in Section 3.1 and Section 3.2, were applied to the Klyuchevskoy volcano image archive for 2017, which includes 519,792 items. At the first stage, the defective images, generated in poor transmission network conditions, were filtered, by checking the last two bytes of each image, which should be equal to ‘0xff0xd9′ for correct JPEG-images. At the second stage, the inference of trained DCEC model from Section 3.1, was carried out, dividing archive in two clusters: 302,520 of day and 217,272 of night images. Next, the visibility estimation algorithm was applied to the day cluster. Figure 11 shows the distribution of images by calculated visibility estimate α .
The proposed visibility ranking approach allows to sort thousands of images by their potential usability. This made it possible to create a labelled dataset for future CNN training.
For day images, we define the next three classes, which describes the most distinctive and important states:
  • unclear—region of interest is not visible because of clouds, different sunlight effects, etc.
  • clear—good visibility, no significant volcanic activity.
  • active—good visibility, volcano activity detected.
To assemble the dataset, labelled with defined classes, we completed the following steps.
  • In the subset of images with α < 0.3, the 5000 images were randomly selected and labelled as class «unclear».
  • In the subset of images with α > 0.8 (good visibility rank), we select a series with Klyuchevskoy volcano activity/inactivity, using VONA/KVERT information releases [35] as a reference. Using this approach, 5000 images of class «active» and 5000 images of class «clear» were selected. During this procedure, the images were found, which have partially visible contours and estimate corrected with frequency characteristics, so, have a good potential usability. However, the region of interest (volcano crater) on them is closed by cloud or sunspot. Such images give poor information for determining possible volcano activity. Considering that, such images found were revised by experts and included in «unclear» subset. To keep the size subset the same, the corresponding number of random images selected at step 1. were deleted. Images with partial contours visibility and the region of interest obscured account 24% of «unclear» subset.
The total dataset size is 15,000 images, with an equal number of images for each class to make it balanced.
There are different state-of-art deep CNN architectures for feature extraction and image classification: ResNet [36], DenseNet [37], EfficientNet [38]. The last one has the smallest memory footprint and performs faster [38], so we adopted an EfficientNet-B0 architecture for volcano images classification. It consists of input convolutional layer with 3-channel input with size 224 × 244, followed by sequence of Mobile Inverted Residual Bottleneck Blocks [38] for feature maps extraction (Figure 12). The final convolution layer is the adaptive 2d pooling followed by a fully connected layer of size 3, by number of predicted classes.
We conduct model training from scratch and didn’t use any pretrained model weights (i.e., on ImageNet database [39]). The Nvidia Tesla P100 GPU was used for calculations. No special input data augmentation was applied, except a resize to 224 × 224 and normalization. Instead of standard ImageNet per-channel mean and std values, the following values were calculated for our dataset:
m e a n =   0.4237 ,   0.4078 ,   0.3958
s t d = 0.2341 ,   0.2291 ,   0.2340
The 10-fold cross-validation was applied to dataset, with the StratifiedKFold from Scikit-learn package [40]. This preserves class balance in different folds. The model was trained and validated 10 times with the following hyperparameters:
  • number of epochs—40;
  • batch size—64;
  • learning rate—0.001.
Figure 13 shows training and validation loss by epoch for each cross-validation fold.
The metric averaged by cross-validation folds are shown on Figure 14. The best average accuracy of 91% with a standard deviation of 5.28 is achieved at 24 epoch.
The confusion matrix overall folds presented on Figure 15. The model predicts ‘unclear’ class better, while the ‘active’ class the most difficult to predict correctly.
Figure 16 shows examples of correctly classified images.
The big variety of scenes with clouds leads to incorrect classifications for some images. Figure 17 shows examples of misclassification:
  • (a)—image marked as ‘unclear’, while actual volcano rim visibility allows to observe possible activity.
  • (b)—image, false labelled as ‘active’ due to cloud cover.
  • (c)—the scene marked as ‘unclear’, but ash plume is visible behind the scattered clouds.
  • (d)—actual light emission on image classified as clear because light emission is almost masked by the volcano itself.
Figure 17. Examples of incorrectly classified scenes of Klyuchevskoy volcano: (a) image with true label ‘clear’ classified as ‘unclear’; (b) image with true label ‘unclear’ classified as ‘active’; (c) image with true label ‘active’ classified as ‘unclear’; (d) image with true label ‘active’ classified as ‘clear’.
Figure 17. Examples of incorrectly classified scenes of Klyuchevskoy volcano: (a) image with true label ‘clear’ classified as ‘unclear’; (b) image with true label ‘unclear’ classified as ‘active’; (c) image with true label ‘active’ classified as ‘unclear’; (d) image with true label ‘active’ classified as ‘clear’.
Remotesensing 13 04747 g017

4. Discussion

Analysing the results of the conducted study, we conclude that the unsupervised clustering is an approach to determine day and night images, which is possibly applicable for any visible band camera. We conducted experiments on inference the DCEC model which was trained on the Klyuchevskoy volcano image archive, with data archive for 2018 of the camera capturing the Sheveluch volcano, and received good results for day/night clusters. This is achieved by sufficient size of the training set, so when model extracts feature maps from many images and adjusts weights to cluster embedded features into two classes, it finally ignores any object edges semantics, focusing on general illumination semantics. For cameras considered in this paper with frame rate 1/60 s we assembled training dataset from a one-year archive, taking every third image. This rule, perhaps, should be expanded for any other camera with higher frame rate (i.e., take every 10th image, for example), or reduced for lower frame rate, using all archive for DCEC model training.
The developed algorithms for volcano visibility estimate in the image cannot be used for a full analysis of volcanic events, since they only consider the visibility of the observed object. However, having this image ranking approach, we exclude a big amount of uninformative data from further analysis. This makes it possible to efficiently work with observation data.
The EfficientNet convolutional neural network trained on labelled dataset shows the average accuracy of 91% in classification daytime images. A distinctive feature of our proposed solution is that the image classification as well as detection of images with possible volcano activity is carried out without specifying the region of interest, like it is done in paper [20]. This makes it possible to use them for an automated classification of images continuously received from the observation network under the conditions when camera lens position is changed. At the same time, our approach does not extract specific signs of volcanic activity, like lava flows [25], but allows to categorize image itself, filtering out the images with no valuable information and detecting images with possible volcano activity, which can further be analysed in detail. At the same time, due to the distinctive features of the relief of Klyuchevskoy volcano, the neural network trained on its images will show a lower classification accuracy (in ‘clear’ and ‘active’ classes) for images of volcanoes, whose visual contour differs significantly (i.e., Sheveluch volcano). This may require assembling the additional dataset with the implemented ranking algorithm, and the subsequent CNN training using the transfer learning methods.
The obtained results show the efficiency of convolutional neural networks application to problems of classification of volcano images. An analysis of cases with misclassified images suggests the next promising study direction—semantic segmentation of ash and steam-gas emissions using neural networks, for example, based on the Mask-RCNN or UNet architecture. This will allow to detect signs of volcanic activity in conditions, where image region of interest is partially obscured by clouds or other phenomena. With emission segmented on image, it is also possible to analyse volcanic events in more detail. However, assembling the labelled dataset for training these kinds of neural networks is an even more time-consuming task and will be a subject of our future research.

5. Conclusions

The article presents the results of the development of algorithms for volcano image labelling and classification. The proposed solutions were adopted on the archive of the video observation system for the volcanoes of Kamchatka. The Klyuchevskoy volcano image dataset was assembled, consisting of 15,000 images, and marked into three classes of scenes (public link provided in Data Availability Statement). It is the first dataset of this kind for the volcanoes of Kamchatka. The CNN trained with this dataset could further be integrated into online processing of continuous volcano video observations. Presented results can be applied to volcano images captured with visible band cameras to study volcanic activity of other stratovolcanoes that occupy most of the video frame, like Popocatepetl and Colima volcanoes of Mexico (https://webcamsdemexico.com/webcams/volcanes/, accessed on 21 November 2021).

Author Contributions

Conceptualization, S.K., A.K. and A.S.; validation, S.K. and I.U.; methodology, A.K. and S.K.; manuscript writing, S.K., A.S., I.U., A.K. and O.G.; data processing, S.K. and I.U. All authors discussed the results and contributed to the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

The reported study was funded by the Russian Foundation for Basic Research, project number 20-37-70008.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available Klyuchevskoy volcano image labelled dataset was assembled in this study. This data can be found here: http://dl.signal.febras.net/KLYU2017_labelled.zip (accessed on 21 November 2021).

Acknowledgments

The studies were carried out using the resources of the Center for Shared Use of Scientific Equipment “Center for Processing and Storage of Scientific Data of the Far Eastern Branch of the Russian Academy of Sciences” [41], funded by the Russian Federation represented by the Ministry of Science and Higher Education of the Russian Federation under project No. 075-15-2021-663.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

GOESgeostationary operational environmental satellite
GNSSglobal navigation satellite system
ANNartifitial neural network
MLPmultilayer perceptron
WANwide area network
KVERTKamchatkan Volcanic Eruption Response Team
VONAvolcano observatory notice for aviation
IRinfrared
BLOBbinary large object
VNIRvisible and near-infrared
CNNconvolutional neural network
HSVhue, saturation, value
DCECdeep convolutional embedded clustering
MSEmean squared error
GPUgraphics processing unit

References

  1. Girina, O.A.; Loupian, E.A.; Sorokin, A.A.; Melnikov, D.V.; Romanova, I.M.; Kashnitskii, A.V.; Uvarov, I.A.; Malkovsky, S.I.; Korolev, S.P.; Manevich, A.G.; et al. Comprehensive Monitoring of Explosive Volcanic Eruptions of Kamchatka; IVS DVO RAN: Petropavlovsk-Kamchatsky, Russia, 2018; 192p. [Google Scholar]
  2. Girina, O.; Kramareva, L.; Loupian, E.; Melnikov, D.; Manevich, A.; Sorokin, A.; Uvarov, I.; Kashnitskiy, A.; Bourtsev, M.; Marchenkov, V.V.; et al. The use of Himawari satellite data for monitoring Kamchatka volcanoes. Sovrem. Probl. Distantsionnogo Zondirovaniya Zemli Kosm. 2017, 14, 65–76. [Google Scholar] [CrossRef]
  3. Horváth, Á.; Girina, O.A.; Carr, J.L.; Wu, D.L.; Bril, A.A.; Mazurov, A.A.; Melnikov, D.V.; Hoshyaripour, G.A.; Buehler, S.A. Geometric estimation of volcanic eruption column height from GOES-R near-limb imagery—Part 2: Case studies. Atmos. Chem. Phys. Discuss. 2021, 21, 12207–12226. [Google Scholar] [CrossRef]
  4. Cigna, F.; Tapete, D.; Lu, Z. Remote Sensing of Volcanic Processes and Risk. Remote Sens. 2020, 12, 2567. [Google Scholar] [CrossRef]
  5. Gordeev, E.I.; Girina, O.A.; Lupyan, E.A.; Sorokin, A.; Kramareva, L.S.; Efremov, V.Y.; Kashnitskii, A.V.; Uvarov, I.A.; Burtsev, M.A.; Romanova, I.; et al. The VolSatView information system for Monitoring the Volcanic Activity in Kamchatka and on the Kuril Islands. J. Volcanol. Seismol. 2016, 10, 382–394. [Google Scholar] [CrossRef]
  6. Korolev, S.P.; Romanova, I.M.; Malkovsky, S.I.; Sorokin, S.I. Service-oriented interface to access scientific data for study and state operational monitoring of volcanoes of kamchatka and northern kuriles. Syst. Means Inform. 2018, 28, 88–98. [Google Scholar] [CrossRef]
  7. Malkovsky, S.I.; Sorokin, A.A.; Korolev, S.P. Improving the system of numerical simulation of volcanic ash propagation using the PUFF model. Russ. J. Earth Sci. 2017, 17, ES5003. [Google Scholar] [CrossRef] [Green Version]
  8. Romanova, I.M.; Girina, O.A.; Manevich, A.G.; Melnikov, D.V.; Gorbach, N.V. Information systems VOKKIA and KVERT for the analysis of the activity of the volcanoes of Kamchatka and the Kuriles. In Proceedings of the V International Scientific and Practical Conference “Information Technologies and High-Performance Computing”, Khabarovsk, Russia, 16–19 September 2019; Pacific State University: Khabarovsk, Russia, 2019; pp. 278–282. [Google Scholar]
  9. Sorokin, A.A.; Korolev, S.P.; Romanova, I.M.; Girina, O.A.; Urmanov, I.P. The Kamchatka Volcano Video Monitoring System. In Proceedings of the 6th International Workshop on Computer Science and Engineering, Tokyo, Japan, 17–19 June 2016; Science and Engineering Institute: Los Angeles, CA, USA, 2016; pp. 734–737. [Google Scholar]
  10. Sansivero, F.; Vilardo, G. Processing Thermal Infrared Imagery Time-Series from Volcano Permanent Ground-Based Monitoring Network. Latest Methodological Improvements to Characterize Surface Temperatures Behavior of Thermal Anomaly Areas. Remote Sens. 2019, 11, 553. [Google Scholar] [CrossRef] [Green Version]
  11. Bombrun, M.; Jessop, D.; Harris, A.; Barra, V. An algorithm for the detection and characterisation of volcanic plumes using thermal camera imagery. J. Volcanol. Geotherm. Res. 2018, 352, 26–37. [Google Scholar] [CrossRef]
  12. Wood, K.; Thomas, H.; Watson, M.; Calway, A.; Richardson, T.; Stebel, K.; Naismith, A.; Berthoud, L.; Lucas, J. Measurement of three dimensional volcanic plume properties using multiple ground based infrared cameras. ISPRS J. Photogramm. Remote Sens. 2019, 154, 163–175. [Google Scholar] [CrossRef]
  13. Andò, B.; Pecora, E. An advanced video-based system for monitoring active volcanoes. Comput. Geosci. 2006, 32, 85–91. [Google Scholar] [CrossRef]
  14. Harirchian, E.; Kumari, V.; Jadhav, K.; Rasulzade, S.; Lahmer, T.; Das, R.R. A Synthesized Study Based on Machine Learning Approaches for Rapid Classifying Earthquake Damage Grades to RC Buildings. Appl. Sci. 2021, 11, 7540. [Google Scholar] [CrossRef]
  15. Sreejaya, K.P.; Basu, J.; Raghukanth, S.T.G.; Srinagesh, D. Prediction of Ground Motion Intensity Measures Using an Artificial Neural Network. Pure Appl. Geophys. 2021, 178, 2025–2058. [Google Scholar] [CrossRef]
  16. Wang, J.; Nie, G.; Gao, S.; Wu, S.; Li, H.; Ren, X. Landslide Deformation Prediction Based on a GNSS Time Series Analysis and Recurrent Neural Network Model. Remote Sens. 2021, 13, 1055. [Google Scholar] [CrossRef]
  17. Piscini, A.; Lombardo, V. Volcanic hot spot detection from optical multispectral remote sensing data using artificial neural networks. Geophys. J. Int. 2014, 196, 1525–1535. [Google Scholar] [CrossRef] [Green Version]
  18. Piscini, A.; Carboni, E.; Del Frate, F.; Grainger, R.G. Simultaneous retrieval of volcanic sulphur dioxide and plume height from hyperspectral data using artificial neural networks. Geophys. J. Int. 2014, 198, 697–709. [Google Scholar] [CrossRef] [Green Version]
  19. Dye, B.C.; Morra, G. Machine learning as a detection method of Strombolian eruptions in infrared images from Mount Erebus, Antarctica. Phys. Earth Planet. Inter. 2020, 305, 106508. [Google Scholar] [CrossRef]
  20. Witsil, A.J.; Johnson, J.B. Volcano video data characterized and classified using computer vision and machine learning algorithms. Geosci. Front. 2020, 11, 1789–1803. [Google Scholar] [CrossRef]
  21. Sorokin, A.; Korolev, S.; Malkovsky, S.; Ras, C.C.F. The Signal automated information system: Research and operational monitoring of dangerous natural phenomena in the Russian Far East. Sovrem. Probl. Distantsionnogo Zondirovaniya Zemli Kosm. 2019, 16, 238–248. [Google Scholar] [CrossRef]
  22. Netcam XL Brochure. Available online: http://www.stardot-tech.com/netcamxl/netcam-xl-brochure.pdf (accessed on 1 November 2021).
  23. Computar H6Z0812M. Available online: https://computar.com/resources/files_v2/224/H6Z0812M_Spec0604.pdf (accessed on 1 November 2021).
  24. Bertucco, L.; Coltelli, M.; Nunnari, G.; Occhipinti, L. Cellular neural networks for real-time monitoring of volcanic activity. Comput. Geosci. 1999, 25, 101–117. [Google Scholar] [CrossRef]
  25. Viteri, F.; Barrera, K.; Cruz, C.; Mendoza, D. Using computer vision techniques to generate embedded systems for monitoring volcanoes in Ecuador with trajectory determination. J. Eng. Appl. Sci. 2018, 12, 3164–3168. [Google Scholar]
  26. Zou, Z.; Gan, H.; Huang, Q.; Cai, T.; Cao, K. Disaster Image Classification by Fusing Multimodal Social Media Data. ISPRS Int. J. Geo-Inf. 2021, 10, 636. [Google Scholar] [CrossRef]
  27. Yahya, A.A.; Tan, J.; Hu, M. A Novel Handwritten Digit Classification System Based on Convolutional Neural Network Approach. Sensors 2021, 21, 6273. [Google Scholar] [CrossRef]
  28. Shi, C.; Tan, C.; Wang, T.; Wang, L. A Waste Classification Method Based on a Multilayer Hybrid Convolution Neural Network. Appl. Sci. 2021, 11, 8572. [Google Scholar] [CrossRef]
  29. Wikipedia: Sunrise Equation. Available online: https://en.wikipedia.org/wiki/Sunrise_equation (accessed on 1 September 2021).
  30. Saha, B.; Davies, D.; Raghavan, A. Day Night Classification of Images Using Thresholding on HSV Histogram. U.S. Patent US9530056B2, 27 December 2016. [Google Scholar]
  31. Zayed, H.; Taha, M.; Khalifa, T. Day/night detector for vehicle tracking in traffic monitoring systems. Int. J. Comput. Inf. Eng. 2016, 10, 98–104. [Google Scholar]
  32. Guo, X.; Liu, X.; Zhu, E.; Yin, J. Deep Clustering with Convolutional Autoencoders. In Lecture Notes in Computer Science, International Conference on Neural Information Processing; Liu, D., Xie, S., Li, Y., Zhao, D., El-Alfy, E.S., Eds.; Springer: Cham, Switzerland, 2017; Volume 10635, pp. 373–382. [Google Scholar] [CrossRef]
  33. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef]
  34. Kamaev, A.N.; Urmanov, I.P.; Sorokin, A.; Karmanov, D.A.; Korolev, S.P. Images analysis for automatic volcano visibility estimation. Comput. Opt. 2018, 42, 128–140. [Google Scholar] [CrossRef]
  35. KVERT: Volcanic Danger Prognosis for Aviation. Available online: http://www.kscnet.ru/ivs/kvert/van/index?type=1 (accessed on 1 September 2021).
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef] [Green Version]
  37. Huang, G.; Liu, Z.; Van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  38. Tan, M.; Le, Q.V. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning (ICML-19), Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
  39. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.F. Imagenet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  40. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  41. Sorokin, A.A.; Makogonov, S.V.; Korolev, S. The Information Infrastructure for Collective Scientific Work in the Far East of Russia. Sci. Tech. Inf. Process. 2017, 44, 302–304. [Google Scholar] [CrossRef]
Figure 1. The conceptual observation network dataflow on the example of videocamera capturing the Klyuchevskoy volcano. Images fetched periodically and transmitted via wide area network (WAN) to automated information system (AIS) “SIGNAL”.
Figure 1. The conceptual observation network dataflow on the example of videocamera capturing the Klyuchevskoy volcano. Images fetched periodically and transmitted via wide area network (WAN) to automated information system (AIS) “SIGNAL”.
Remotesensing 13 04747 g001
Figure 2. Examples of region of interest shift for different observation periods of Klyuchevskoy volcano: (a) 6 July 2016; (b) 19 December 2016; (c) 17 July 2017; (d) 11 August 2017; (e) 7 May 2018.
Figure 2. Examples of region of interest shift for different observation periods of Klyuchevskoy volcano: (a) 6 July 2016; (b) 19 December 2016; (c) 17 July 2017; (d) 11 August 2017; (e) 7 May 2018.
Remotesensing 13 04747 g002
Figure 4. The results of images day/night clustering (blue line, 0—night, 1—day) in compare with: (a) sunset period (dotted line—sunset begin; dashed line—sunset end); (b) sunrise period (dashed line—sunrise begin, dotted line—sunrise end).
Figure 4. The results of images day/night clustering (blue line, 0—night, 1—day) in compare with: (a) sunset period (dotted line—sunset begin; dashed line—sunset end); (b) sunrise period (dashed line—sunrise begin, dotted line—sunrise end).
Remotesensing 13 04747 g004
Figure 5. Examples of incorrectly clustered images: (a) Klyuchevskoy volcano daytime image at 03:01 a.m. 3 January 2018, false labelled as night; (b) Klyuchevskoy volcano nighttime image at 13:05 p.m. 20 December 2018, false marked as day.
Figure 5. Examples of incorrectly clustered images: (a) Klyuchevskoy volcano daytime image at 03:01 a.m. 3 January 2018, false labelled as night; (b) Klyuchevskoy volcano nighttime image at 13:05 p.m. 20 December 2018, false marked as day.
Remotesensing 13 04747 g005
Figure 6. The process of parametric contours building: (a) extraction of branch points (red crosses) and of edge points (blue dots) from discrete edge map; (b) recursive multistep process to split pixel chain and construct polyline: at each step split point is selected in pixel with maximum distance to the line, if this distance exceeds the d m a x threshold.
Figure 6. The process of parametric contours building: (a) extraction of branch points (red crosses) and of edge points (blue dots) from discrete edge map; (b) recursive multistep process to split pixel chain and construct polyline: at each step split point is selected in pixel with maximum distance to the line, if this distance exceeds the d m a x threshold.
Remotesensing 13 04747 g006
Figure 7. The example of parametric contours built for image of Klyuchevskoy volcano.
Figure 7. The example of parametric contours built for image of Klyuchevskoy volcano.
Remotesensing 13 04747 g007
Figure 8. Examples of reference contours of Klyuchevskoy volcano, extracted for different values of γ : (a) γ = 0.5 —contours occurred in at least 50% of reference images; (b) γ = 1.0 —contours occurred in all reference images.
Figure 8. Examples of reference contours of Klyuchevskoy volcano, extracted for different values of γ : (a) γ = 0.5 —contours occurred in at least 50% of reference images; (b) γ = 1.0 —contours occurred in all reference images.
Remotesensing 13 04747 g008
Figure 9. The graphs of used base functions, which determines the influence of the parameters w j on high, mid and low frequencies.
Figure 9. The graphs of used base functions, which determines the influence of the parameters w j on high, mid and low frequencies.
Remotesensing 13 04747 g009
Figure 10. Graphs for frequency characteristics for different images of Klyuchevskoy volcano in compare with reference characteristics f ¯ : (a) volcano is obscured by clouds ( ρ = 0.08 ), (b) volcano is partially obscured by clouds ( ρ = 0.38 ), (c) volcano is clearly visible ( ρ = 0.99 ).
Figure 10. Graphs for frequency characteristics for different images of Klyuchevskoy volcano in compare with reference characteristics f ¯ : (a) volcano is obscured by clouds ( ρ = 0.08 ), (b) volcano is partially obscured by clouds ( ρ = 0.38 ), (c) volcano is clearly visible ( ρ = 0.99 ).
Remotesensing 13 04747 g010
Figure 11. A distribution of images count by visibility estimate. Each bar represents count of images with estimate below the corresponding x-value.
Figure 11. A distribution of images count by visibility estimate. Each bar represents count of images with estimate below the corresponding x-value.
Remotesensing 13 04747 g011
Figure 12. The EfficientNet-B0 architecture diagram used for volcano image classification.
Figure 12. The EfficientNet-B0 architecture diagram used for volcano image classification.
Remotesensing 13 04747 g012
Figure 13. Loss graphs for convolutional neural network training and validation for all cross-validation folds: (a) training loss by epoch; (b) validation loss by epoch.
Figure 13. Loss graphs for convolutional neural network training and validation for all cross-validation folds: (a) training loss by epoch; (b) validation loss by epoch.
Remotesensing 13 04747 g013
Figure 14. Loss and validation accuracy during convolutional neural network training and validation, averaged for all cross-validation folds: (a) training and validation loss by epoch; (b) validation accuracy value by epoch. The thick orange line is the average over all folds.
Figure 14. Loss and validation accuracy during convolutional neural network training and validation, averaged for all cross-validation folds: (a) training and validation loss by epoch; (b) validation accuracy value by epoch. The thick orange line is the average over all folds.
Remotesensing 13 04747 g014
Figure 15. The confusion matrix overall folds of training the CNN for volcano image classification.
Figure 15. The confusion matrix overall folds of training the CNN for volcano image classification.
Remotesensing 13 04747 g015
Figure 16. Examples of Klyuchevskoy volcano images, classified by convolutional neural network: (a) uninformative image of class ‘clear’; (b) visible volcano with no significant activity—class ‘clear’; (c) image with volcano activity detected—class ‘active’.
Figure 16. Examples of Klyuchevskoy volcano images, classified by convolutional neural network: (a) uninformative image of class ‘clear’; (b) visible volcano with no significant activity—class ‘clear’; (c) image with volcano activity detected—class ‘active’.
Remotesensing 13 04747 g016
Table 1. The inventory and archive summary (as of the date of submission) of video observation system for Kamchatka volcanoes.
Table 1. The inventory and archive summary (as of the date of submission) of video observation system for Kamchatka volcanoes.
VolcanoVideocameraObservation PointObservation PeriodImages Count
1KlyuchevskoyStarDot NetCam XLKlyuchi2009–20201,927,537
2Axis P1343KlyuchiSince 20113,784,981
3Axis P1365KlyuchiSince 2020903,018
4SheveluchStarDot NetCam XLKlyuchiSince 20093,961,225
5Axis P1343KlyuchiSince 20114,441,012
6Axis P1343KlyuchiSince 20181,914,259
7AvachinskySanyo HD 4000Petropavlovsk-KamchatskySince 20113,812,263
8Axis P1365Petropavlovsk-KamchatskySince 2020877,305
9GorelyStartDot Netcam SCPetropavlovsk-KamchatskySince 2011449,654
Table 2. Main specification of the video camera, considered in the current study.
Table 2. Main specification of the video camera, considered in the current study.
CharacteristicValue
Camera modelStarDot NetCam XL 3MP
Sensor1/2″ CMOS sensor
Focal length8–48 mm
Max. Aperture Ratio1:1.2
Image resolution1024 × 768
FPS1/60
Distance to crater30 km
Table 3. Constant parameters for frequency characteristics computation, calculated for camera, observing Klyuchevskoy volcano.
Table 3. Constant parameters for frequency characteristics computation, calculated for camera, observing Klyuchevskoy volcano.
w 1 w 2 w 3 b ε b , w j
16.298.501.471.400.28
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Korolev, S.; Sorokin, A.; Urmanov, I.; Kamaev, A.; Girina, O. Classification of Video Observation Data for Volcanic Activity Monitoring Using Computer Vision and Modern Neural NetWorks (on Klyuchevskoy Volcano Example). Remote Sens. 2021, 13, 4747. https://doi.org/10.3390/rs13234747

AMA Style

Korolev S, Sorokin A, Urmanov I, Kamaev A, Girina O. Classification of Video Observation Data for Volcanic Activity Monitoring Using Computer Vision and Modern Neural NetWorks (on Klyuchevskoy Volcano Example). Remote Sensing. 2021; 13(23):4747. https://doi.org/10.3390/rs13234747

Chicago/Turabian Style

Korolev, Sergey, Aleksei Sorokin, Igor Urmanov, Aleksandr Kamaev, and Olga Girina. 2021. "Classification of Video Observation Data for Volcanic Activity Monitoring Using Computer Vision and Modern Neural NetWorks (on Klyuchevskoy Volcano Example)" Remote Sensing 13, no. 23: 4747. https://doi.org/10.3390/rs13234747

APA Style

Korolev, S., Sorokin, A., Urmanov, I., Kamaev, A., & Girina, O. (2021). Classification of Video Observation Data for Volcanic Activity Monitoring Using Computer Vision and Modern Neural NetWorks (on Klyuchevskoy Volcano Example). Remote Sensing, 13(23), 4747. https://doi.org/10.3390/rs13234747

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop