Next Article in Journal
A Noncommutative Model of Cosmology with Two Metrics
Previous Article in Journal
Symmetric Performance Analysis for Mechanical Properties of Sustainable Asphalt Materials under Varying Temperature Conditions: An Application of DT and NDT Digital Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Infrared Objects in Manifold Space Using Kullback-Leibler Divergence of Gaussian Distributions of Image Points

1
School of Electronics and Information, Jiangsu University of Science and Technology, Zhenjiang 212003, Jiangsu, China
2
College of Computer Science and Engineering, Xi’an University of Technology, Xi’an 710048, China
3
Faculty of Applied Mathematics, Silesian University of Technology, 44-100 Gliwice, Poland
4
Department of Applied Informatics, Vytautas Magnus University, Kaunas 44404, Lithuania
*
Authors to whom correspondence should be addressed.
Symmetry 2020, 12(3), 434; https://doi.org/10.3390/sym12030434
Submission received: 3 February 2020 / Revised: 1 March 2020 / Accepted: 5 March 2020 / Published: 8 March 2020

Abstract

:
Infrared image recognition technology can work day and night and has a long detection distance. However, the infrared objects have less prior information and external factors in the real-world environment easily interfere with them. Therefore, infrared object classification is a very challenging research area. Manifold learning can be used to improve the classification accuracy of infrared images in the manifold space. In this article, we propose a novel manifold learning algorithm for infrared object detection and classification. First, a manifold space is constructed with each pixel of the infrared object image as a dimension. Infrared images are represented as data points in this constructed manifold space. Next, we simulate the probability distribution information of infrared data points with the Gaussian distribution in the manifold space. Then, based on the Gaussian distribution information in the manifold space, the distribution characteristics of the data points of the infrared image in the low-dimensional space are derived. The proposed algorithm uses the Kullback-Leibler (KL) divergence to minimize the loss function between two symmetrical distributions, and finally completes the classification in the low-dimensional manifold space. The efficiency of the algorithm is validated on two public infrared image data sets. The experiments show that the proposed method has a 97.46% classification accuracy and competitive speed in regards to the analyzed data sets.

1. Introduction

Feature detection and matching are the basis of many image processing applications in the computer vision domain [1,2,3,4,5] and elsewhere [6]. Infrared small object detection is a focus of ongoing research in numerous areas, such as aircraft tracking [7], ship detection [8], 3D scene reconstruction [9] and video surveillance [10]. Infrared small object recognition in difficult environments such as those with a complex background and object clutter or those with low illumination is highly important, and is a difficult task in infrared search and tracking systems [11]. Contrary to visible light images, the infrared images do not have color information, while luminance is influenced by the thermal radiation of the object and background. Moreover, small infrared objects miss texture features due to the long sensing distance [12]. As a result, common object tracking methods based on visible features cannot recognize the difference between the object and its background in the infrared image [13].
Many different approaches have been proposed for infrared object tracking such as saliency extraction [9], multiscale patch-based contrast measure and a temporal variance filter [14], feature learning and fusion, reliability weight estimation based on nonnegative matrix factorization [15], Poisson reconstruction and the Dempster-Shafer theory [16], three-dimensional scalar field [17], a double-layer region proposal network (RPN) [18], Siamese convolution network [19], a mixture of Gaussians with modified flux density [20], spatial-temporal total variation regularization and weighted tensor [21], two-stage U-skip context aggregation network [22], histogram similarity map based on the Epanechnikov kernel function [23], quaternion discrete cosine transform [24], non-convex optimization [25], Mexican-hat distribution of pixels [26], and Schatten regularization with reweighted sparse enhancement [27].
Manifold learning assumes that the infrared object images to be classified are distributed as a set of points on the manifold space [28,29,30,31]. The purpose of manifold learning is to put forward a representation method for mapping manifold data point sets to a low-dimensional space [32]. The prior knowledge of the low-dimensional manifold of an image can be effectively used for the image reconstruction method, as has been demonstrated for the computer tomography images in [33]. The current infrared image classification methods in manifold space mainly map high-dimensional infrared object data point sets to low-dimensional space, and then complete classification of infrared object data points. Knowing the intrinsic structure of data, efficient manifold-based image classification methods can be constructed [34].
The infrared object classification method on the manifold takes advantage of one property of the manifold space, so that the manifold space can be regarded as a small piece of Euclidean space locally [35]. It attempts to obtain the distribution information of the infrared object data point set in the entire manifold with all the low-dimensional local maps. Most of the infrared object classification methods on the manifold are developed under the concept of describing the relationship between points in the data point set in high-dimensional manifold space [36]. In order to better describe the local relationship of infrared object data points on high-dimensional manifolds, initial research focused on the topological relationships between local region data points. The local linear embedding (LLE) method was proposed to describe the local data points in the manifold space by measuring the Euclidean distance between data points [37]. LLE considers that each point can be represented by its surrounding points. The distance between the surrounding data points and the object point within the neighborhood is used as the weight. However, LLE cannot well reconstruct a high-dimensional manifold data point set with unevenly distributed data point sets. In order to further describe the distance relationship between the data points of the infrared object image on the manifold, the isometric feature mapping (ISOMAP) method [38] introduced the concept of geodesic distance. The idea of this method is to construct a feature neighborhood map of infrared object image data points, which can represent the local information of the infrared object image data point set in a high-dimensional manifold space. However, this method is more suitable for scenarios where the manifold space of infrared image data points is relatively flat, and the calculation cost is relatively high when calculating the optimal route in the neighborhood of the infrared object image data point set [39]. The Laplace feature mapping algorithm (LE) [40] is a different idea to using the Euclidean distance between points on the manifold. LE expresses the local relationship of the infrared object image data point set in the manifold space through the graph theory. However, LE does not perform well when the data points on the manifold are far away from each other.
Image classification using non-Euclidean manifolds such as a Grassmann manifold and the Symmetric Positive Definite (SPD) manifold [41] is becoming increasingly more attractive. The weights of the iterative manifold embedding (IME) layer are learned by unsupervised strategy, which has been used to analyze the intrinsic manifolds of data sets with missing data [42]. The distribution of image data in multi-view manifold space can be captured by Multi-view Generative Adversarial Network (GAN), which can map the shape and view manifolds in a lower dimensionality latent space [43]. The Wasserstein-driven low-dimensional manifold model (W-LDMM) can be used for noise estimation, image denoising and noisy image inpainting tasks [44]. On the other hand, Riemannian manifolds can be used to visualize geometric transformations in images [45], which has many useful applications for image augmentation aiming to improve the accuracy of classification for small data sets. Pixels corresponding to different image classes tend to be segmented better on the Riemannian manifold than in the spectral space, since image points mapped from the Gaussian probability distribution are radially distributed on the skewed surface of Riemannian manifold [46]. However, due to the specific characteristics of Riemannian manifolds, the traditional machine learning methods often fail on them [47], which motivates the exploration of new feature extraction and classification methods.
Currently, there are still many problems to be addressed in infrared object classification, such as the occlusion of objects, change of object position and change of light. [48]. Improvements in the accuracy of infrared object classification when the infrared object data set has only a few samples (the “small data” problem [49]) and lack of prior knowledge have become important research directions for manifold learning in infrared object classification.
The novelty and contribution of this paper is outlined as follows. We propose a novel manifold learning algorithm for infrared object detection and classification that uses Kullback-Leibler (KL) Divergence to minimize the loss function between two symmetrical distributions of points (the distribution in the manifold space and the distribution of the points of the infrared image), and finally completes the classification in the low-dimensional space.

2. Materials and Methods

To improve the classification accuracy of infrared objects, this manuscript proposes a manifold learning method for classification. The proposed algorithm (see Figure 1) constructs a high-dimensional manifold space by using the infrared object image pixels as the manifold space dimension. The constructed high-dimensional manifold is then mapped into a low-dimensional space. By describing the local probability distribution information of the infrared image data points, the reduced-dimensional data set can accurately retain the high-dimensional information. Finally, the difference between the probability distribution of the data point set in manifold space and the probability distribution of the low-dimensional space is minimized using the KL Divergence.

2.1. Construction of High-Dimensional Manifold Space

Each pixel of the infrared object image is used as each dimension in the manifold space to construct a high-dimensional manifold space. The infrared objects of various categories appear as data points in this manifold space. The infrared object image manifold constructed in this paper is a differential manifold, but its local space can be understood as Euclidean-style space. The definition of this high-dimensional infrared object manifold is as follows.
If every point p in X has an open neighborhood Y X , and in addition Y and an open subset in the Euclidean space R n are homeomorphic, then X is an n -dimensional topological manifold.
Let X be a d -dimensional manifold and V = { ( U α , φ α ) } α I is a set of X coordinate cards. V is a C k differential structure of X when the following conditions are met.
The manifold M formed by { U α : α I } is an open cover, and the mapping of φ α : U α to φ α ( U α ) R d is homeomorphic. If U α U β , and the double mapping φ α   φ β 1 : φ β ( U α U β ) to φ α ( U α U β ) and its inverse mapping are all k times differentiable, then ( U α , φ α ) is compatible with ( U β , φ β ) .

2.2. Low-Dimensional Mapping and Classification Method of Infrared Object Manifold Space

After the high-dimensional manifold space is constructed, the dimension of the manifold space is still high, and it needs to be reduced to complete the classification of infrared objects. In order to improve the efficiency, we have adopted a dimensionality reduction method for the manifold space. The Gaussian distribution is used to describe the distribution characteristics of the infrared object image data point set in the high-dimensional manifold space, and the Student’s t-distribution is used to describe the distribution characteristics of infrared object image data point set after dimensionality reduction in the low-dimensional space. Finally, the difference of two distributions is minimized to complete the infrared object classification process. The steps are mainly divided into the description of the distribution information of the infrared object image data points in two spaces and the process of obtaining the minimum value of the difference between two distributions.

2.2.1. Projection of Infrared Object Image Data Points

The infrared object image data point set is projected onto a plane by random projection. On this plane, all data point sets show a discrete distribution. The data set of infrared object image data in the n -dimensional manifold space X can be defined by Equations (1)–(3).
x i , x j X ,
x i = ( x i ( 1 ) , x i ( 2 ) , x i ( 3 ) ) T ,
x j = ( x j ( 1 ) , x j ( 2 ) , x j ( 3 ) ) T
Since the infrared object image data points are projected onto a two-dimensional plane, when calculating the distance between these points and the surrounding neighboring points, the distance L p between x i and x j can be defined by the Euclidean distance using the following formula:
L 2 ( x i , x j ) = ( l = 1 n | x i ( l ) x j ( l ) | 2 ) 1 2

2.2.2. Construction of the KNN Map of the Infrared Object Image Data Points

The k-Nearest Neighbor (KNN) map of the infrared object image data points can be constructed by the above steps, and the obtained KNN map initially describes the local features of the manifold data point set. The local feature information of the infrared object image data in the manifold space should also have the same local characteristics after being projected into the low-dimensional space. For infrared object image data points in the high-dimensional space, the KNN maps can be used to represent the weighted image maps of the surrounding data points to the object data points. Here a weight is the Euclidean distance between points. We assume that a normal distribution describes the distribution probability of these neighboring points. For example, there are two infrared object image data points x i and x j in a high-dimensional manifold space. The infrared object image data point x i is the center of the Gaussian distribution. x i uses probability P j | i to select x j as its nearest neighbor. P j | i of the neighboring points is inversely proportional to the distance of x j and can be expressed as:
P j | i = exp ( x i x j 2 / 2 σ i 2 ) k i exp ( x i x k 2 / 2 σ i 2 ) ,
where σ i is the variance of the Gaussian distribution with x i as the center point.
In order to avoid the congestion problem when the Gaussian distribution is projected into the two-dimensional space, Student’s t-distribution is used to describe the local relationship of the infrared object image data points in the low-dimensional space. The t-distribution in the two-dimensional space is derived from the Gaussian distribution of the infrared object image data point set in the manifold space. Assuming that the probability distribution of the neighborhood near the point set of the infrared object image data in the high-dimensional space can be represented by the normal distribution N ( μ , σ 2 ) , the mean value of the T distribution in the low-dimensional space can be derived by:
u ¯ = 1 n i = 1 n u i ,
where u i is the infrared object image data points in the KNN domain and u ¯ is its mean.
Similarly, the variance of these points can be derived from the following formula.
s 2 = 1 n 1 i = 1 n ( u i u ¯ ) 2 ,
The distribution of the infrared object image data points in the two-dimensional space is determined by the variance and standard deviation of the normal distribution in the manifold space and the number of data points. Therefore, the T distribution random variable can be constructed by:
t = u ¯ μ s n
In this way, the probability distribution of the infrared object image data points in the two-dimensional space can be constructed according to the T distribution:
q i | j = ( 1 + y i y j 2 ) 1 k l ( 1 + y k y l 2 ) 1 ,
where q i | j is the probability distribution of the infrared object image data points in the 2D space.
In order to maintain the symmetry of the two probability distributions, a uniform symmetrical distance function is introduced as shown in the following formula:
p i j = p j | i + P i | j 2 n

2.2.3. Dimensionality Reduction

Since the distribution of high-dimensional space P j | i and the distribution of low-dimensional space q i | j should be as similar as possible, the KL divergence is used as a loss function to minimize the difference of this two distributions. The process of dimension reduction and classification is converted to the process of obtaining the minimum value of the loss function:
C = i j p j | i l o g p j | i q i | j
In order to control overfitting that occurs due to the small number of samples, a degree of confusion is also set to avoid overfitting. The degree of confusion can be defined by
P e r p ( p i ) = 2 H ( P i ) ,
where H ( P i ) is the Shannon entropy with p i defined as follows:
H ( P i ) = j p j | i log 2 p j | i
The degree of confusion changes in proportion to the entropy. With this feature, the value of entropy can be changed by adjusting the degree of confusion.
The next step is to construct the objective function. The positive samples of the infrared object image data points projected into the 2D space should be clustered, and the negative samples of the infrared object image data points should be placed far away from the positive samples. The weight between the points can be defined by Equation (14),
P ( e i j = 1 ) = f ( y i y j 2 ) ,
where y i and y j represent two points in the low-dimensional space. A binary edge e i j between these two points has a weight value of 1, and P ( e i j = 1 ) represents the probability that these two points exist. When the distance between y i and y j becomes closer, the value P ( e i j = 1 ) becomes larger. The expression with weight w i j is used in practical applications as follows:
P ( e i j = w i j ) = P ( e i j = 1 ) w i j
The positive sample set in the infrared object image data point set is defined as E , and the negative sample set is defined as E ¯ . These sets are obtained from the KNN diagram as follows.
O = ( i , j ) E p ( e i j = 1 ) w i j ( i , j E ¯ ) ( 1 p ( e i j = 1 ) ) γ ,
where γ is the weight of the negative samples.
The optimization process can be understood as maximizing the probability of weighted edges of positive samples in the KNN graphs and minimizing the probability of weighted edges of negative samples in the KNN graphs. For the convenience of calculation, the above optimization formula can be transformed into the following formula:
O = ( i , j ) E w i j P ( e i j = 1 ) ( i , j ) E ¯ γ ( 1 p ( e i j = 1 ) )
The negative sample E ¯ increases the computational complexity, and it is not easy to directly use gradient descent for training. Therefore, a negative sampling algorithm is selected in this paper, and a negative sample is formed based on a randomly selected number of infrared object image data points that conform to the noise distribution P n ( j ) . The objective function is expressed by Equation (18).
O = ( i , j ) E w i j ( p ( e i j = 1 ) + k = 1 M E j k ~ p n ( j ) γ log ( 1 p ( e i j = 1 ) ) )

2.2.4. Classification in Infrared Object Manifold Space

Finally, the low-dimensional representation of the manifold space and the classification results of different categories of objects are obtained. First, we find the distance between the input infrared object image and each point set in a two-dimensional space. Then, a set of points with the smallest distance from the input object image is considered as the category to which the object image belongs.
The problems of gradient vanishing and gradient explosion still occur during training. This problem can be solved by converting the edge between two points into a binary edge with the number of w i j . When more edges with large weights are converted into binary edges, the calculation cost will become higher. Therefore, we use random sampling in these transformed binary edges to solve the problem of high computational cost. In the optimization loss function, we use the asynchronous stochastic gradient descent (ASGD) algorithm [50] to improve the execution performance.

2.2.5. Illustration of the Method Stages

To illustrate the different stages of the method operation, we used the thermal infrared image “ambassador_morning” from the CSIR-CSIO Moving Object Thermal Infrared Imagery Dataset (MOTIID) [51]. Figure 2 shows the probability distributions calculated for different classes of image data points and the constructed KNN tree.
Figure 3 illustrates the low-dimensional embedding of the “ambassador_morning” image into a two-dimensional manifold and the corresponding tracking result. Note that the sets of closely related image points form clusters in the manifold space, which correspond to the particular areas of the target infrared image such as the tracked object.

3. Experimental Verification

The hardware platform used in this experiment is a Microsoft Surface Pro laptop with Intel Core M3-7Y30 1.61GHz CPU and 4 GB RAM.
The experiments in this paper used the following infrared object image data sets.
  • Two infrared video sequences named “8_quadrocopter1” and “8_horse” in the open-source LTIR data set (v1.0) of the Computer Vision Laboratory of Linköping University [52].
  • The infrared video sequence named “data1” in the data set for dim-small object detection and tracking of aircraft in infrared image sequences of the ATR Key Laboratory of National University of Defense Technology [7].
  • An infrared video sequence named “6a” in the OSU Color-Thermal Database data set [53].
  • The four infrared image sequences of the CSIR-CSIO Moving Object Thermal Infrared Imagery Dataset (MOTIID) data set named “ambassador_morning”, “auto_partially_occluded”, “bike_far” and “dog_evening” [51].
The characteristics of image sequences are summarized in Table 1.
Eight kinds of objects are selected in the eight infrared object image data sets. There are 20 samples of each object. The size of these infrared object images is resized to 40 × 40 px, and the corresponding category label data is also added. The data set information is shown in Table 2, while the examples of images are shown in Figure 4.
After reducing the dimensionality of the manifold space of the infrared object images, different categories of infrared object image data points are represented in the form of point sets in the low-dimensional space. The set of points with different label numbers in Figure 5 represents different categories of infrared object images. We can see that different categories of infrared objects have been effectively classified, and there are obvious gaps between different categories. There were no misclassifications observed, partly because of the small number of samples, which were separated in the manifold space well.
The reason for the clear classification results is that the characteristics of different infrared object types are quite different from each other. The number of infrared object categories we used is small, so the results of infrared object classification do not overlap with each other. The background changes of categories 1, 2, 4, 5, 7, 8 are not obvious, and the differences between objects and background are more obvious. These six categories of infrared object images have similar image distribution characteristics, so the infrared object point sets of these categories are closer in the figure. The point set of cars (category 3) and dog (category 6) is far from the point sets of other categories. This is because the pixel values of these two categories of infrared objects change greatly during the movement. Therefore, the position of their point set is farther away from several other categories of infrared objects.
We compare the proposed algorithm with three image classification algorithms: convolutional neural network (CNN) [14], multi-class Support Vector Machine (SVM) from LIBSVM [54], and multi-label lazy KNN (ML-KNN) [55] in terms of operation speed and classification accuracy. The comparison of the classification results of different algorithms used in this paper is shown in Figure 6.
As we can see from Figure 4, CNN (Convolutional Neural Network) is not as good as the proposed algorithm because of its complex structure and a small number of data set samples, which does not allow us to train the network effectively. Although SVM (Support Vector Machine) is faster in finding classification hyperplanes in the high-dimensional space, it cannot accurately divide the infrared object image data points. As a result, the accuracy of SVM is lower than the accuracy of the proposed algorithm. The ML-KNN (Multi-label Lazy KNN) algorithm is more accurate, but it takes more time to calculate the result when compared to the algorithm proposed in this paper.
The results are summarized in Table 3 using typical classification assessment metrics [56]. Here FPR is False Positive Rate and AUC is Area under Receiver Operating Characteristic (ROC) Curve.

4. Discussion and Final Remarks

We have developed and implemented an infrared object classification method for infrared images with mainly static backgrounds, i.e., under the condition that there are few movements in the background. For this type of image, our method achieved a high accuracy of 97.46%, which exceeded the accuracy of other methods using state-of-the-art object classifiers such as CNN, SVM or ML-KNN. The algorithm proposed in this paper can establish a high-dimensional manifold space of infrared object images and can classify different categories of infrared objects. Particularly, the proposed method can successfully perform infrared object classification even if only a small number of images are available for training. Most importantly, our method can effectively work with small data samples, on which the deep learning networks cannot be trained effectively. Finally, our experiments verify that the proposed algorithm can effectively classify different categories of infrared objects in the manifold space.
The achieved result demonstrates that our method could already be used for several applications such as infrared security cameras. Based on the main concepts used in the development of the method presented in this paper, we plan to work further on the development of new methods for infrared object tracking in images with dynamic backgrounds and cluttered object space, focusing on such applications as autonomous driving or military applications (such as those described in [57,58]) based on the concepts and ideas which were successfully validated in this paper.

Author Contributions

Conceptualization, K.L., W.W., and R.L.; methodology, K.L., W.W., and R.L.; software, H.G., Z.Z., and K.L.; validation, H.G., Z.Z., K.L., W.W., and R.L.; formal analysis, H.G., Z.Z., K.L., W.W., and M.W.; investigation, H.G., Z.Z., K.L., W.W., R.L., and R.D.; resources, H.G., Z.Z., K.L., and R.L.; writing—original draft preparation, H.G., Z.Z., K.L., and W.W.; writing—review and editing, R.D., and M.W.; visualization, H.G., Z.Z., and K.L.; supervision, W.W., and R.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 61671222) and Development Program of Shaanxi Province (No. 2018ZDXM-GY-036) and Shaanxi Key Laboratory of Intelligent Processing for Big Energy Data (No. IPBED7).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dara, S.; Tumma, P. Feature Extraction by Using Deep Learning: A Survey. In Proceedings of the Second International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 29–31 March 2018; pp. 1795–1801. [Google Scholar] [CrossRef]
  2. Gabryel, M.; Damaševičius, R. The image classification with different types of image features. In Proceedings of the International Conference on Artificial Intelligence and Soft Computing ICAISC, Zakopane, Poland, 11–15 June 2017; pp. 497–506. [Google Scholar] [CrossRef]
  3. Zhou, B.; Duan, X.; Ye, D.; Wei, W.; Woźniak, M.; Damaševičius, R. Heterogeneous image matching via a novel feature describing model. Appl. Sci. 2019, 9, 4792. [Google Scholar] [CrossRef] [Green Version]
  4. Zhou, B.; Duan, X.; Ye, D.; Wei, W.; Woźniak, M.; Połap, D.; Damaševičius, R. Multi-level features extraction for discontinuous object tracking in remote sensing image monitoring. Sensors 2019, 19, 4855. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Zhou, B.; Duan, X.; Wei, W.; Ye, D.; Wozniak, M.; Damasevicius, R. An adaptive local descriptor embedding zernike moments for image matching. IEEE Access 2019, 7, 183971–183984. [Google Scholar] [CrossRef]
  6. Riaz, F.; Azad, M.A.; Arshad, J.; Imran, M.; Hassan, A.; Rehman, S. Pervasive blood pressure monitoring using Photoplethysmogram (PPG) sensor. Future Gener. Comput. Syst. 2019, 98, 120–130. [Google Scholar] [CrossRef] [Green Version]
  7. Hui, B.; Song, Z.; Fan, H.; Zhong, P.; Hu, W.; Zhang, X.; Ling, J.; Su, H.; Jin, W.; Zhang, Y.; et al. A dataset for dim-small object detection and tracking of aircraft in infrared image sequences. China Sci. Data 2019, 1–12. [Google Scholar] [CrossRef]
  8. Li, Y.; Li, Z.; Zhu, Y.; Li, B.; Xiong, W.; Huang, Y. Thermal infrared small ship detection in sea clutter based on morphological reconstruction and multi-feature analysis. Appl. Sci. 2019, 9, 3786. [Google Scholar] [CrossRef] [Green Version]
  9. Ma, Y.; Wang, Y.; Mei, X.; Liu, C.; Dai, X.; Fan, F.; Huang, J. Visible/Infrared combined 3D reconstruction scheme based on nonrigid registration of multi-modality images with mixed features. IEEE Access 2019, 7, 19199–19211. [Google Scholar] [CrossRef]
  10. Younsi, M.; Diaf, M.; Siarry, P. Automatic multiple moving humans detection and tracking in image sequences taken from a stationary thermal infrared camera. Expert Syst. Appl. 2020, 146, 113171. [Google Scholar] [CrossRef]
  11. Chen, Y.; Song, B.; Du, X.; Guizani, M. Infrared small object detection through multiple feature analysis based on visual saliency. IEEE Access 2019, 7, 38996–39004. [Google Scholar] [CrossRef]
  12. Zhang, K.; Yang, K.; Li, S.; Chen, H. A difference-based local contrast method for infrared small object detection under complex background. IEEE Access 2019, 7, 105503–105513. [Google Scholar] [CrossRef]
  13. Li, L.; Zhou, F.; Zheng, Y.; Bai, X. Reconstructed saliency for infrared pedestrian images. IEEE Access 2019, 7, 42652–42663. [Google Scholar] [CrossRef]
  14. Gao, J.; Lin, Z.; An, W. Infrared small object detection using a temporal variance and spatial patch contrast filter. IEEE Access 2019, 7, 32217–32226. [Google Scholar] [CrossRef]
  15. Lan, X.; Ye, M.; Shao, R.; Zhong, B.; Jain, D.K.; Zhou, H. Online non-negative multi-modality feature template learning for RGB-assisted infrared tracking. IEEE Access 2019, 7, 67761–67771. [Google Scholar] [CrossRef]
  16. Li, J.; Huo, H.; Sui, C.; Jiang, C.; Li, C. Poisson reconstruction-based fusion of infrared and visible images via saliency detection. IEEE Access 2019, 7, 20676–20688. [Google Scholar] [CrossRef]
  17. Ma, M. Infrared pedestrian detection algorithm based on multimedia image recombination and matrix restoration. Multimed. Tools Appl. 2019, 1–16. [Google Scholar] [CrossRef]
  18. Qu, H.; Zhang, L.; Wu, X.; He, X.; Hu, X.; Wen, X. Multiscale object detection in infrared streetscape images based on deep learning and instance level data augmentation. Appl. Sci. 2019, 9, 565. [Google Scholar] [CrossRef] [Green Version]
  19. Shen, G.; Zhu, L.; Lou, J.; Shen, S.; Liu, Z.; Tang, L. Infrared multi-pedestrian tracking in vertical view via siamese convolution network. IEEE Access 2019, 7, 42718–42725. [Google Scholar] [CrossRef]
  20. Sun, Y.; Yang, J.; Li, M.; An, W. Infrared small-faint object detection using non-i.i.d. mixture of gaussians and flux density. Remote Sens. 2019, 11, 2831. [Google Scholar] [CrossRef] [Green Version]
  21. Sun, Y.; Yang, J.; Long, Y.; An, W. Infrared small object detection via spatial-temporal total variation regularization and weighted tensor nuclear norm. IEEE Access 2019, 7, 56667–56682. [Google Scholar] [CrossRef]
  22. Wang, H.; Shi, M.; Li, H. Infrared dim and small object detection based on two-stage U-skip context aggregation network with a missed-detection-and-false-alarm combination loss. Multimed. Tools Appl. 2019, 1–22. [Google Scholar] [CrossRef]
  23. Yun, S.; Kim, S. TIR-MS: Thermal infrared mean-shift for robust pedestrian head tracking in dynamic object and background variations. Appl. Sci. 2019, 9, 3015. [Google Scholar] [CrossRef] [Green Version]
  24. Zhang, P.; Wang, X.; Wang, X.; Fei, C.; Guo, Z. Infrared small object detection based on spatial-temporal enhancement using quaternion discrete cosine transform. IEEE Access 2019, 7, 54712–54723. [Google Scholar] [CrossRef]
  25. Zhang, T.; Wu, H.; Liu, Y.; Peng, L.; Yang, C.; Peng, Z. Infrared small object detection based on non-convex optimization with lp-norm constraint. Remote Sens. 2019, 11, 559. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, Y.; Zheng, L.; Zhang, Y. Small infrared object detection via a mexican-hat distribution. Appl. Sci. 2019, 9, 5570. [Google Scholar] [CrossRef] [Green Version]
  27. Zhou, F.; Wu, Y.; Dai, Y.; Wang, P. Detection of small object using Schatten 1/2 quasi-norm regularization with reweighted sparse enhancement in complex infrared scenes. Remote Sens. 2019, 11, 2058. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, K.; Li, X. Infrared small dim object detection based on region proposal. Optik 2019, 182, 961–973. [Google Scholar] [CrossRef]
  29. Deng, L.; Zhang, J.; Zhu, H. Infrared moving point object detection using a spatial-temporal filter. Infrared Phys. Technol. 2018, 95, 122–127. [Google Scholar] [CrossRef]
  30. Nie, J.; Qu, S.; Wei, Y.; Zhang, L.; Deng, L. An infrared small object detection method based on multiscale local homogeneity measure. Infrared Phys. Technol. 2018, 90, 186–194. [Google Scholar] [CrossRef]
  31. Ge, H.; Zhu, Z.; Lou, K. Tracking video target via particle filtering on manifold. Inf. Technol. Control. 2019, 48, 538–544. [Google Scholar] [CrossRef]
  32. Zhu, J.Y.; Krähenbühl, P.; Shechtman, E.; Efros, A.A. Generative Visual Manipulation on the Natural Image Manifold. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 597–613. [Google Scholar]
  33. Cong, W.; Wang, G.; Yang, Q.; Li, J.; Hsieh, J.; Lai, R. CT image reconstruction on a low dimensional manifold. Inverse Probl. Imag. 2019, 13, 449–460. [Google Scholar] [CrossRef] [Green Version]
  34. Luo, F.; Huang, Y.; Tu, W.; Liu, J. Local manifold sparse model for image classification. Neurocomputing 2019, 382, 162–173. [Google Scholar] [CrossRef]
  35. Bernstein, A.; Kuleshov, A.; Yanovich, Y. Manifold Learning in Regression Tasks. In Proceedings of the International Symposium on Statistical Learning and Data Sciences, Egham, UK, 20–23 April 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 414–423. [Google Scholar]
  36. Bai, S.; Bai, X.; Tian, Q. Scalable Person Re-Identification on Supervised Smoothed Manifold. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2530–2539. [Google Scholar]
  37. Zhu, B.; Liu, J.Z.; Cauley, S.F.; Rosen, B.R.; Rosen, M.S. Image reconstruction by domain-transform manifold learning. Nature 2018, 555, 487. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Tenenbaum, J.B.; de Silva, V.; Langford, J.C. A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef]
  39. Calandra, R.; Peters, J.; Rasmussen, C.E.; Deisenroth, M.P. Manifold Gaussian Processes for Regression. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, Canada, 24–29 July 2016; pp. 3338–3345. [Google Scholar]
  40. Lu, J.; Wang, G.; Deng, W.; Moulin, P.; Zhou, J. Multi-Manifold Deep Metric Learning for Image Set Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1137–1145. [Google Scholar]
  41. Wei, D.; Shen, X.; Sun, Q.; Gao, X.; Yan, W. Prototype learning and collaborative representation using Grassmann manifolds for image set classification. Pattern Recognit. 2020, 100, 107123. [Google Scholar] [CrossRef]
  42. Xu, J.; Wang, C.; Qi, C.; Shi, C.; Xiao, B. Iterative manifold embedding layer learned by incomplete data for large-scale image retrieval. IEEE Trans. Multimed. 2019, 21, 1551–1562. [Google Scholar] [CrossRef]
  43. Cui, J.; Li, S.; Xia, Q.; Hao, A.; Qin, H. Learning multi-view manifold for single image based modeling. Comput. Gr. 2019, 82, 275–285. [Google Scholar] [CrossRef]
  44. He, R.; Feng, X.; Wang, W.; Zhu, X.; Yang, C. W-LDMM: A wasserstein driven low-dimensional manifold model for noisy image restoration. Neurocomputing 2020, 371, 108–123. [Google Scholar] [CrossRef]
  45. Liu, T.; Shi, Z.; Liu, Y. Visualization of the image geometric transformation group based on riemannian manifold. IEEE Access 2019, 7, 105531–105545. [Google Scholar] [CrossRef]
  46. Zhao, X.; Li, Y.; Wang, H. Manifold based on neighbour mapping and its projection for remote sensing image segmentation. Int. J. Remote Sens. 2019, 40, 9304–9320. [Google Scholar] [CrossRef]
  47. Liu, X.; Ma, Z.; Niu, G. Mixed region covariance discriminative learning for image classification on riemannian manifolds. Math. Prob. Eng. 2019, 2019, 1261398. [Google Scholar] [CrossRef] [Green Version]
  48. Lu, J.; Tan, Y.P.; Wang, G. Discriminative multimanifold analysis for face recognition from a single training sample per person. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 39–51. [Google Scholar] [CrossRef] [PubMed]
  49. Qi, G.-J.; Luo, J. Small Data Challenges in Big Data Era: A Survey of Recent Progress on Unsupervised and Semi-Supervised Methods. arXiv 2019, arXiv:1903.11260. [Google Scholar]
  50. Dean, J.; Corrado, G.; Monga, R.; Chen, K.; Devin, M.; Le, Q.V.; Mao, M.Z.; Ranzato, M.A.; Senior, A.W.; Tucker, P.A.; et al. Large Scale Distributed Deep Networks. In Proceedings of the Neural Information Processing Systems NIPS, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1232–1240. [Google Scholar]
  51. Akula, A.; Ghosh, R.; Kumar, S.; Sardana, H.K. Moving object detection in thermal infrared imagery using spatiotemporal information. JOSA A 2013, 30, 1492–1501. [Google Scholar] [CrossRef]
  52. Berg, A.; Ahlberg, J.; Felsberg, M. A Thermal Object Tracking Benchmark. In Proceedings of the 12th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Karlsruhe, Germany, 25–28 August 2015; pp. 1–6. [Google Scholar]
  53. Davis, J.; Sharma, V. Background-Subtraction using Contour-based Fusion of Thermal and Visible Imagery. Comput. Vision Image Underst. 2007, 106, 162–182. [Google Scholar] [CrossRef]
  54. Chang, C.; Lin, C. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  55. Zhang, M.-L.; Zhou, Z.-H. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef] [Green Version]
  56. Tharwat, A. Classification Assessment Methods. Available online: https://www.sciencedirect.com/science/article/pii/S2210832718301546 (accessed on 2 February 2020).
  57. d’Acremont, A.; Fablet, R.; Baussard, A.; Quin, G. CNN-Based Target Recognition and Identification for Infrared Imaging in Defense Systems. Sensors 2019, 19, 2040. [Google Scholar] [CrossRef] [Green Version]
  58. Ivanovas, A.; Ostreika, A.; Maskeliūnas, R.; Damaševičius, R.; Połap, D.; Woźniak, M. Block Matching Based Obstacle Avoidance for Unmanned Aerial Vehicle. In Proceedings of the Artificial Intelligence and Soft Computing, ICAISC, Zakopane, Poland, 3–7 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; Volume 10841, pp. 58–69. [Google Scholar] [CrossRef]
Figure 1. Schematic outline of the algorithm proposed in this article.
Figure 1. Schematic outline of the algorithm proposed in this article.
Symmetry 12 00434 g001
Figure 2. Illustration of the construction of the k-Nearest Neighbor (KNN) tree (right) using probability distributions of image points (left) for the “ambassador_morning” image. The red color shows a tree node corresponding to the tracked infrared object.
Figure 2. Illustration of the construction of the k-Nearest Neighbor (KNN) tree (right) using probability distributions of image points (left) for the “ambassador_morning” image. The red color shows a tree node corresponding to the tracked infrared object.
Symmetry 12 00434 g002
Figure 3. Illustration of the embedding of the image points into a two-dimensional manifold (left) and the corresponding tracked infrared object (right) for the “ambassador_morning” image. The red color outlines the cluster of a closely related set of image points in the two-dimensional manifold and the tracked object.
Figure 3. Illustration of the embedding of the image points into a two-dimensional manifold (left) and the corresponding tracked infrared object (right) for the “ambassador_morning” image. The red color outlines the cluster of a closely related set of image points in the two-dimensional manifold and the tracked object.
Symmetry 12 00434 g003
Figure 4. Sample images of different infrared object categories.
Figure 4. Sample images of different infrared object categories.
Symmetry 12 00434 g004aSymmetry 12 00434 g004b
Figure 5. Classification results of different infrared object categories.
Figure 5. Classification results of different infrared object categories.
Symmetry 12 00434 g005
Figure 6. Comparison of calculation time and accuracy of different algorithms.
Figure 6. Comparison of calculation time and accuracy of different algorithms.
Symmetry 12 00434 g006
Table 1. Characteristics of image sequences.
Table 1. Characteristics of image sequences.
Image SequenceNo. of Images (Frames)Resolution, pxBit Depth
8_quadrocopter1178640 × 4808
8_horse348324 × 2568/16
data1398256 × 25624
6a1652320 × 2408
ambassador_morning155640 × 48024
auto_partially_occluded219640 × 48024
bike_far202640 × 48024
dog_evening69640 × 48024
Table 2. Description of different infrared object categories.
Table 2. Description of different infrared object categories.
CategoryDistanceBackgroundLabel
HorseFarGround1
PlaneFarSky2
CarCloseGround3
QuadcopterFarWall4
TricycleCloseGround5
GogCloseGround6
MotorcycleFarGround7
PedestrianCloseGround8
Table 3. Summary of performance characteristics of infrared object tracking methods. CNN: Convolutional Neural Network. SVM: Support Vector Machine. ML-KNN: Multi-label Lazy KNN.
Table 3. Summary of performance characteristics of infrared object tracking methods. CNN: Convolutional Neural Network. SVM: Support Vector Machine. ML-KNN: Multi-label Lazy KNN.
MetricCNNSVMML-KNNProposed
Accuracy94.8%92.4%96.2%97.46%
FPR0.0440.0930.0540.032
Precision0.9320.9110.9550.963
Recall0.9500.9310.9680.988
F-score0.9410.9210.9610.975
AUC0.9670.9580.9790.987

Share and Cite

MDPI and ACS Style

Ge, H.; Zhu, Z.; Lou, K.; Wei, W.; Liu, R.; Damaševičius, R.; Woźniak, M. Classification of Infrared Objects in Manifold Space Using Kullback-Leibler Divergence of Gaussian Distributions of Image Points. Symmetry 2020, 12, 434. https://doi.org/10.3390/sym12030434

AMA Style

Ge H, Zhu Z, Lou K, Wei W, Liu R, Damaševičius R, Woźniak M. Classification of Infrared Objects in Manifold Space Using Kullback-Leibler Divergence of Gaussian Distributions of Image Points. Symmetry. 2020; 12(3):434. https://doi.org/10.3390/sym12030434

Chicago/Turabian Style

Ge, Huilin, Zhiyu Zhu, Kang Lou, Wei Wei, Runbang Liu, Robertas Damaševičius, and Marcin Woźniak. 2020. "Classification of Infrared Objects in Manifold Space Using Kullback-Leibler Divergence of Gaussian Distributions of Image Points" Symmetry 12, no. 3: 434. https://doi.org/10.3390/sym12030434

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop