Next Article in Journal
Immersed Boundary Methods for Simulations of Biological Flows in Swimming and Flying Bio-Locomotion: A Review
Next Article in Special Issue
PEiD: Precise and Real-Time LOS/NLOS Path Identification Based on Peak Energy Index Distribution
Previous Article in Journal
Multiplicative Vector Fusion Model for Detecting Deepfake News in Social Media
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sample Reduction-Based Pairwise Linear Regression Classification for IoT Monitoring Systems

School of Information Science and Engineering, University of Jinan, Jinan 250022, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(7), 4209; https://doi.org/10.3390/app13074209
Submission received: 22 February 2023 / Revised: 12 March 2023 / Accepted: 24 March 2023 / Published: 26 March 2023

Abstract

:
At present, the development of the Internet of Things (IoT) has become a significant symbol of the information age. As an important research branch of it, IoT-based video monitoring systems have achieved rapid developments in recent years. However, the mode of front-end data collection, back-end data storage and analysis adopted by traditional monitoring systems cannot meet the requirements of real-time security. The currently widely used edge computing-based monitoring system can effectively solve the above problems, but it has high requirements for the intelligent algorithms that will be deployed at the edge end (front-end). To meet the requirements, that is, to obtain a lightweight, fast and accurate video face-recognition method, this paper proposes a novel, set-based, video face-recognition framework, called sample reduction-based pairwise linear regression classification (SRbPLRC), which contains divide SRbPLRC (DSRbPLRC), anchor point SRbPLRC (APSRbPLRC), and attention anchor point SRbPLRC (AAPSRbPLRC) methods. Extensive experiments on some popular video face-recognition databases demonstrate that the performance of proposed algorithms is better than that of several state-of-the-art classifiers. Therefore, our proposed methods can effectively meet the real-time and security requirements of IoT monitoring systems.

1. Introduction

At present, the development of the Internet of Things [1,2] has become a significant symbol of the information age, and the video monitoring systems are an important basic aspect within the IoT field, which can be used in many real word scenarios, such as intelligent elderly care monitoring systems, intelligent access control systems, and intelligent anti-theft systems.
Traditional monitoring systems usually consist of two parts, the front-end cameras, and the back-end server, and their basic processes are usually as follows: the front-end cameras are first used to collect the monitoring data, then the collected data will be uploaded to the back-end server for storage and processing; finally, the useful information is obtained through human inspection or some intelligent algorithms. However, such manner has the following disadvantages: (1) uploading the collected data to the server will cost a lot of time, resulting in a failure to meet real-time requirement; (2) compared with PCs and other devices, IoT devices are generally weak in performance and more vulnerable to attacks, resulting in their inability to ensure the security of data during transmission; (3) the collected video data usually need to be analyzed and recognized using intelligent algorithms. However, the existing video face-recognition methods still have some defects. For the above problems (1) and (2), the currently mature solution is to introduce edge computing [3] and migrate the data analysis process to the edge end. Specifically, some intelligent algorithms to each edge end (for example, deploy them to the surveillance cameras) are firstly deployed, in order to process the collected data in real time; secondly, the recognition or analysis results processed by the algorithm are returned to the back-end (server) for storage, or even alarming. In such a manner, we can not only effectively reduce the amount of data transmission and meet the real-time requirements, but also gain higher security, because the transmission data no longer comprise the original data but the analysis results. However, this scheme has high requirements for the intelligent algorithms to be deployed, that is, the intelligent algorithms need to have fewer parameters (because the memory of the edge device is limited), faster computing speed and higher recognition accuracy. In view of this practical demand, this paper will focus on the research of video face-recognition methods applicable to IoT monitoring systems.
So far, face recognition is already widely used in everyday life, including sign-in systems, fugitive tracking systems, etc. Many related algorithms have been developed [4,5], such as the principal component analysis (PCA) algorithm [6], linear regression classification (LRC) algorithm [7], support vector machine (SVM) algorithm [8], linear discriminant analysis (LDA) algorithm [9], k-nearest neighbor (K-NN) approach [10], canonical correlation analysis (CCA) method [11], and the sparse representation-based classifications (SRC) algorithm [12]. All of these single image-based image recognition algorithms have achieved satisfactory performances; however, with the development of imaging technology, a great number of videos is being produced everyday, and how to recognize video [13] data, i.e., how to measure the distance between videos, is still a challenging problem. Many researchers have pointed out that temporal information is not important for face recognition; hence, the term video face recognition in this study refers to set-based video face recognition, i.e., where temporal information is not considered.
When compared to a single image, videos and image sets contain useful information, such as varying expressions, varying poses, and varying illumination conditions, that describe the objects within in the images. Hence, it is very important to study the classification problem from the perspective of image sets, i.e., the image set classification task. The purpose of image set classification (or set-based video recognition) is to assign labels to the probe set (or video) by measuring the similarities between the gallery videos (or sets) and the probe videos (or sets). When compared with the single image-based recognition task, the set-based video recognition task can directly calculate the labels for whole videos, without classifying each frame of the videos separately, which can effectively accelerate the calculation speed. Since every video includes a wide range of appearance variations, the key to set-based video face recognition lies in two strategies: the representation of the videos and the accurate measurement of the distance between different videos. From the perspective of model representation and measurement learning, many set-based video face-recognition methods have been presented in recent decades, in order to identify the optimal recognition performance, including the discriminant canonical correlations (DCC) algorithm [14], covariance discriminative learning (CDL) algorithm [15], dual-linear regression classification (DLRC) algorithm [16], manifold discriminant analysis (MDA) algorithm [17], pairwise linear regression classification (PLRC) algorithm [18], image set-based collaborative representation and classification (ISCRC) algorithm [19], and set-level joint sparse representation classification (SJSRC) algorithm [20]. Among these algorithms, LRC-based methods have gradually attracted the attention of a greater number of researchers because of their superior performance.
Recently, the PLRC algorithm [18] has been proven to be a valid image set-classification method. By analyzing PLRC, the distance between images has been identified as the key point in the construction of unrelated sets. However, previously developed unrelated set construction strategies (i.e., S1 and S2 in [18]) are not optimal distance metric methods. Fortunately, based on collaborative and sparse representation, Gao et al. proposed another two unrelated set construction strategies: S3 and S4 [21]. In addition, when the size of the videos or image sets is small, PLRC can work well (see Figure 4). Therefore, how to make PLRC suitable for large-size videos has become an interesting problem. Gao et al. proposed the kernel PLRC (KPLRC) algorithm [21], with the aim of increasing the dimensionality of image samples in an image set, in order to increase the linear separability and overcome this issue. Unfortunately, KPLRC not only increases the classification accuracy, but also the additional computational overhead.
Motivated by this, and starting from the direction of decreasing the number of images in each video, this paper proposes a novel video face recognition or image set classification framework, named sample reduction-based PLRC (SRbPLRC), which consists of divide SRbPLRC (DSRbPLRC), anchor point SRbPLRC (APSRbPLRC), and attention anchor point SRbPLRC (AAPSRbPLRC) methods. Specifically, the DSRbPLRC algorithm was evaluated first, which simply divides large-size videos into several small-size videos/subsets randomly. However, DSRbPLRC increases the number of videos, which may reduce the computing speed, and it does not consider the influence of any noise or outliers that may exist in the videos. Then, we looked at the APSRbPLRC algorithm, which first uses the clustering method to divide each video into several sub-videos, and then the centroid and mean values of each sub-video were used as anchor points. APSRPLCR can greatly reduce the number of samples in videos and does not increase the number of videos. Thus, it can realize set-based video face recognition efficiently and accurately. Additionally, APSRPLCR can overcome the influence of noise and outliers to some extent. Nevertheless, APSRPLCR considers all images within each sub-video to be of equal importance while computing the anchor points; however, we believe that a good anchor point construction strategy should adaptively weigh and combine the images in each sub-video/subset. Thus, the AAPSRbPLRC algorithm was developed, which utilizes an attention mechanism to learn the optimal anchor points. An overall flow chart of this algorithm is shown in Figure 1. Finally, several experiments were conducted on some popular databases (i.e., Honda, Mobo, and YTC) to demonstrate the effectiveness of our proposed method for large-size video face recognition. The main contributions of this paper are given as follows:
  • A simple, yet effective, sample reduction-based pairwise linear regression classification framework is proposed, which consists of three methods, and these methods are augmented one by one.
  • By introducing the attention mechanism, our attention anchor point SRbPLRC method can adaptively weigh and combine the images in each cluster, so as to achieve faster and effective image set classification.
  • Experimental results can demonstrate the effectiveness of our proposed three methods for large size video face recognition. Therefore, our proposed methods can effectively meet the real-time and security requirements of IoT monitoring system.
The rest of this paper is organized as follows. In Section 2, we briefly review the existing set-based video face-recognition algorithms. In Section 3, we present the SRbPLRC framework in detail. The experiments and the results analysis are discussed in Section 4. In Section 5, we present our conclusions.

2. Related Works

In this section, we briefly review some related works, which cover Internet of Things and video face-recognition methods for IoT monitoring systems.

2.1. Internet of Things

The definition of IoT is as follows: “through various information sensors, collect any object or process that needs to be monitored, connected and interacted in real time, collect all kinds of required information such as sound, light, heat, electricity, mechanics, chemistry, biology and location, and realize the universal connection between things and people, and realize the intelligent perception, recognition and management of things and processes through various possible network access”. According to this definition, there are many research branches in IoT, such as IoT basic theory and technology, IoT data management and middleware technology, IoT electrical automation and remote monitoring, etc.
In the IoT basic theory and technology branch, researches mainly study the IoT communication protocol and node technology, sensor network architecture deployment and performance evaluation, information security and privacy, multi-sensor information fusion, etc. For example, regarding the IoT scenario, Cauteruccio et al. had developed many works on social networks [22,23,24]. For instance, in the literature [24], they proposed a new method to investigate anomalies in a multiple IoT scenario. In the literature [22], the authors proposed computing the scope of a social object in a multi-IoT scenario; and in another study [23], they proposed three new measures of betweenness centrality, specifically conceived for a multiple IoT scenario. Besides, Nicolazzo et al. [25] proposed a privacy-preserving approach to prevent feature disclosure in a multiple IoT scenario, i.e., a scenario where objects can be organized into (partially overlapped) networks that interact with each other.
In the IoT data management and middleware technology branch, researches mainly study massive data processing and analysis technology, as well a sdata storage, query, mining, analysis and fusion technology, in other words, applying artificial intelligence (AI) algorithms to IoT systems, and thus producing the AIoT concept. The AI technology is usually used to interpret and respond to some human-to-machine and machine-to-machine data flows in real time, and it can reduce latency, as well as increase the privacy and real-time intelligence of the system at the edge. This also means that fewer data need to be sent and stored on the server. Note that the research in this paper just belongs to this research field.

2.2. Video Face-Recognition Methods for IoT Monitoring Systems

Video- or image set-based face-recognition has been actively studied for decades. In this paper, we only considered inputs of orderless sets of face images. Since many researchers have pointed out that temporal information is not important for face-recognition tasks, existing methods that exploit temporal dynamics are not considered here.
Up until now, existing video recognition or image set classification approaches have been roughly grouped into three classes: model representation-based methods, metric learning-based methods, and simultaneous model representation- and metric learning-based methods.
Model representation-based methods mainly focus on how to build more discriminative model representations. For example, the discriminant canonical correlations (DCC) algorithm [14] used subspaces to model videos or image sets, and employed the canonical correlation metric to measure the similarities between two videos. The log-Euclidean metric learning (LEML) algorithm [26] and covariance discriminative learning (CDL) algorithm [15] both used covariance matrices to model videos and logarithm kernels, in order to learn the distance between videos or image sets. Single Gaussian or Gaussian mixture models (GMM) used probability distributions to model videos, and utilized existing K-L divergence to measure the similarities between the distributions. The disadvantage of model representation-based methods is that they ignore the importance of distance metric learning.
Different from model representation-based approaches, metric learning-based methods mainly use existing strategies to represent videos, and focus on how to learn accurate distance metrics. Cevikalp and Triggs [27] modeled image sets and videos as affine hulls and convex hulls, and then proposed two metric learning methods: the convex hull-based image set distance (CHISD) algorithm and the affine hull-based image set distance (AHISD) algorithm. They then developed two optimization functions to learn distance metrics, based on collaborative representation. The SANP [28] method also used affine hulls to represent videos, and employs a new distance learning method, which tries to identify the SNAP between two videos. The ISCRC algorithm [19] represents probe videos as regularized/convex hulls and uses all gallery videos to collaboratively reconstruct these hulls in order to learn the discriminative distances. Dual-linear regression classification (DLRC) [16] took LRC as its baseline and extended it to video recognition. Specifically, DLRC used spanning subspaces to represent image sets and videos, and tried to learn more suitable set-based distances. Inspired by DLRC, the concept of unrelated sets was first defined in PLRC, which learned related distance metrics, unrelated distance metrics, and combination metrics to improve classification performance. Unfortunately, experiments have shown that even though DLRC and PLRC can achieve satisfactory performances in small-size video recognition or image set classification tasks, they do not perform well on large-size videos. Additionally, the existing unrelated set construction strategies (i.e., S1 and S2 in PLRC) are not optimal. So, how to construct more discriminative unrelated sets and extend the methods to large-size video recognition tasks remain challenging problems. Gao et al. proposed the KPLRC algorithm, which uses two new, unrelated set-constructing strategies, and tries to increase the dimensionality of the image samples in each image set to overcome the limitations of PLRC. Unfortunately, KPLRC not only increases the classification accuracy but also the additional computational overhead. Inspired by this, a novel SRbPLRC framework is proposed in this paper, which combines the two new unrelated set construction strategies from KPLRC with three extensions of SRbPLRC, for large-size video recognition or image set classification.
Simultaneous model representation- and metric learning-based methods consider model representation and metric learning at the same time, in order to achieve more accurate video recognition. For instance, the joint metric learning-based class-specific representation (JMLC) [29] framework modeled image sets as affine hulls, and simultaneously learned the related and unrelated distance metrics. The projection metric learning (PML) algorithm [30] used subspaces to represent videos, and learned the projection metric directly using Grassmann manifolds, in order to reduce the computation costs. The MDA algorithm [17] and manifold–manifold distance (MMD) algorithm [31] both modeled videos as local linear models. Thus, the similarities between two videos are then characterized by the distance between their corresponding local linear models. The covariate-relation graph (CRG) algorithm [32] used graphs to represent videos, and tried to learn the graph guide distance metric. The multi-model fusion metric learning (MMFML) algorithm [33] was developed because it was thought that using one model to represent each image set was not enough; therefore, this algorithm used multi-models to jointly model each video or image set and adopted a distance fusion method. Other methods in the literature [34,35,36] also used subspace points, which lie on Grassmannian manifolds, to represent videos, and different discriminant analysis approaches have been developed to learn more effective distance metrics.

3. SRbPLRC Framework for Large-Size Video Face Recognition

In one study [18], the PLRC algorithm has been developed and a definition of unrelated sets has been introduced, with the aim of improving discriminative information. According to that definition of unrelated sets, N i samples need to be selected from all classes (except the i t h class) to construct an unrelated set. Additionally, two unrelated set construction strategies have been developed in PLRC, which are denoted as S1 and S2. In S1, response vectors are reconstructed using a complete dataset X = [ X 1 , X 2 , , X C ] R q × L and a probe set Y R q × n , where L is the sum of all N i and N i is the number of samples in X i . Obviously, q , L , n satisfy the inequality q < L + n , which indicates that S1 does not satisfy the basic conditions of the DLRC algorithm. Hence, the learned reconstruction coefficient of S1 is unbelievable (see Figure 4). Inspired by S1, two novel unrelated set construction strategies have further been proposed in the literature [21]: S3 and S4. The details of these strategies are shown in Figure 2. However, although KPLRC overcomes the disadvantages of PLRC with the help of S3 and S4, its computing time is very high. Aiming at obtaining an effective and efficient video face recognition method, we developed the new SRbPLRC framework by decreasing the number of images in each video. Note that our proposed framework will borrow the above mentioned unrelated set-construction strategies.

3.1. DSRbPLRC Algorithm

Suppose that there are C large-size videos/image sets, denoted as { X 1 , , X C } , with each video containing N i images and each image being x m i R q × 1 . Additionally, suppose that the image vector x m i is normalized to [0, 1]. Using these definitions, the i t h gallery video or image set can be formulated as X i = { x j i | j = 1 , , N i } and the probe video or image set can be represented as Y = { y j | j = 1 , , n } , where n is the number of images in the probe video Y.
The DSRbPLRC algorithm aims to divide large-size videos into several small-size sub-videos to improve its ability to deal with the large-size video recognition task. Assuming that t is the division threshold, i.e., the number of samples in the divided video is no larger than t, the related (or unrelated) gallery video X i can be randomly divided into X i 1 , X i 2 , , X i a , where a = [ N i / t + 1 ] is the number of the divided sub-videos, and
X i j = { x 1 i j , , x t i j } , j = 1 , , a 1 , X i a = { x 1 i a , , x N i t ( a 1 ) i a } ,
where i means the sub-video X i j or X i a comes from the original i t h gallery video, and j means that it is the j t h sub-video of X i .
Similarly, the probe video Y can be divided into Y 1 , Y 2 , , Y b , where b = [ n / t + 1 ] is the number of the divided sub-videos of Y, and
Y e = { y 1 e , , y t e } , e = 1 , , b 1 , Y b = { y 1 b , , y n t ( b 1 ) b } ,
where e means the e t h sub-video of Y.
The DSRbPLRC algorithm uses spanning subspaces to represent each sub-video; thus, the gallery set X i j can be represented as X i j α = [ x 1 i j , , x t i j ] α R q × 1 , i = 1 , , C , j = 1 , , a , which can represent all images in X i j . The probe set Y e can be represented as Y e η = [ y 1 e , , y t e ] η , e = 1 , , b . Finally, the unrelated set U i j that corresponds to the gallery set X i j can be represented as U i j γ = [ u 1 i j , u 2 i j , , u N i i j ] γ R q × 1 .
By minimizing the Euclidean distance between Y e η and X i j α (or U i j γ ), we obtain the following optimization problem:
min | | X i j α Y e η | | 2 2 s . t . i α i = 1 , i η i = 1 ,
and
min | | U i j γ Y e η | | 2 2 s . t . i γ i = 1 , i η i = 1 .
Then, using the least squares algorithm, the related and unrelated representation coefficients β e , r i j and β e , u i j between ( X i j , Y e ) and ( U i j , Y e ) , respectively, can be computed as follows:
β e , r i j = ( S e , r i j T S e , r i j ) 1 S e , r i j T s e , r i j , β e , u i j = ( S e , u i j T S e , u i j ) 1 S e , u i j T s e , u i j ,
where s e , r i j = y m e a n e x m e a n i j , s e , u i j = y m e a n e u m e a n i j , S e , r i j = [ X ¯ i j , Y ¯ e ] , S e , u i j = [ U ¯ i j , Y ¯ e ] , X ¯ i j = [ x 1 i j x m e a n i j , , x N i i j x m e a n i j ] , β e , r i j = [ α ; η ] , β e , u i j = [ γ ; η ] , and the definitions of Y ¯ e , U ¯ i j are the same as in X ¯ i j .
Thus, the related metric distance d e , r i j and the unrelated metric distance d e , u i j between X i j and Y e can be calculated as follows:
d e , r i j = | | s e , r i j S e , r i j β e , r i j | | 2 , d e , u i j = | | s e , u i j S e , u i j β e , u i j | | 2 .
Finally, the combination metric d e i j = d e , r i j / d e , u i j is used to obtain the label of sub-video Y e (the detail can be found in [29]). Thus, the label of probe video Y can be obtained by using the majority voting strategies.

3.2. APSRbPLRC Algorithm

From the above subsection, we can see that the DSRbPLRC algorithm increases the number of videos, which may reduce the computing speed, and it does not consider the influence of any noise or outliers that may exist in the videos. Thus, the APSRbPLRC algorithm is introduced in this subsection, which first uses the clustering method to divide each video into several sub-videos, and then uses the centroid and mean values of each sub-video as anchor points.
Anchor points are the most representative points in videos or image sets, and the subsets formed by anchor points can approximately replace the original videos or sets. In order to decrease the size of image sets or videos and improve efficiency, subsets of anchor points are constructed by extracting the common characters from each image and singling out the most representative points. An illustration of this process is shown in Figure 3.
In this study, the hierarchical divisive clustering (HDC) algorithm [31] is used as the clustering method to extract the clusters, i.e., to divide each video into several sub-videos. Specifically, the HDC algorithm regards each image set or video as a manifold and measures the non-linearity degrees of different clusters according to the deviations between the Euclidean distances and geodesic distances. This ensures that there is a similar number of images in the extracted clusters, and that the samples in each cluster have similar visual characteristics. Then, the mean image of each cluster is computed as the anchor point, which can in turn represent the whole subset.
After obtaining the anchor points, the size of the videos is greatly reduced, and due to the necessary information being preserved by the anchor points, recognition rates can also be retained. We should point out that other anchor points extraction strategies can also be used, such as the K-means clustering approach, spectral clustering algorithms, and one-class SVMs.
In this study, for C gallery videos { X 1 , , X C } and a probe video Y, the HDC algorithm is first used to extract the anchor points of each video, so we could obtain the following anchor point gallery and probe videos: { X 1 a p , , X C a p } and Y a p (where X i a p contains N i a p representative images, Y a p contains n a p representative images, and N i a p N i , n a p n is satisfied). As with the DSRbPLRC algorithm, we could then directly obtain the label of Y using Equations (5) and (6).

3.3. AAPSRbPLRC Algorithm

The APSRbPLRC algorithm uses the mean image of each subset as the anchor point; however, we believe that a good anchor point construction strategy should adaptively weigh and combine the images in each subset. Hence, we adopted an attention mechanism to obtain more discriminative anchor points.
Suppose that we obtained subset { x 1 , , x m } using the HDC algorithm, and then initialized the query vector q 0 R q × 1 . We can then obtain the score of each image x i by computing the dot product as follows:
e k = ( q 0 ) T x k ,
where T is the transpose operation. The attention distribution α k can then also be computed, as follows:
α k = exp ( e k ) j exp ( e j ) .
Finally, the weighted average of subset can be obtained as follows:
r 0 = k = 1 m α k x k .
The obtained r 0 can be input into the following equation to obtain the new query vector q 1 , q 2 :
q 1 = T a n h ( W 1 r 0 + b 1 ) , q 2 = T a n h ( W 2 q 1 + b 2 ) .
Finally, the anchor point r 2 of the subset generated by q 2 can be obtained using Equations (7)–(9).
In order to solve the parameters in Equation (10), the subspace-based contrastive loss is used in this paper. Suppose that { X 1 a p , , X C a p } are anchor point gallery videos computed after the above attention mechanism, then the distance d i j between X i a p and X j a p can be computed by Equations (5) and (6). Thus, the subspace based contrastive loss can be defined as:
L = i j y i j d i j + ( 1 y i j ) max ( 0 , m d i j )
where y i j equals 1 or 0 ( y i j = 1 means that X i a p and X j a p belong to the same class, whereas y i j = 0 means that X i a p and X j a p belong to different classes).
After obtaining more discriminative anchor points for each video, the AAPSRbPLRC method can effectively produce the labels for the probe video.
In summary, when we suppose that each sub-video or subset has a similar number of images, then the PLRC algorithm needs to compute C large-size distances to classify one probe video, while the DSRbPLRC algorithm needs to compute C × a × b small-size distances, and the APSRbPLRC and AAPSRbPLRC algorithms both need to compute C small-size distances.

4. Experiments

4.1. Experimental Settings

Next, we verified the effectiveness of the proposed framework for large-size video recognition tasks. In this study, the division threshold t in DSRbPLRC was set as t = 100 . There are two parameters for the APSRbPLRC algorithm: the threshold parameter θ and the nearest neighbor parameter k. According to [31], it is known that larger θ values imply fewer local models (thus, higher efficiency) but larger linearity deviations, and vice versa. Thus, the parameter θ was selected by searching from 1 , 2 , , 5 , while parameter k was selected by searching from 1 , 2 , , 12 . In the AAPSRbPLRC algorithm, the parameter m was set as 1.5.

4.2. Methodology

In this subsection, we describe the comparison experiments that we conducted, in order to compare our proposed method with some state-of-the-art set-based video recognition methods, including DCC [14], AHISD, CHISD [27], SANP [28], MMD [31], COV [15], ISCRC [19], LEML [26], PML [30], the covariate-relation graph (CRG) algorithm [32], DLRC [16], and PLRC (PLRC-I and PLRC-II) [18]. The implementation codes of these algorithms were provided by the original authors, and their parameters were empirically tuned, according to the recommendations in the original references. The attention mechanism in our AAPSRbPLRC algorithm was trained using a standard backpropagation algorithm. On the Honda and Mobo databases, we randomly divided the gallery videos into two small videos, as each class only has one gallery video.

4.3. Databases

In this study, three popular large-size databases were used in our experiments: the Honda database, Mobo database, and YouTube Celebrities (YTC) database. The Honda/UCSD database contains 59 video sequences from 20 different people and there are about 300∼500 frames in each sequence. For fair comparison, 20 sequences were randomly selected for training (one sequence was selected from each class), and the other 39 video sequences were selected for testing. So, the video frames from each video sequence became a face image set. The Mobo database contains 96 video sequences from 24 different people, with each person having four video sequences in each class. In our experiments, we randomly selected one sequence from each person for training, and used the remaining three video sequences for testing. The YTC database is a large-size video recognition database, which consists of 1910 video clips from 47 celebrities on the YouTube website. In each clip, there are hundreds of frames. In our experiments, we randomly selected three video clips or image sets from each celebrity for training and six video clips or image sets for testing. All of the face images in these databases were resized to 20 × 20 .

4.4. SRbPLRC for Large-Size Video Recognition

On the Honda database, the biggest video sequence contains 618 images. On the Mobo database, the biggest video sequence includes 340 images, while on the YouTube Celebrities database, the biggest video sequence includes 349 frames. Hence, we will perform large-size video recognition experiments on these databases.
Figure 4 illustrates the recognition rates of 10 random experiments using the DLRC and PLRC methods. The videos were from the Honda and Mobo databases and contained varying numbers of frames. In the experiments, PLRC-III and PLRC-IV are the combinations of PLRC and the S3 and S4 unrelated set construction strategies. From the figure, it can clearly be observed that the DLRC, PLRC-I, and PLRC-II methods only work well on small videos (when there are fewer than 150 frames in each video). Specifically, on the Honda database, the performance of the DLRC, PLRC-I, and PLRC-II methods is significantly reduced when the number of frames is larger than 150. For the Mobo database, the demarcation point for the DLRC, PLRC-I, and PLRC-II methods is also 150 frames. We also noticed that the PLRC-III and PLRC-IV methods perform better than the DLRC, PLRC-I, and PLRC-II methods.
In subsequent experiments, the Honda, Mobo and YouTube Celebrities databases, with all their frames, were used as the large-sized videos. The average classification accuracies of the DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms are shown in Figure 5.
Figure 5 lists a quantitative evaluation (i.e., the average classification accuracy) of the DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms for large-size videos. It can be observed that our proposed method outperforms PLRC in all cases, which indicates that our SRbPLRC framework is effective, i.e., the DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms can overcome the limitations of PLRC to some extent. We also found that the DSRbPLRC and APSRbPLRC algorithms achieve similar performances, while the AAPSRbPLRC algorithm achieves a better performance, which indicates that the attention mechanism could find more discriminative anchor points. From Figure 5, it can also be seen that the unrelated S3 and S4 set -construction strategies outperform the other unrelated subspace-construction strategies. These quantitative and qualitative comparisons confirm the applicability of the proposed method for large-size video recognition.
Finally, Table 1 enumerates the performances of the different methods on the different databases when using all frames. The results of DSRbPLRC, APSRbPLRC and AAPSRbPLRC are the best classification accuracies selected from Figure 4 and Figure 5. As can be seen from the table, we observed the following: (1) the proposed method performs better than PLRC, which indicates that our proposed approach is effective for the large-sized video face-recognition task; (2) the AAPSRbPLRC algorithm obtains the best classification accuracy on all databases, while the APSRbPLRC algorithm achieves a sub-optimal classification performance on the Mobo database, which means that decreasing the number of images in the videos used can overcome the shortcomings of the DLRC and PLRC algorithms, and that the attention mechanism is helpful for the large-sized video face-recognition task.

4.5. Computation Times

In this subsection, we present the testing times for identifying a query video on the Honda and Mobo databases, using different methods. Here, only the unrelated S1 and S2 set-construction strategies were used for all methods. The average computing times are shown in Table 2. From the table, we can see the following: (a) compared to PLRC, our APSRbPLRC and AAPSRbPLRC algorithms require less time for testing, possibly because our anchor point method can effectively decrease the set size; (b) compared to the other methods, the DSRbPLRC algorithm requires the most testing time, which is due to the fact that the DSRbPLRC algorithm needs to solve more optimization problems.

5. Conclusions

In this study, a new set-based video face recognition framework for IoT monitoring systems, called sample reduction-based pairwise linear regression classification, was proposed, which can effectively classify large-size videos. Subsequently, three classifiers (DSRbPLRC, APSRbPLRC, and AAPSRbPLRC) were also developed by decreasing the number of images in each video. Our experimental results on some popular databases demonstrate the superiority of our DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms. Specifically, qualitative and quantitative comparisons between our proposed method, DLRC, and PLRC for the large-size video face-recognition task demonstrate that the DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms can overcome the limitations of PLRC. We also find that the AAPSRbPLRC method achieves the best classification results on all databases.
The main limitations of the proposed framework include the fact that we did not consider the importance of feature learning; only one model feature is used in our framework, while multi-model features are very common in IoT monitoring systems. Therefore, the unrelated set construction strategy can be further improved. Hence, our future studies will combine our framework with deep neural networks, classify multi-feature videos, and design new unrelated set-construction strategies. Besides this, we will also research the accelerating deploy problem in IoT systems using these algorithms.

Author Contributions

Conceptualization, X.G. and W.H.; Data curation, W.H.; Formal analysis, W.H.; Funding acquisition, X.G.; Investigation, X.G.; Methodology, X.G. and S.N.; Project administration, X.G.; Resources, S.N.; Software, W.H.; Supervision, S.N.; Validation, Y.C.; Visualization, Y.C.; Writing—original draft, X.G.; Writing—review and editing, W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China (grant number: 62101213), the Shandong Provincial Natural Science Foundation (grant number: ZR2020QF107), the Development Program Project of Youth Innovation Team of Institutions of Higher Learning in Shandong Province, and the Big Data Project of University of Jinan (grant number: XKY1926).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pedditi, R.B.; Debasis, K. Energy Efficient Routing Protocol for an IoT-Based WSN System to Detect Forest Fires. Appl. Sci. 2023, 13, 3026. [Google Scholar] [CrossRef]
  2. Azfar, S.; Nadeem, A.; Ahsan, K.; Mehmood, A.; Siddiqui, M.S.; Saeed, M.; Ashraf, M. An IoT-Based System for Efficient Detection of Cotton Pest. Appl. Sci. 2023, 13, 2921. [Google Scholar] [CrossRef]
  3. Chang, D.-M.; Hsu, T.-C.; Yang, C.-T.; Yang, J. A Data Factor Study for Machine Learning on Heterogenous Edge Computing. Appl. Sci. 2023, 13, 3405. [Google Scholar] [CrossRef]
  4. Ullah, W.; Hussain, T.; Baik, S.W. Vision transformer attention with multi-reservoir echo state network for anomaly recognition. Inf. Process. Manag. 2023, 60, 103289. [Google Scholar] [CrossRef]
  5. Ullah, W.; Hussain, T.; Khan, Z.A.; Haroon, U.; Baik, S.W. Intelligent dual stream CNN and echo state network for anomaly detection. Knowl.-Based Syst. 2022, 253, 109456. [Google Scholar] [CrossRef]
  6. Turk, M.; Pentland, A. Eigenfaces for recognition. J. Cogn. Neurosci. 1991, 3, 71–86. [Google Scholar] [CrossRef] [PubMed]
  7. Naseem, I.; Togneri, R.; Bennamoun, M. Linear regression for face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 31, 2106–2112. [Google Scholar] [CrossRef]
  8. Gao, X.; Fan, L.; Xu, H. A novel method for classification of matrix data using twin multiple rank smms. Appl. Soft Comput. 2016, 48, 546–562. [Google Scholar] [CrossRef]
  9. Belhumeur, P.N.; Hespanha, J.P.; Kriegman, D.J. Eigenfaces vs. fisher-faces: Recognition using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 711–720. [Google Scholar] [CrossRef] [Green Version]
  10. Hastie, T.; Tibshirant, R. Discriminant adaptive nearest neighbor classification. IEEE Trans. Pattern Anal. Mach. Intell. 1996, 18, 607–616. [Google Scholar] [CrossRef] [Green Version]
  11. Gao, X.; Sun, Q.; Yang, J. MRCCA: A novel CCA based method and its application in feature extraction and fusion for matrix data. Appl. Soft Comput. 2018, 62, 45–56. [Google Scholar] [CrossRef]
  12. Wright, J.; Yang, A.Y.; Ganesh, A.; Sastry, S.S.; Ma, Y. Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 210–227. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. ur Rehman, A.; Belhaouari, S.B.; Kabir, M.A.; Khan, A. On the Use of Deep Learning for Video Classification. Appl. Sci. 2023, 13, 2007. [Google Scholar] [CrossRef]
  14. Kim, T.K.; Kittler, J.; Cipolla, R. Discriminative learning and recognition of image set classes using canonical correlations. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1005–1018. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, R.; Guo, H.; Davis, L.S.; Dai, Q. Covariance discriminative learning: A natural and efficient approach to image set classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2496–2503. [Google Scholar]
  16. Chen, L. Dual linear regression based classification for face cluster recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 2673–2680. [Google Scholar]
  17. Wang, R.; Chen, X. Manifold discriminant analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 429–436. [Google Scholar]
  18. Feng, Q.; Zhou, Y.; Lan, R. Pairwise linear regression classification for image set retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4865–4872. [Google Scholar]
  19. Zhu, P.; Zuo, W.; Zhang, L.; Shiu, S.C.; Zhang, D. Image set-based collaborative representation for face recognition. IEEE Trans. Inf. Forensics Secur. 2014, 9, 1120–1132. [Google Scholar] [CrossRef] [Green Version]
  20. Zheng, P.; Zhao, Z.Q.; Gao, J.; Wu, X. A set-level joint sparse representation for image set classification. Inf. Sci. 2018, 448, 75–90. [Google Scholar] [CrossRef]
  21. Gao, X.; Sun, Q.; Xu, H.; Gao, J. Sparse and collaborative representation based kernel pairwise linear regression for image set classification. Expert Syst. Appl. 2020, 140, 112886. [Google Scholar] [CrossRef]
  22. Cauteruccio, F.; Cinelli, L.; Fortino, G.; Savaglio, C.; Terracina, G.; Ursino, D.; Virgili, L. An approach to compute the scope of a social object in a Multi-IoT scenario. Pervasive Mob. Comput. 2020, 67, 101223. [Google Scholar] [CrossRef]
  23. Cauteruccio, F.; Terracina, G.; Ursino, D.; Virgili, L. Redefining Betweenness Centrality in a Multiple IoT Scenario. In Proceedings of the 1st Workshop on Artificial Intelligence and Internet of Things Co-Located with the 18th International Conference of the Italian Association for Artificial Intelligence, Rende, Italy, 19–20 November 2019; pp. 16–27. [Google Scholar]
  24. Cauteruccio, F.; Cinelli, L.; Corradini, E.; Terracina, G.; Ursino, D.; Virgili, L.; Savaglio, C.; Liotta, A.; Fortino, G. A framework for anomaly detection and classification in Multiple IoT scenarios. Future Gener. Comput. Syst. 2021, 114, 322–335. [Google Scholar] [CrossRef]
  25. Nicolazzo, S.; Nocera, A.; Ursino, D.; Virgili, L. A privacy-preserving approach to prevent feature disclosure in an IoT scenario. Future Gener. Comput. Syst. 2020, 105, 502–519. [Google Scholar] [CrossRef]
  26. Huang, Z.; Wang, R.; Shan, S.; Li, X.; Chen, X. Log-euclidean metric learning on symmetric positive definite manifold with application to image set classification. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 720–729. [Google Scholar]
  27. Cevikalp, C.; Triggs, B. Face recognition based on image sets. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2567–2573. [Google Scholar]
  28. Zhao, Z.Q.; Xu, S.T.; Liu, D.; Tian, W.D.; Jiang, Z.D. A review of image set classification. Neurocomputing 2019, 335, 251–260. [Google Scholar] [CrossRef]
  29. Gao, X.; Niu, S.; Wei, D.; Liu, X.; Wang, T.; Zhu, F.; Dong, J.; Sun, Q. Joint Metric Learning-Based Class-Specific Representation for Image Set Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022; Early Access. [Google Scholar] [CrossRef]
  30. Huang, Z.; Wang, R.; Shan, S.; Chen, X. Projection metric learning on grassmann manifold with application to video based face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 140–149. [Google Scholar]
  31. Wang, R.; Shan, S.; Chen, X.; Gao, W. Manifold-manifold distance with application to face recognition based on image set. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, Alaska, 23–28 June 2008; pp. 1–8. [Google Scholar]
  32. Chen, Z.; Jiang, B.; Tang, J.; Luo, B. Image set representation and classification with attributed covariate-relation graph model and graph sparse representation classification. Neurocomputing 2017, 226, 262–268. [Google Scholar] [CrossRef]
  33. Gao, X.; Sun, Q.; Xu, H.; Wei, D.; Gao, J. Multi-model fusion metric learning for image set classification. Knowl.-Based Syst. 2019, 164, 253–264. [Google Scholar] [CrossRef]
  34. Wei, D.; Shen, X.; Sun, Q.; Gao, X. Discrete Metric Learning for Fast Image Set Classification. IEEE Trans. Image Process. 2022, 31, 6471–6486. [Google Scholar] [CrossRef] [PubMed]
  35. Wei, D.; Shen, X.; Sun, Q.; Gao, X.; Ren, Z. Sparse Representation Classifier Guided Grassmann Reconstruction Metric Learning with Applications to Image Set Analysis. IEEE Trans. Multimed. 2022; Early Access. 2022. [Google Scholar] [CrossRef]
  36. Wang, T.; Shi, P. Kernel Grassmannian distances and discriminant analysis for face recognition from image sets. Pattern Recognit. Lett. 2009, 30, 1161–1165. [Google Scholar] [CrossRef]
Figure 1. An illustration of the proposed framework. Aiming at obtaining an effective and efficient video face recognition method, a new SRbPLRC framework is proposed, by decreasing the number of images in each video. This framework contains DSRbPLRC, APSRbPLRC, and AAPSRbPLRC methods.
Figure 1. An illustration of the proposed framework. Aiming at obtaining an effective and efficient video face recognition method, a new SRbPLRC framework is proposed, by decreasing the number of images in each video. This framework contains DSRbPLRC, APSRbPLRC, and AAPSRbPLRC methods.
Applsci 13 04209 g001
Figure 2. A diagram of the unrelated S3 and S4 set-constructing strategies. In S3, the sparse representation is used to learn the distances, while in S4, the collaborative representation is used.
Figure 2. A diagram of the unrelated S3 and S4 set-constructing strategies. In S3, the sparse representation is used to learn the distances, while in S4, the collaborative representation is used.
Applsci 13 04209 g002
Figure 3. An illustration of the proposed DSRbPLRC and APSRbPLRC algorithms. DSRbPLRC directly divides a large-size video into several small-size sub-videos, and each sub-video is used for classification. Different from DSRbPLRC, APSRbPLRC uses the HDC algorithm to extract anchor points, which will not increase the number of videos.
Figure 3. An illustration of the proposed DSRbPLRC and APSRbPLRC algorithms. DSRbPLRC directly divides a large-size video into several small-size sub-videos, and each sub-video is used for classification. Different from DSRbPLRC, APSRbPLRC uses the HDC algorithm to extract anchor points, which will not increase the number of videos.
Applsci 13 04209 g003
Figure 4. The experimental results of DLRC and PLRC methods from two popular databases, with varying numbers of frames.
Figure 4. The experimental results of DLRC and PLRC methods from two popular databases, with varying numbers of frames.
Applsci 13 04209 g004
Figure 5. The average recognition rates of the DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms, in terms of large-sized videos.
Figure 5. The average recognition rates of the DSRbPLRC, APSRbPLRC, and AAPSRbPLRC algorithms, in terms of large-sized videos.
Applsci 13 04209 g005
Table 1. The average classification accuracies of the different methods on the three popular databases (%), where the bold denotes the best results.
Table 1. The average classification accuracies of the different methods on the three popular databases (%), where the bold denotes the best results.
MethodHondaMoboYTC
DCC96.6788.5461.73
MMD97.1894.1067.23
LEML97.1888.6950.60
PML96.6789.6966.13
CRG88.4670.4849.38
COV100.090.3569.18
SANP93.6093.8766.70
AHISD96.3493.1365.13
CHISD95.4494.2163.40
DLRC30.5477.3037.31
PLRC-I65.1384.6749.89
PLRC-II67.9586.0349.43
DSRbPLRC96.5995.9667.18
APSRbPLRC96.3696.8368.11
AAPSRbPLRC100.097.9872.23
Table 2. The testing times for identifying a query video, on the Honda and Mobo databases.
Table 2. The testing times for identifying a query video, on the Honda and Mobo databases.
MethodHondaMobo
PLRC-I2.033.38
PLRC-II6.339.38
DSRbPLRC-I20.3830.89
DSRbPLRC-II80.80100.8
APSRbPLRC-I1.282.89
APSRbPLRC-II4.988.83
AAPSRbPLRC-I1.793.21
AAPSRbPLRC-II5.678.87
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, X.; Hu, W.; Chu, Y.; Niu, S. Sample Reduction-Based Pairwise Linear Regression Classification for IoT Monitoring Systems. Appl. Sci. 2023, 13, 4209. https://doi.org/10.3390/app13074209

AMA Style

Gao X, Hu W, Chu Y, Niu S. Sample Reduction-Based Pairwise Linear Regression Classification for IoT Monitoring Systems. Applied Sciences. 2023; 13(7):4209. https://doi.org/10.3390/app13074209

Chicago/Turabian Style

Gao, Xizhan, Wei Hu, Yu Chu, and Sijie Niu. 2023. "Sample Reduction-Based Pairwise Linear Regression Classification for IoT Monitoring Systems" Applied Sciences 13, no. 7: 4209. https://doi.org/10.3390/app13074209

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop