Next Article in Journal
A High Resolution Capacitive Sensing System for the Measurement of Water Content in Crude Oil
Previous Article in Journal
A Sensitive Sensor Cell Line for the Detection of Oxidative Stress Responses in Cultured Human Keratinocytes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Compact Methodology to Understand, Evaluate, and Predict the Performance of Automatic Target Recognition

1
School of Electrical Science and Engineering, National University of Defense Technology, 137 Yanwachi Street, Changsha 410073, China
2
State Key Laboratory of Complex Electromagnetic Environment Effects on Electronics and Information System (CEMEE), Luoyang 471003, China
*
Author to whom correspondence should be addressed.
Sensors 2014, 14(7), 11308-11350; https://doi.org/10.3390/s140711308
Submission received: 5 December 2013 / Revised: 23 May 2014 / Accepted: 9 June 2014 / Published: 25 June 2014
(This article belongs to the Section Physical Sensors)

Abstract

: This paper offers a compacted mechanism to carry out the performance evaluation work for an automatic target recognition (ATR) system: (a) a standard description of the ATR system's output is suggested, a quantity to indicate the operating condition is presented based on the principle of feature extraction in pattern recognition, and a series of indexes to assess the output in different aspects are developed with the application of statistics; (b) performance of the ATR system is interpreted by a quality factor based on knowledge of engineering mathematics; (c) through a novel utility called “context-probability” estimation proposed based on probability, performance prediction for an ATR system is realized. The simulation result shows that the performance of an ATR system can be accounted for and forecasted by the above-mentioned measures. Compared to existing technologies, the novel method can offer more objective performance conclusions for an ATR system. These conclusions may be helpful in knowing the practical capability of the tested ATR system. At the same time, the generalization performance of the proposed method is good.

Graphical Abstract

1. Introduction

1.1. The ATR Technology and Performance Analysis for It

Automatic target recognition (ATR) is the capability for an algorithm or equipment to recognize targets or objects based on the data obtained from sensors [1,2]. ATR is an essential component of intelligence systems implemented with various types of sensors [1,3]. Therefore, it is of great importance to have an objective and quantitative performance evaluation measure for an ATR system [1].

ATR technology is widely employed as the essential technique in advanced systems such as within the military [4,5], security [6,7] and modern medical science [8]. It enables a radar to catch its object of interest [9,10], helps a seeker find the fixed target in a complicated scenario [3,11], and makes the accurate diagnosis possible with medical sensors [12,13]. Nowadays, ATR is frequently typified by the application of radars and optical-sensors [4].

The primary principle in ATR is inverse theory, which, enlightened by the feeding signal collected with certain types of sensors, makes decisions on the information related to the intended target [3,9] (see “ATR system” in figure of Section 2.1). For example, people may know that ATR can be viewed as an inverse problem in the fields of electromagnets and acoustics: targets of interest are sensed, the sensed signatures are then transmitted to the detectors, and the main purpose of ATR is to use these signatures to classify the original targets [1,5,14].

As for the components of an ATR system, the employed sensor can be a polarimetric infrared, a hyperspectral device, or an ultra-wide band radar [3]. Many kinds of classifiers are investigated, such as model-based classifier, statistical based classifier, phenomenological modeling classifier, context information based classifier and information fusion classifier [9]. With the rapid advances in sensor technology, flexible field programmable gate array (FPGA), high performance computer, the art of ATR is becoming more pertinent to a much wider group of scientist and engineers than before [1,14]. As we proceed into the future, there will be more and more research/applications of ATR technologies and ATR systems [2].

However, given the changing environment and the limitation of the sensors, this system sometimes runs into trouble [15]. For example, the same kind of cell and the diseased tissue being observed may vary in shape, size and even quality [16,17], the vehicles being investigated may shift in velocity, pitch and direction [6], a certain type of sensor can only collect limited information from the target [3], and this point is further complicated by the fact that so many systems and factors are involved in the signal processing course of an ATR system [11].

Given the facts mentioned previously, the performance evaluation for ATR systems (PE-ATR) and the performance prediction for ATR systems (PP-ATR) continue to be studied by many experts in the field [18,19]. As aforementioned, the application of ATR in radars and photo-sensors is frequently found. Consequently, the literature on performance evaluation in those ATR systems maintains the major part in this area [20,21]. There are evaluation technologies for ATR in radar systems [22,23], performance assessment work on ATR algorithms employing motion imagery [24], performance prediction testbed for an image-based ATR algorithm [20], etc. When reviewing the available technologies, most of the literatures are from the radar and photo-sensor related discipline. Typical performance evaluation and performance prediction technologies for an ATR system are as follows.

1.2. Scaling the Operating Condition for the ATR System

ATR systems operate and are tested under certain conditions. These conditions may be regarded as subsets of a multi-dimensional space of conditions [25]. To obtain an objective conclusion, the operating condition should be considered in the performance evaluation course [9]. However, the scenarios of the ATR system are many and varied [26]. It is of great importance to develop some approaches to scale the operating condition (throughout this paper, unless otherwise stated, “operating condition” is all the scenarios an ATR system can be applied) for an ATR system [27]. Unfortunately, the literature on scaling the operating condition for an ATR system is limited. Available works are as follows:

  • Operating condition description with concepts. Generally, there are four sets of conditions: operating (here, “operating condition” is only the operational condition), testing, modeled, and training [25]. The relation of them can be shown in a Venn diagram [25]. Although this approach is not a quantified way, it is helpful in discussing the performance of an ATR system.

  • For image metric-based ATR and synthetic aperture radar (SAR) ATR, the operating condition is sometimes quantified in the way of image characterization [28]. As a fundamental idea, the concept of “Extended Operating Condition (EOC)” is defined [27,29]. EOC is an operating condition “away” from the trained condition [27]. Experiments shown the tested SAR ATR performance (recognition rate) was very sensitive to the EOCs tested [27,30].

  • With respect to the condition in signal processing in ATR, amplitude affection factor (AAF) and signal to noise ratio affect factor (SNRF) are developed [9]. In view of the condition of feature extraction, extend of recognition (ER) is proposed [9]. These metrics are further applied in denning performance evaluation indexes and building performance evaluation models [9].

1.3. Performance Evaluation for ATR Systems

As mentioned above, the performance evaluation for ATR systems is studied by many experts. The available technologies can be divided into two main groups according to the framework: model-based and data-based approaches [9]. When estimating the performance for an ATR system, model-based approaches usually work with a performance model such as the expected measures of effectiveness (MOE), robust evaluation model and independence evaluation model [9,19]. The data-based approaches directly calculate some indexes from the recognition output such as recognition rate and false alarm [21]. In practice, these two approaches are often combined in performance evaluating for an ATR system.

1.3.1. Basic Performance Evaluation Indexes

In PE-ATR, some basic measures like probability of detection (PD), recognition rate (RR), and false alarm rate (PFA) are generally employed facilities [9]. Estimating the performance bound is concerned in the early years [31].

Performance concepts are also introduced to compare performance across ATR technologies [25]. Two classes of concepts are proposed [25]. One class is referred to as performance. It includes accuracy, extensibility, robustness, and utility [25]. These Performance concepts consider the relationship between the test data, the training data, and data from modeled conditions [25]. The other class is called cost. It includes efficiency, scalability, and synthetic trainability [25]. The latter class of concepts put the cost into three categories: data-collection, data storage, and data-processing [25].

The confusion matrix (CM) is another widely used performance evaluation approach [19,32]. It can be easily configured and employed for a diverse set of ATR systems. The matrix is a square grid with a single row and a single column corresponding to each category defined in the data set. The (i, j) cell in the matrix is the number of predicted classifications on category j that correspond to the truth source of category i [19].

It should be noted that, in some related works, the quantities to show the performance of an ATR system are all referred to as “performance metric” or “character of performance”. However, they are referred to as “performance indexes” in this work hereafter.

1.3.2. Performance Evaluation Based on Performance Modeling

In PE-ATR, many scientists work with performance models and/or evaluation models [33]. The existing performance models and evaluation models can be classified as: (a) models based on probability, statistics, and random processes [9,34]; (b) models based on Bayesian approach [35]; (c) models based on information theory approach [35]; (d) subsystem performance models [36]; (e) other performance models [37].

(a)

Models Based on Probability, Statistics, and Random Processes

Series of performance indexes are built based on probability, statistics, and random processes: measurement of recognition rate (MRR), measurement of false recognition rate (MFRR), mean of MRR, variance of MRR, the independence of MRR to operating condition, etc. [9].

(b)

Models Based on Bayesian Approach

As for the Bayesian approach, probability distributions are used to represent the variability in target and background signatures [35]. To apply the method, assumptions are usually made (such as the use of Gaussian distributions and independence of information sources) to ensure mathematical tractability. However, these assumptions are not always practical enough [35].

(c)

Models Based on Information Theory Approach

This kind of model casts the recognition problem as a communication process [35]. Information theory brings in the notation of entropy and measures of relative information to try to figure out how information and thus performance is lost along the processing course [35]. It may suffer from the problem when assumptions do not match reality closely enough. This kind of performance indexes have been applied in evaluating SAR ATR [38].

(d)

Subsystem Performance Models or Performance Model for Certain Metric

The computational burden is an important metric for image recognition [16]. It is further considered for image recognition of high resolution radar sensors [16].

For SAR ATR, polarization and resolution may affect the performance [36,39]. This can be studied with the help performance curves (probability of detection to false alarms) [36]. For some ATR algorithms, performance curves at all three ATR stages (detection, discrimination, and classification) for certain combination of polarization and resolution were studied by the Lincoln Laboratory [36,39].

Performance evaluation of the subsystem of an ATR system is meaningful. The reliability analysis of the sensor employed in ATR is of interest [40]. Performance indexes are built on two fundamental issues: reasonable dissimilarity among evidences, and adaptive combination of static and dynamic discounting [40]. These measures are helpful to optimize the mentioned ATR algorithm [40].

(e)

Other Performance Models

To study the potentialities of polarimetric SAR interferometry (POLInSAR) in developing a new classification methods for ships, performance evaluation has been performed to accomplish a trade-off between geometry description accuracy and method robustness in reference feature vectors (or patterns) [37]. Experiments showed a low number of vectors could lead to an overestimation of the classification rate, and an excessive number of patterns would make quite similar geometries to be classified in different classes [37].

1.3.3. Receiver Operating Characteristic Analysis and Similar Approaches

Receiver operating characteristic (ROC) analysis is a broadly used performance analysis tool in signal processing and communications [34,41]. Researchers have introduced this notation into PE-ATR. A three-dimensional (3-D) ROC trajectory was presented to compare competing target recognition algorithms when unknown targets are present in the data [34]. In understanding the tradeoffs between the probability of rejection and other two performance measures commonly used in detection problems, it is a useful tool for SAR image analysis [34,42].

Scientists also extended the conventional ROC analysis from single-signal detection to detection and classification of multiple signals [41,43]. Applications showed it was a flexible utility in PE-ATR [41].

An extension of the ROC method is the analysis of performance bounds in different scenarios [15]. Some analytical characters on PE-ATR are obtained under complicated, non-Gaussian models and optimized system parameters [15]. For targets composed of a constellation of geometrically-simple reflectors, lower and upper bounds on the probability of correct classification are estimated in SAR ATR [44,45]. In performance evaluation for sidescan sonar target classification, some common bounds are derived to show the properties of ATR [46]. In pose estimation related to ATR, Hilbert-Schmidt lower bounds for estimators on matrix Lie groups is defined and validated [47].

Another extension of ROC method is confidence intervals for ATR performance evaluation index [48,49]. The provided confidence interval estimator includes proportion estimation based on Binomial distribution and rate estimation based on Poisson distribution. Under the Bayesian posterior distribution, this estimator is substantially more accurate than other similar approaches [48].

Automatic fingerprint recognition is a interdisciplinary field. It includes image processing, pattern recognition, computer technology, and so on. The confidence interval is compared between different automatic fingerprint recognition algorithms [50]. A performance model is built based on statistics. It can be applied to estimate the uniqueness of the template in classifiers [50].

1.3.4. Performance Evaluation Framework

Performance evaluation indexes assess the capability in various aspects. However, people sometimes seek an integrated conclusion in some different sides [21]. Therefore, performance evaluation framework is concerned and investigated. Generalized performance model is built based on fuzzy comprehensive evaluation, fuzzy integration and fuzzy cluster analysis [9]. These performance models can offer an algorithm-independent view of the ATR performance [21].

1.3.5. Other Performance Evaluation Methods

Underwater target recognition is challenging due to the presence of noise, point-spread function effects resulting from camera or media inhomogeneities [51]. Image compression transform is sometimes applied. Performance evaluation method of data compression transforms is then developed to achieve low-distortion images that eases the burden of classifiers [51].

For automatic face recognition systems, the effect of racial and gender demographics on estimating the accuracy of algorithms is considered [52]. It was reported that differences in the match threshold was required to obtain a false alarm rate of 0.001 when demographic controls on the non-matched identity (race or gender) pairs varied [52].

1.3.6. Performance Evaluation System or Performance Evaluation Testbed

As for PE-ATR software or a testbed, an example is given where Python (an open source scripting language) and OpenEV (a viewing and analysis tool) have been incorporated [53]. This testbed gives important insight into the risks as well as the successful use of open source language in ATR [53].

An experimental system called automated instrumentation and evaluation (Auto-I) is developed [32]. Auto-I has a module for automatic adaptation of algorithms parameters using algorithms performance models, optimization and artificial intelligence techniques [32]. The presented design of Auto-I is modular, it can be interfaced to other ATR systems except for the ATR system in [32].

For image-based target detection, a complete truthing system is developed [54]. It is named “the Scoring, Truthing, And Registration Toolkit (START) [54]”. This toolkit can align the images of the identical scene to a common reference frame. Then, “truthing” is applied to specify target identity, position, orientation, and other scene characteristics [54]. Finally, “scoring” is used to evaluate the performance of certain algorithms as compared to the specified truth [54].

1.4. Performance Prediction for ATR Systems

Compared to performance evaluation work, the existing performance prediction methods are fairly limited [20]. To some extend, the available performance prediction methods are extending work of PE-ATR.

1.4.1. Basic Performance Prediction Approaches for ATR Systems

Based on image measures quantifying the intrinsic difficulty of ATR, a performance forecaster is developed [20]. The performance measures include: constant false alarm rate (CFAR), power spectrum signature, probability of edge, etc. This algorithm offers a method for predicting ATR performance based on information extracted directly from the imagery [20]. The statistical accuracy is another basic index in performance predicting [55].

A generally employed performance prediction index is performance bound, namely, upper bound [56] and lower bound [21]. In this approach, the frequently considered performance include: detection rate, false alarm rate and recognition rate.

1.4.2. Performance Prediction Based on Performance Modeling

When predicting the performance for an ATR system, performance models are widely employed [20]. Simple models are easy to configure, but they cannot accurately quantify performance [57]. Detailed models may freely respond to the scenario, however, the detailed models are difficult to investigate [35,57].

When the features are distorted by uncertainty (occlusion and/or clutter) in both feature locations and magnitudes, the performance of an ATR system is especially difficult to predict. A practical way is to estimate the performance bound for the system [57]. For a vote-based object recognition system, forecasting lower and upper bound recognition ability is implemented [57]. This approach takes object model similarity into account, so that when models of objects are more similar to each other, then the desired recognition rate is lower [57].

The parameters of ATR algorithms can be used for predicting the performance for an ATR system [58]. The levels of robustness and invariance of parameters are employed as predictive indicators of ATR performance along with self refusal capabilities of the ATR algorithms [58].

A model of the subsystem of an ATR system can be introduced in forecasting the performance for the system [59]. One of the methods models the capability of the classifier. The classifier is based on a Bayes match between vector of extracted scattering features and a vector of predicted features. Uncertainty in both extracted and predicted features are included in the match metric (evaluation index) [59]. With scattering centers extracted from measured SAR imagery of ten targets, experiments show that the proposed match metric (evaluation index) is helpful in predicting the performance for an ATR system [59].

To estimate and predict the computational error of an ATR system, scientists developed error probability distribution method [60,61]. It is resolved from error function that is derived from the parse tree which represents a given ATR algorithm [60,61]. Field tests of performance prediction were performed in terms of computational accuracy, cost, and portability. The results show the prediction is reasonable [60,61].

Algorithm-independent predicting of the ATR performance is highly welcomed. To facilitate evaluation of performance tradeoffs for SAR designs, performance predictions are performed including both parameter selections (e.g., bandwidth and transmit power) and added domains of SAR observation, such as 3-D, full polarimetry, aspect diversity, and/or frequency diversity [62]. Discussion is made about performance of 3-D SAR includes parameter tradeoffs of various height resolutions at the target, and various numbers of sensors [62]. This work is significant in supporting SAR ATR designation.

1.4.3. Other Performance Prediction Method

To optimize the speech recognition performance in a computer assisted language learning system, a decision tree-based method is incorporated to predict possible speaking errors made by non-native speakers [63]. Trials of the language learning system and the performance prediction were conducted [63]. Positive feedback was reported [63].

The confidence interval is compared between different automatic fingerprint verification algorithms [50]. A performance model is built based on statistics. It can be applied to estimate the uniqueness of the template in classifiers [50].

1.4.4. Performance Prediction System or Performance Prediction Testbed

The afore-mentioned image measures (CFAR, power spectrum signature and probability of edge) are applied in a software which is implemented to validate the performance of some infrared (IR) image-based ATR algorithms [20]. For an imagery automatic target detection (ATD) system, these metrics are also employed in a software tool developed at Los Alamos National Laboratory [64]. A prototype software is developed to reveal the computational error of an ATR system [60,61].

1.5. Limitations of the Available Approaches on Performance Evaluation and Performance Prediction for an ATR System

Based on the materials presented above, the time-line of the evolution in PE-ATR is summarized in Figure 1. In the performance evaluation and the performance prediction work for an ATR system, the aforementioned methods offer choices for us. However, there are still remarkable weaknesses in this area:First of all, in the calculating course, most of the performance evaluation and the performance prediction approaches have not taken the operating condition into account. As a result, the performance evaluation and the performance prediction output may lack of objectiveness [19].

Secondly, the performance evaluation methods available can not work flexibly and no general reference frame has yet been built [22,41]. Furthermore, some of the performance evaluation indexes are too simple to reveal the problem-solving capability of an ATR system [65].

In addition, there are few perfected performance prediction tools that can be used to field test at present [66].

Therefore, in PE-ATR and PP-ATR, sound methodologies that are flexible to the scenario while exercising objectiveness are key topics [3].

1.6. Designation Objective of this Work and Its Layout

The contribution of this paper includes: (a) a measure to scale the operating condition for ATR; (b) the definition of performance evaluation indexes; (c) the construction of performance evaluating and performance predicting function. As a result, a novel approach is developed for the performance evaluation work in ATR. Compared to the existing methods, this approach is compacted, scenario adapting and easy to configure. In the evaluation or prediction course, this novel approach takes the operating condition into account, an objective conclusion may be arrived at.

In organizing this paper, the problem and its background are analyzed firstly. The key ideas of this work are explained. These are the main contents in Section 1. The rest of the data is organized as follows:

  • The majority of our work concerns the performance evaluation and performance evaluation work in ATR. This is further detailed in Section 2.

    The general idea of this methodology is summarized in Section 2.1.

    In Section 2.2, some similar technics related to ATR is identified and the ATR system's output is classified. The sample size in various experiments is resolved.

    To offer an objective evaluation conclusion, ATR system's condition is scaled in Section 2.3. The proposed index is enlightened by the measures of similarity in pattern recognition.

    In Sections 2.4 and 2.5, the performance evaluation work is implemented with performance evaluation indexes and an evaluation function. The proposed performance evaluation indexes are built based on the probability and mathematical statistics. The most important principles are the tests of statistical hypothesis: the hypothesis test of distribution specialty and the hypothesis test of independence.

    In Section 2.6, the performance predicting is realized with a generalized function. Based on the idea in expert prediction (EP, a branch of machine learning), the proposed performance predicting approach is built.

  • To confirm the practicability of this work, experiments are implemented in Section 3. The ATR algorithms setup and the data are explained in Section 3.1. Simulation results and the analysis of them are shown in Sections 3.2–3.4. Comprehensive topics related to this work are discussed in Section 3.5.

  • In Section 4, a summary is provided and the future topics are suggested.

In view of the proposed indexes, this work spans a number of scientific disciplines, and there are many references concerning those topics, though the related scientific background has not been presented in the text. However, the scientific background is figured out for each proposed index.

2. The Algorithm to Understand, Evaluate, and Predict the Performance of an ATR System

2.1. The Idea to Evaluate an ATR System's Performance

Because the ATR system is flexible and many constituent components interact in a complicated way, it is impossible to model an ATR system's output as the function of all the effective factors. A more viable approach (the idea in this work) is to observe the input and the corresponding output, and to determine the comprehensive performance in handling a certain target [26]. In carrying out the theoretics part of this work, we follow the listed steps below.

  • The definition of ATR is firstly investigated. The ATR system's output is classified. These are the foundation of the entire work.

  • Secondly, an index is proposed to scale the operating condition for recognition. This index can be further utilized in developing the performance evaluation index and performance evaluation function.

  • Thirdly, a series of evaluation index is developed. The precision, the robustness and the independence of the recognition output are measured.

  • The fourth step is building a performance evaluation function. The proposed evaluation indexes and the operating condition are integrated. A general conclusion may be arrived at with this function.

  • The final step is developing an algorithm to predict the ATR system's performance.

The idea and the main contribution is shown in Figure 2.

2.2. The Definition of Automatic Target Recognition, the Identification of Some Similar Technics, and the Classification of an ATR System's Output

As most researchers will admit, the main component of ATR is a signal processing course which trains the system with information regarding the concerned target in advance. The system can then be used to make decisions on the input signal about the potential target. Usually, its output is used for further decision making or action. Typically, there are three terms relating to this system: “classification”, “recognition” and “identification”. Some scientists have discussed this point [1,3]. Here, ATR is divided between automatic target discrimination (ATD) and automatic target identification (ATI). If the feed signal contains information from the trained target, the processing course is then called ATR. If there is only information from an untrained target in the collected signal, the processing course is called ATD, which, in nature, discriminates the signal as “having no information related to any trained target”. Moreover, if there is no information from any target in the obtained signal, then the processing course is called ATI. This, essentially identifies the signal as “having no information related to any target at all”.

The difference of these three technologies related to ATR is shown in Figure 3.

With these preparations, the output of an ATR system can be classified as in Table 1, where the variable nij for each category is the corresponding sample size when there are N tests in total, N = i = 1 3 j = 1 4 n i j; and [Info.], [R], [D], [I] stand for “information”, “recognition”, “discrimination”, and “identification”, respectively. It can be seen that there are three types of signal fed to the sensor: “target A”, “an untrained target”, and “no target”. So, i = 1, 2, 3. There are four types of output of the ATR system. So, j = 1, 2, 3, 4. Each false decision in these activities can be classified into a false type, as is shown in Table 1.

The designation of Table 1 is as follows. When the feeding signal containing information from target A, “False [R]” is the name of the decision type that the ATR system's output is another trained target other than target A, “Omitted [R]” is the name of the decision type that the ATR system's output is “cannot figure out the target type, ” and “True [R]” is the name of the decision type that the ATR system's output is target A.

2.3. Scaling the Condition for Recognition

In order to judge the ATR systems in an objective way, one must scale the condition for recognition. This is measured by a novel developed quantity called “Innovation for recognition (INR)”, which, through calculating the distance of the samples inside a certain target type and the distance among different target types, indicates the degree of difficulty in recognizing a certain trained target.

Firstly, for the testing samples (testing data) and the training samples (training data), the distance of the target's feature column vector between them is considered.

Suppose there are t1 different types of training targets in the system, the targets are distinguished by features in m dimensions, x(i1,i2) is the feature column vector of the target's testing samples. i1 shows the serial number of the target. i2 indicates the serial number of the sample. Here, i1 = 1, 2, …, t1, i2 = 1, 2, …, S ( i 1 w ). S ( i 1 w ) is the total number of the testing samples, (i3,i4) is the feature column vector of the target's training samples. i3 shows the serial number of the target. i4 indicates the serial number of the sample. Here, i3 = 1, 2, …, t1, i4 = 1, 2, …, S ( i 3 t ). S ( i 3 t ) is the total number of the training samples. As a result, for target i1 and target i3,

δ ( x ( i 1 , i 2 ) , x ˜ ( i 3 , i 4 ) ) = [ ( x ( i 1 , i 2 ) x ˜ ( i 3 , i 4 ) ) T ( x ( i 1 , i 2 ) x ˜ ( i 3 , i 4 ) ) ] 1 / 2
is the distance of the feature column vector from these two sets of samples.

Then, the INR of target i1 can be solved by:

d ˜ ( i 1 ) = 1 ( t 1 1 ) · i 2 = 1 S ( i 1 w ) ( i 3 = 1 t 1 , i 3 i 1 i 4 = 1 S ( i 3 t ) δ 1 ( · ) ) [ i 1 = 1 t 1 S ( i 1 w ) ] [ i 3 = 1 t 1 , i 3 i 1 S ( i 3 t ) ] · 1 1 S ( i 1 w ) S ( i 1 t ) · i 2 = 1 S ( i 1 w ) ( i 4 = 1 S ( i 1 t ) δ 2 ( · ) ) = S ( i 1 w ) S ( i 1 t ) i 2 = 1 S ( i 1 w ) ( i 3 = 1 t 1 , i 3 i 1 i 4 = 1 S ( i 3 t ) δ 1 ( · ) ) ( t 1 1 ) [ i 1 = 1 t 1 S ( i 1 w ) ] [ i 3 = 1 t 1 , i 3 i 1 S ( i 3 t ) ] i 2 = 1 S ( i 1 w ) ( i 4 = 1 S ( i 1 t ) δ 2 ( · ) )
where δ 1 ( · ) = δ ( x ( i 1 , i 2 ) , x ˜ ( i 3 , 1 4 ) ) , δ 2 ( · ) = δ ( x ( i 1 , i 2 ) , x ˜ ( i 1 , i 4 ) ). In Equation (2),
i 2 = 1 S ( i 1 w ) ( i 3 = 1 t 1 , i 3 i 1 i 4 = 1 S ( i 3 t ) δ ( · ) ) [ i 1 = 1 t 1 S ( i 1 w ) ] [ i 3 = 1 t 1 , i 3 i 1 S ( i 3 t ) ]
indicates the distance between the feature column vectors from (a) the testing samples of the training target i1, and (b) the training samples of all the training targets except for target i1. Here,
1 S ( i 1 w ) S ( i 1 t ) · i 2 = 1 S ( i 1 w ) ( i 4 = 1 S ( i 1 t ) δ ( · ) )
shows the distance between the feature column vectors from (a) the testing samples of the training target i1, and (b) the training samples of the target i1. 1 ( t 1 1 ) brings in an average among t1 1 targets. The normalized form of INR is used generally,
d ( i 1 ) = d ˜ ( i 1 ) / d 0
where d 0 = max ( d ˜ ( i 1 ) ) is the maximum value for all possible operating conditions. For a certain ATR system handling a certain target, the lower the INR, the more difficult it is to perform the recognition task.

In building the INR index, the related principle is the knowledge of feature extraction in pattern recognition, as is detailed in many literatures [67].

2.4. Performance Evaluation Indexes

For a practical ATR system, an accurate and robust output is overwhelmingly welcomed. It is important that the result should be independent to the run condition, or at least, should be influenced as little as possible. The following capacities are concerned:

(a)

The general approach of the recognition output (GARO). GARO weighs the recognition output, on the basis of whether or not it comes up with the desired level on correct decisions. Suppose the sample size in Table 1 fulfills the requirements in hypothesis testing. There are two schemes for GARO: naked GARO (n-GARO) and GARO with cost (c-GARO), denoted by I1 and I2 respectively,

I 1 = n 11 j = 2 4 n 1 j · n 23 j = 1 4 , j 3 n 2 j · n 34 j = 1 3 n 3 j
I 2 = ω 11 n 11 j = 2 4 ω 1 j n 1 j · ω 23 n 23 j = 1 4 , j 3 ω 2 j n 2 j · ω 24 n 34 j = 1 3 ω 3 j n 3 j
Here, ωij ≥ 1, i, j = 1, 2, 3, 4 are the assigned value of cost, usually, ω11 = ω23 = ω34 = 1. The cost in c-GARO is introduced to distinguish the risk of different types of decisions. These costs are empirically set according to the scenario.

If any fraction of I1 and/or I2 fall(s) into the 0/0 form, it is then set to 1.

The n-GARO is introduced with the knowledge of “summary measures” in statistics [68,69], while the c-GARO is found with the knowledge of “summary measures” in statistics and “numeric analysis” in engineering mathematics [68,70].

(b)

The robustness of the recognition output (RRO). RRO checks whether the operating output samples have the same distribution as that in the training course. RRO is revealed by the distribution specialty of GARO, through a rank-sum test. The related knowledge is “hypothesis testing” in statistics [68,69].

Suppose that there are n1 samples of n-GARO in the training course, while there are n2 samples of n-GARO in the testing course, all these samples are obtained under the same INR level. That is to say, within the same INR confidence interval. The Wilcoxon rank-sum test is applied [68]. Let R1 stand for the rank summation of the training samples, then,

I 3 = { 1 + | ln [ 2 R 1 n 1 ( n 1 + 1 ) n 1 ( n 1 + 2 n 2 + 1 ) 2 R 1 ] | } 1
is the normalized RRO. It shows whether the two concerned sample sets are subject to a uniform distribution, the idealized value of I3 is 1. Proof of this point can be found in [68,69].

The RRO with cost has not been touched here, as it can be arrived at in a similar manner.

(c)

The independence of the recognition output to condition (IRO). Through the hypothesis test of independence, IRO estimates the independence of the recognition output to condition. Here, the hypothesis test of independence shows the influence (or impact) of the testing condition on the ATR system's performance. The related knowledge is “hypothesis testing” in statistics [68,69].

The two sets involved in the test are INR and n-GARO. There are s1 subclasses in INR and s2 subclasses in n-GARO, P(INR = i, n-GARO = j) = pij, ∀i ∈ [1, s1], ∀j ∈ [1, s2]. The population (INR, n-GARO) has a sample size m. mij is the sample size when INR is in its ith subclass and n-GARO is in its jth subclass, mi. is the sample size when INR is in its ith subclass and all n-GARO subclasses. pi· = mi·/m·m·j and p·j have the similar meaning for n-GARO. Let:

χ 2 = i = 1 s 1 j = 1 s 2 ( m i j m p i · · p · j ) 2 m p i · · p · j = i = 1 s 1 j = 1 s 2 ( m i j m i · m · j / m ) 2 m i · m · j / m
stand for the test statistic, the threshold is η, then, IRO is arrived at by:
I 4 = [ 1 + 1 Λ | ln ( χ 2 / η ) | ] 1
where Λ is the variation range of d(i1), the idealized value of I4 is 1.

Further materials related to the hypothesis test of independence can be found in [68,71].

2.5. The Way to Understand and Evaluate the Performance of an ATR System

On the basis of the previous work, the performance of an ATR system can be interpreted and evaluated in two ways. One is to list the value of INR and the corresponding evaluation indexes. This can be easily realized, but the result can not be understood well by people outside of this field. Another way is to introduce a comprehensive function from these parameters, namely, the quality factor of the ATR system (QF-ATR) in attacking target i1,

Q ( i 1 ) = ( I 1 ( i 1 ) I 3 ( i 1 ) I 4 ( i 1 ) ) / d ( i 1 )
i1 = 1, 2, …, t1, through applying the Monte Carlo test, the final comprehensive comment may be obtained. Here, Q(i1) is the expression of QF-ATR in the calculating course. It should be noted that for I4(i1) in calculating QF-ATR, d(i1) is the mean value of its variation interval. When recognizing a certain target under a certain situation, the larger the QF-ATR is, the better performance the system maintains.

In a similar way, the QF-ATR with cost is resolved.

The QF-ATR index is introduced with the knowledge of “summary measures” in statistics and “numeric analysis” in engineering mathematics [68,70].

2.6. Predicting the Performance of an ATR System

Performance prediction work can be classified into three situational categories: (a) forecasting the performance for a repeated test with a familiar system and target; (b) predicting the performance for a tested system on a newly trained target; (c) figuring out the capability for a new ATR system on a familiar or a novel target. As an example, I1 is chosen as the performance index to be predicted.

First, to estimate the performance in managing a trained target i1 for a repeated test with INR equaling dj(i1), the test records of this target, with consistent INR, are taken from the database. These records are the seeds for forecasting work. A term imposed on the newly born set is that, its sample size should be no less than the requirements originating from the corresponding hypothesis testing. This term is effective here as well as in the following cases.

Another mission, is that to estimate the performance for the first test in coping with a newly trained target in. In order to proceed with the forecasting work, the database of training output is consulted. While the operating conditions are much more substantial, it is supposed that the target's INR states fall into dj(in), j = 1, 2, …, J in the training course. For predicting the performance in a certain state dm(in), m = 1, 2, …, M, the training record, whose INR is dm(in) ± o, is taken out. These records are the seeds for performing the prediction work. Here, o is a reasonable tiny quantity in practice. For example, o = 0.05 × dj(in).

The third one, but not the least important, is that to predict the performance for a newly developed ATR system, people may consult the systems with similar approaches in processing same or similar targets. The procedures are not duplicated as they are similar to those in the previous situations. One should be aware that even for the same target within a uniform environment, the INR may be different in different systems.

Once the preparation has been completed, a novel developed prediction methodology, referred to as “context-probability (CP)”, is applied. CP is useful for estimation and forecasting work in complicated systems such as an ATR system, where there are many different variables interacting in a complex fashion that can not be figured out in clear expressions. In addition, the system may provide increasingly accurate and robust results by incorporating historical data into the calculations. So, the new measure should take into account both sequential information and probability. The procedures are:

(a)

Collecting the seeds for prediction according to their sequence, here, I1(l), I1(2), …, I1(k) are harvested;

(b)

Calculating the context weight for the collected seeds, ω 1 ( m ) = e ( m k ) / j = 1 k e ( j k ), m = 1, 2, …, k;

(c)

Calculating the probability weight for the collected seeds, ω 2 ( m ) = Δ ( m ) / j = 1 k Δ ( j ), m = 1, 2, …, k; where Δ(m) and Δ(j) subject to the identical form: Δ ( μ ) = t = 1 k | I 1 ( t ) ( r = 1 k I 1 ( r ) / k ) ) | | I 1 ( μ ) ( r = 1 k I 1 ( r ) / k ) ) | , μ = 1 , 2 , , k ;

(d)

Calculating the general weight for the collected seeds, ω ( m ) = ω 1 ( m ) ω 2 ( m ) j = 1 k ω 1 ( j ) ω 2 ( j ); m = 1, 2, …, k;

(e)

Releasing the forecasting result for the system, I 1 ( k + 1 ) = j = 1 k ω ( j ) I 1 ( j ).

It is clear that for this kind of weighted average prediction, there is a group of choices for the weight average strategy. The above-mentioned way is one of them. The principal requirements for the weight average strategy are: (a) the fresher the data point, the larger the weight is; (b) the less distance between the data point and the mean value, the larger the weight is; (c) the final weight vector should be a normalized one.

As mentioned before, when one takes the knowledge of probability, statistics, and weighted average prediction into mind, a kind of performance prediction method is realized. Aside from this predicting method, one can forecast an ATR system's performance by using a machine learning facility called expert prediction [72], or with a data processing technology called bootstrapping [73]. In most situations, this method outperforms the others in that both the sequence and the probability are considered.

The flow diagram of the prediction algorithm is shown in Figure 4.

2.7. Summary of the Proposed Methodologies

As we have witnessed, the compilation of this work has thus far comprised of the performance evaluation measure for an ATR system, the performance prediction method for an ATR system, and a quantity to scale the operating condition is developed. The proposed methodologies are collected in Table 2. The relation among these performance indexes is shown in Figure 2. In Table 2, “SCR” means “Scaling the Condition for Recognition”.

3. Experiments

To validate this novel methodology, a series of simulations have been undertaken. A sampling of results follows. Before starting the discussion of the simulation, we should emphasize that the experiments here are: (a) to check whether the evaluation conclusion is in accordance with the performance inference; (b) to check whether the performance prediction output is proper compared to the practical performance; and (c) to validate whether the methodology can be applied to a variety range of ATR systems. Therefore, when performing experiments, there are 3 kind of ATR systems being tested. The capability of the proposed methods to be applied in various ATR systems is thus validated. Moreover, two similar ATR systems are considered. This is to check the ability of distinguishing the performance of similar ATR systems in similar scenarios.

3.1. The ATR Algorithms Setup and the Data

The proposed methodology in this work can be applied to all ATR systems and algorithms. However, the algorithms under consideration in the experiments are limited. There are 4 ATR algorithms taken into account: a SAR ATR method based on a global scattering center model [74], an improved approach for target discrimination in high-resolution SAR images [75], and an electrocardiograph (ECG) waveform recognition algorithm based on sparse decomposition and neural network (NN) [76]. They are named as Sys1, Sys2, and Sys3A respectively; a modified electroencephalograph (EEG) signal recognition measure based on empirical mode decomposition (EMD) and autoregression (AR), namely, Sys3B, is developed and validated to compare the performance results as in Sys3A.

Sys1 is configured according to [74] (recognizing targets I, II, and III, and is referred to as recognizing target 1, 2, and 3 in this work). Sys2 is implemented from [75] (recognizing target 6, 7, and 9, and is referred to as recognizing T6, T7, and T9 throughout this work). Sys3A is accomplished in conformity to [76] (recognizing P Pulse and T Pulse in this work). The EMD subsystem of Sys3B in feature extraction is directly implemented with respect to the EMD subsystem in [77]. The classifier in Sys3B is realized according to the classifier in [78]. The other subsystems in Sys3B and Sys3A are identical.

Sys1 and Sys2 are trained and tested with the data from [74,75], respectively; while Sys3A is trained and tested with the data from PhysioNet [79]. Sys3B is applied to the same data as in Sys3A.

The EEG data of University of California Irvine (UCI) arises from a large study to examine EEG correlates of genetic predisposition to alcoholism [80]. It contains measurements from 64 electrodes (medical sensors) placed on the scalp sampled at 256 Hz. Both the training portion and the test portion of the large data set are applied. The ECG data from PhysioNet applied are ECG [Class 1; core] long-term ST database.

3.2. Selected Simulation Results and Analysis

3.2.1. Partial Results (Performance Evaluation and Performance Prediction) and Analysis

Some of the performance evaluation results are given in Table 3, while the performance forecasting results and validation there of are shown in Table 4. The “PE” in these tables means “performance evaluation.” For each record of the performance evaluation indexes, the original sensing and recognizing sample size is 150 times. As for each record in Tables 3 and 4, it is obtained using the Monte Carlo test with 50 runs. The principle of performance model based on fuzzy integration (PM-FI) is detailed in [9]. The performance indexes (I1/I3/I4) are considered in PM-FI. The weight in these three indexes are all set as 1.

In Table 3, several interesting conclusions can be drawn. First, the recommended methodologies can offer well-founded judgment for the system, as long as the operating condition is varying. Secondly, the QF-ATR consider the performance not only with the output, but also with the operating condition. For example, the I1 level of Sys2 is much better than that of Sys1. At the same time, the value of I3 and I4 from these two systems are almost similar. It is unfortunate, that QF-ATR of Sys2 is about half of Sys1. The reason lies in the condition, as is indicated by INR. Third, this facility can clearly discriminate between systems when they handle identical targets under identical conditions. The evaluation results from Sys3A and Sys3B support this point. It is sure that EMD and AR methods maintain less relevance with the condition than sparse decomposition and neural network methods. The figures are in accordance with the inference.

In Table 4, the gap between the forecasting result and the actual output is slim. However, we should pay attention to the fact that each record is the mean value of 50 original performance prediction runs. The prediction error at each prediction step is still clear, as is shown by figures in the following subsections. The result in Table 4 is exciting. It is obvious that the prediction error of QF-ATR is much stronger than the other indexes. This stems from that QF-ATR is the function of the other variables. All the error will be collected into QF-ATR.

It may seem unusual that the QF-ATR can not strictly subject itself to Equation (11) with the listed I1/I3/I4 and INR. This stems from the fact that all indexes in Tables 3 and 4 are processed individually through the Monte Carlo test. The data has been derived individually from the mean value from each 50 run test. The performance prediction is performed using CP only.

Because the scenarios are not complicated, the prediction results of RR have high precision.

3.2.2. Performance Evaluation with ROC Method and Analysis

The evaluation results with ROC method are presented in Figure 5. Here, QDD is “quadratic distance discriminator”. WQDD is “weighted quadratic distance discriminator”. For Sys2, Sys3A and Sys3B in Figure 5, it may seem unusual that the RR is little decreasing while PFA is greater than a certain value and growing. This stems from the fact that the clutter is too heavy to be effectively processed in those scenarios.

3.2.3. Performance Evaluation with “Confusion Matrix” Method

The evaluation results of confusion matrix method are shown in Tables 5 and 6. In Table 5, “T1” means “Target 1”. The other targets are with the similar name. Here, the settings of the targets for Sys1 are: signal to noise ratio (SNR) is 10 dB, elevation is 10° and the result is arrived at with 500 Monte Carlo simulations [74]. The result of Sys2 is “Experiment and analysis od data provided by the Institute of Electronics, Chinese Academy of Sciences” [75]. In Table 6, “P Pulse” and “T Pulse” are different waveforms which have implications in medical science.

3.3. More Simulation Results With Brief Analysis

To clearly show the capability of the methodology, a mere fraction of the simulation results is presented.

The primary setting of the performance evaluation experiments has been collected in Table 7, where “Figure 6, 0.31” means the INR in Figure 6 for the corresponding system is 0.31. The remaining items follow this rule.

The primary setting of the performance prediction simulations has been collected in Table 8, where “Figure 8, 12” means the number of prediction seeds in Figure 8 is 12. The remaining figures follow this rule.

3.3.1. More Performance Evaluation Results on Sys1 and Sys2

The step-by-step performance evaluation results are presented in Figures 6 and 7. As demonstrated in Figures 6 and 7, even for a certain ATR system regarding a certain target under a certain condition, the performance shakes. However, the difference exists in the shaking range between different systems.

The upper-left part (Figures 6 and 7) suggests that the I1 of Sys1 is much more robust than the I1 of Sys2. For the I3 and the I4, Sys1 and Sys2 are similar in the first scenario. Moreover, there is a modest difference in the I3 and the I4 from Sys1 and Sys2 in the second scenario. One should be aware that each data point in performance evaluation is arrived at from N ATR tests in practice, as is shown in Table 1, as well as subjecting to Equations (6), (8), (10) and (11).

In Figure 6, it may seem unusual that the I1 of Sys2 is much better than those of Sys1, while the QF-ATR of Sys1 overwhelms those of Sys2. The reason lies in the difference of INR, which shows that the recognition condition is much worse for Sys1 than it is in Sys2.

As presented in these data, the performance of Sys1 is more robust than Sys2 in these two scenarios.

3.3.2. More Performance Prediction Results on Sys1 and Sys2

Detailed performance prediction results of the above-mentioned Sys1 and Sys2 are given accordingly (Figures 8, 9, 10, 11, 12, 13, 14 and 15). It can be seen that the performance prediction algorithm developed in this work is able to forecast the performance of an ATR system. One should note that each predicted data point here and thereafter is arrived at from a different number of prediction seeds (shown in Table 8), and subjects to the prediction procedures. The actual output is also obtained from N tests (shown in Table 1).

From Figures 8, 9, 10, Figures 12, 13, 14, one may know that, for a given ATR system under a certain condition, the fewer the prediction seeds, the more flexible of the prediction ability In most occasions, the especially poor match of the prediction results exists in the initial part. As the prediction continues, the error turns to decline.

3.3.3. Further Performance Evaluation Results on Sys3A and Sys3B

Figures 16 and 17 represent the performance evaluation results on Sys3A and Sys3B. Because the target being recognized is the same one and the designation of these two systems is similar, the actual outputs from Sys3A and Sys3B maintain a similar tendency. However, the performance marks from the proposed method are different. This result confirmed that, even in a challenging evaluation work, the newly developed methodology is suitable for evaluating the ATR system's performance.

In the upper-left part (Figures 16 and 17), the I1 of Sys3A is much more better than that of Sys3B. For the I3 and the I4, Sys3A and Sys3B are similar in these two scenarios. In the lower-right part (Figures 16 and 17), it is clear that the QF-ATR of Sys3A overwhelms those of Sys3B.

3.3.4. Further Performance Prediction Results on Sys3A and Sys3B

Detailed performance prediction results of the above-mentioned Sys3A and Sys3B are given respectively (Figures 18, 19, 20, 21, 22, 23, 24 and 25). These results confirmed that the proposed performance prediction method works well in forecasting the performance of Sys3A and Sys3B. While error exists in individual parts, the predicting accuracy is almost as well as that can be expected.

3.4. Comparison between the Existing Technologies and the Proposed Methodologies in Performance Evaluation For an ATR System

Based on the materials presented above, a comparison between the existing technologies and the proposed methodologies is performed in Table 9. As afore-mentioned, most of the existing performance prediction methods are extending work of performance evaluation technologies. Therefore, the comparison between performance prediction methods is not presented. Readers are encouraged to finish this work. The meaning of some symbols are list below.

  • LI: Is the operating condition considered in the evaluating course?

  • L2: The objectiveness of the evaluation result.

  • L3: The effectiveness of the method in revealing the performance from various aspect.

  • L4: The generalization of the method.

  • L5: Is the method easy to configure?

3.5. Discussion

As can be seen from the aforementioned data, the proposed methodology can offer reasonable performance evaluation and performance prediction results for the ATR systems. To ensure a practical and reliable mechanism, there are still some extended topics related to this work.

First, for some ATR systems, it may be difficult to determine INR. The features for recognition may be indistinct, or cannot be directly converted into variables, e.g., image, voice, smell and similar items which are used to recognize animals cannot be scaled into feature vectors. For signals that cannot be denoted with feature vectors, the INR is set to 1 temporarily for all s possible situations; then the system makes use of those signals, and the s “faked (because the INR has not been considered)” QF-ATR are arrived at as i, i = 1, 2, …, s, consequently, for the ith situation,

d ( i ) = i = 1 s Q ˜ i / ( Q ˜ i k = 1 s j = 1 s Q ˜ j Q ˜ k )
the QF-ATR with INR is then arrived at.

Second, it is meaningful to settle the sample size and know the degree of confidence in a field test. When the risk is assigned in a field test, the sample size and the degree of confidence can be solved by hypothesis testing.

In addition, if the sample size is less than the demand, bootstrapping can offer some help [68,73].

4. Conclusions

To sum up, this work offers a comprehensive performance analysis tool for ATR systems. For various system processing an identical target under various condition, the evaluation results by this novel facility can reveal the accomplishment of the system by the evaluation indexes and QF-ATR, as is confirmed by the experimental results. At the same time, it has no limitations and presumptions imposed on the system being considered.

For a given ATR system, the INR index can scale the operating condition in an objective way; the evaluation indexes and the evaluation function serve to interpret the system's accomplishments. The QF-ATR factor, like the quality factor in circuits, may reveal the general capabilities of the entire system. All the proposed methodologies is suitable for all existing ATR systems. However, the methodologies are especially helpful for ATR in radars and photo-sensors.

While convenient to exercise, this methodology is unfamiliar at first sight since it is newly proposed. Although it is still too early to determine whether or not this is the most suitable way to conduct PE-ATR, the results it provides will place PE-ATR on a more objective and quantitative footing. It can also serve as a reference for performance analysis of similar systems.

The future research on this topic may origin from:

  • Validation of the methodology with large scale field tests.

  • Application in different ATR systems.

  • Performance evaluation and performance prediction with less samples.

Acknowledgments

The work in this paper has been supported by the National Science Fund for Distinguished Young Scholars of China under Grant No. 61025006, the National Key Lab Program of Complex Electronic-magnetic Environment Effects (CEMEE) (Luoyang, 471003) under Grant No. CEMEE2014K0204B, the National Natural Science Foundation of China under Grant No. 61171133, the Innovation Team Project from Ministry of Education of China under Grant No. IRT1120, and State Natural Science Fund for Distinguished Young Scholars of Hunan China under Grant No. 11zz1010.

Valuable data and reference ATR algorithms came freely from PhysioNet and UCI, Jixiang Sun, Chunguang Wang, Shufang Li, Weidong Zhou, Wenming Cao, Jianxiong Zhou, Gui Gao, which made the simulation here feasible. Julie Kilborn and Arthur Louis have offered good suggestions on the grammar of the manuscript. The authors also wish to extend their sincere thanks to editors and reviewers for their careful reading and fruitful suggestions.

All this support is deeply appreciated.

Glossary of Abbreviations

AAF

amplitude affection factor

AR

autoregression

Auto-I

automated instrumentation and evaluation

CFAR

constant false alarm rate

CM

confusion matrix

ECG

electrocardiograph

EEG

electroencephalograph

EMD

empirical mode decomposition

EOC

extended operating condition

EP

expert prediction

ER

extend of recognition

FPGA

flexible field programmable gate array

IR

infrared

LB-RR

lower bound of recognition rate

MFRR

measurement of false recognition rate

MOE

measures of effectiveness

MRR

measurement of recognition rate

NN

neural network

PE

performance evaluation

PE-ATR

performance evaluation for ATR systems

PM-FI

performance models based on fuzzy integration

POLInSAR

polarimetric SAR interferometry

PP-ATR

performance prediction for ATR systems

QDD

quadratic distance discriminator

ROC

receiver operating characteristic

RR

recognition rate

SAR

synthetic aperture radar

SCR

scaling the condition for recognition

SNR

signal to noise ratio

SNRF

signal to noise ratio affect factor

START

the scoring, truthing, and registration toolkit

UB-RR

upper bound of recognition rate

UCI

University of California Irvine

WQDD

weighted quadratic distance discriminator

Author Contributions

Yanpeng Li was responsible for all the theoretical work here, while Xiang Li, Hongqiang Wang and Zhaowen Zhuang collected and sorted the data for the experiments. Yiping Chen, Yongqiang Cheng, Bin Deng, Liandong Wang, Yonghu Zeng and Lei Gao performed the experiments. Yanpeng Li wrote and revised the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dudgeon, D.E. An overview of automatic target recognition. Linc. Lab. J. 1993, 6, 3–10. [Google Scholar]
  2. Huang, J.; Zhou, Q.; Zhang, X.; Song, E.; Li, B.; Yuan, X. Seismic target classification using a wavelet packet manifold in unattended ground sensors systems. Sensors 2013, 13, 8534–8550. [Google Scholar]
  3. Bhanu, B. Automatic target recognition: State of the art survey. IEEE Trans. Aerosp. Electron. Syst. 1986, 22, 364–379. [Google Scholar]
  4. Wei, Y.; Meng, H.; Liu, Y.; Wang, X. Extended target recognition in cognitive radar networks. Sensors 2010, 10, 10181–10197. [Google Scholar]
  5. Tait, P. Introduction to Radar Target Recognition; The Institution of Electrical Engineers: Herts, UK, 2005. [Google Scholar]
  6. Roth, M.W. Survey of neural network technology for automatic target recognition. IEEE Trans. Neural Netw. 1990, 1, 28–43. [Google Scholar]
  7. Klopčič, J.; Ambrožič, T.; Marjetič, A.; Bogatin, S.; Pulko, B.; Logar, J. Use of automatic target recognition system for the displacement measurements in a small diameter tunnel ahead of the face of the motorway tunnel during excavation. Sensors 2008, 8, 8139–8155. [Google Scholar]
  8. Pei, D.; Ulrich, H.; Schultz, P. A combinatorial approach toward DNA recognition. Science 1991, 253, 1408–1411. [Google Scholar]
  9. Li, Y. Performance Evaluation in Automatic Target Recognition: Foundation, Theoretic Framework and Related Research. Ph.D. Thesis, National University of Defense Technology, Changsha, China, 2004. [Google Scholar]
  10. Gronwall, C.; Gustafsson, E.; Millnert, M. Ground target recognition using rectangle estimation. IEEE Trans. Image Process. 2006, 15, 3400–3408. [Google Scholar]
  11. Zhao, Q.; Principe, J.C. Support vector machines for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 643–654. [Google Scholar]
  12. Diaz-Bolado, A.; Barriere, P.A.; Laurin, J.J. Study of microwave tomography measurement setup configurations for breast cancer detection based on breast compression. Int. J. Antennas Propag. 2013, 2013. [Google Scholar] [CrossRef]
  13. Mitton-Fry, R.M.; DeGregorio, S.J.; Wang, J.; Steitz, T.A.; Steitz, J.A. Poly(A) tail recognition by a viral RNA element through assembly of a triple helix. Science 2010, 330, 1244–1247. [Google Scholar]
  14. Sadjadi, E.; Javid, B. Physics of Automatic Target Recognition; Springer: Berlin, Germany, 2007. [Google Scholar]
  15. Jain, A.; Moulin, P.; Miller, M.I.; Ramchandran, K. Information-theoretic bounds on target recognition performance based on degraded image data. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 1153–1166. [Google Scholar]
  16. López-Rodríguez, P.; Fernández-Recio, R.; Bravo, I.; Gardel, A.; Lázaro, J.L.; Rufo, E. Computational burden resulting from image recognition of high resolution radar sensors. Sensors 2013, 13, 5381–5402. [Google Scholar]
  17. Álvarez Vallina, L.; Yanez, R.; Blanco, B.; Gil, M.; Russell, S.J. Pharmacologic suppression of target cell recognition by engineered T cells expressing chimeric T-cell receptors. Cancer Gene Ther. 2000, 7, 526–529. [Google Scholar]
  18. Nasr, H.; Sadjadi, F. Automatic target recognition algorithm performance evaluation: The bottleneck in the development life cycle. Proceedings of the SPIE, Aerospace Pattern Recognition, Orlando, FL, USA, 27 March 1989.
  19. Bassham, C.B. Automatic Target Recognition Classification System Evaluation Methodology. Ph.D. Thesis, Air Force Institute of Technology, Wright-Patterson AFB, OH, USA, 2002. [Google Scholar]
  20. Ralph, S.K.; Irvine, J.M.; Snorrason, M.; Stevens, M.R.; Vanstone, D. An image metric-based ATR performance prediction testbed. Proceedings of the 34th Applied Imagery and Pattern Recognition Workshop, Washington, DC, USA, 19–21 October 2005; pp. 6–12.
  21. Fu, Q.; He, J. The Methodologies and Applications of Performance Evaluation Technology for Automatic Target Recognition; National Defense Industry Press: Beijing, China, 2013. [Google Scholar]
  22. Lui, H.S.; Shuley, N.V. Performance evaluation of subsurface target recognition based on ultra-wide-band short-pulse excitation. Proceedings of the IEEE International Symposium Antennas and Propagation Society (APSURSI), Toronto, ON, Canada, 11–17 July 2010; pp. 1–4.
  23. Bechar, A.; Meyer, J.; Edan, Y. An objective function to evaluate performance of human-robot collaboration in target recognition tasks. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2009, 39, 611–620. [Google Scholar]
  24. Irvine, J.M. Evaluation of ATR algorithms employing motion imagery. Proceedings of the 30th Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 1–12 October 2001; pp. 57–63.
  25. Ross, T.D.; Westerkamp, L.A.; Zelnio, E.G.; Burns, T.J. Extensibility and other model-based ATR evaluation concepts. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery IV, Orlando, FL, USA, 28 July 1997; Zelnio, E.G., Ed.; Volume 3070.
  26. Zhuang, Z.; Li, X.; Li, Y.; Wang, H. Performance Evaluation Technology for Automatic Target Recognition; National Defense Industry Press: Beijing, China, 2006. [Google Scholar]
  27. Mossing, J.C.; Ross, T.D. An evaluation of SAR ATR algorithm performance sensitivity to MSTAR extended operating conditions. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery V, Orlando, FL, USA, 14–17 April 1998; Zelnio, E.G., Ed.; Volume 3370.
  28. Bhanu, B.; Jones, T.L. Image understanding research for automatic target recognition. IEEEA erosp. Electron. Syst. Mag. 1993, 8, 15–23. [Google Scholar]
  29. Clark, L.; Velten, V.J. Image characterization for automatic target recognition algorithm evaluations. Opt. Eng. 1991, 30, 147–153. [Google Scholar]
  30. Kaufman, V.I.; Ross, T.D.; Lavely, E.M.; Blasch, E.P. Score-based SAR ATR performance model with operating condition dependencies. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery XIV, Orlando, FL, USA, 9 April 2007.
  31. Zelnio, E.G.; Garber, F. A characterization of ATR performance evaluation. Proceedings of the SPIE, Signal Processing, Sensor Fusion, and Target Recognition V, Orlando, FL, USA, 8–10 April 1996.
  32. Nasr, H.N.; Sadjadi, F.A. Automatic evaluation and adaptation of automatic target recognition systems. Proceedings of the SPIE, Signal and Image Processing Systems Performance Evaluation, Orlando, FL, USA, 19–20 April 1990.
  33. Wang, E.; Li, Y.; Li, X. Performance evaluation for automatic target recognition based on cloud theory. Proceedings of the IEEE International Conference on Radar, Adelaide, Australia, 2–5 September 2008; pp. 498–502.
  34. Alsing, S.; Blasch, E.P. Baue, R. 3D ROC surface concepts for evaluation of target recognition algorithms faced with the unknown target detection problem. Procedings of the SPIE Automatic Target Recognition IX, Orlando, FL, USA, 13–17 April 1999.
  35. Dudgeon, D.E. ATR.Performance modeling and estimation. Digit. Signal Process. 2000, 10, 269–285. [Google Scholar]
  36. Novak, L.M.; Halversen, S.D.; Owirka, G.J.; Hiett, M. Effects of polarization and resolution on the performance of a SAR automatic target recognition system. Line. Lab. J. 1995, 8, 49–68. [Google Scholar]
  37. Margarit, G.; Mallorqui, J.J. Assessment of polarimetric SAR interferometry for improving ship classification based on simulated data. Sensors 2008, 8, 7715–7735. [Google Scholar]
  38. Zelnio, E.G.; Garber, F.D. A comparison of SAR ATR performance with information theoretic predictions. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery X, Orlando, FL, USA, 21 April 2003.
  39. Novak, L.M.; Owirka, G.J.; Netishen, C.M. Performance of a high-resolution polarimetric SAR automatic target recognition aystem. Line. Lab. J. 1993, 6, 11–24. [Google Scholar]
  40. Zhu, J.; Luo, Y.; Zhou, J. Sensor reliability evaluation scheme for target classification using belief function theory. Sensors 2013, 13, 17193–17221. [Google Scholar]
  41. Chang, C.I. Multiparameter receiver operating characteristic analysis for signal detection and classification. IEEE Sens. J. 2010, 10, 423–42. [Google Scholar]
  42. Alsing, S.G. The Evaluation of Competing Classifiers. Ph.D. Thesis, Air Force Institute of Technology, Wright-Patterson AFB, OH, USA, 2000. [Google Scholar]
  43. Parker, D.R.; Gustafson, S.C.; Ross, T.D. Receiver operating characteristic and confidence error metrics for assessing the performance of automatic target recognition systems. Opt. Eng. 2005, 44. [Google Scholar] [CrossRef]
  44. Boshra, M.; Bhanu, B. Bounding SAR ATR performance based on model similarity. Proceedings of the SPIE, Synthetic Aperture Radar Algorithms for Imagery VI, Orlando, FL, USA, 5 April 1999.
  45. Zelnio, E.G. Analytic performance bounds on SAR-image target recognition using physics-based signatures. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery VIII, Orlando, FL, USA, 16ȓ19 April 2001.
  46. Sadjadi, F.A. Information theoretic bounds of ATR algorithm performance for sidescan sonar target classifcation. Proceedings of the SPIE, Automatic Target Recognition XV, Orlando, FL, USA, 28 March 2005.
  47. Grenander, U.; Miller, M.I.; Srivastava, A. Hilbert-schmidt lower bounds for estimators on matrix lie groups for ATR. IEEE Trans. Pattern Anal. Mach Intell. 1998, 20, 790–802. [Google Scholar]
  48. Zelnio, E.G. Confidence intervals for ATR performance metrics. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery VIII, Orlando, FL, USA, 16–19 April 2001; Volume 4382.
  49. Zelnio, E.G. Fidelity Score for ATR Performance Modeling. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery XII, Orlando, FL, USA, 28–31 March 2005.
  50. Ren, Q. Research on Recognition Performance of Automatic Fingerprint Verification Algorithms. Ph.D. Thesis, Chinese Academy of Sciences, Beijing, China, 2003. [Google Scholar]
  51. Schmalz, M.S.; Ritter, G.X.; Caimi, F.M. Performance evaluation of data compression transforms for underwater imaging and object recognition. Proceeding of the MTS/IEEE Conference on OCEANS′97, Halifax, NS, Canada, 6–9 October 1997; pp. 1075–1081.
  52. ÓToole, A.J.; Phillips, P.J.; An, X.; Dunlop, J. Demographic effects on estimates of automatic face recognition performance. Image Vis. Comput. 2012, 30, 169–176. [Google Scholar]
  53. Sadjadi, F.A. Open source tools for ATR development and performance evaluation. Proceedings of the SPIE, Automatic Target Recognition XII, Orlando, FL, USA, 1–5 April 2002.
  54. Sadjadi, F.A. A scoring, truthing, and registration toolkit for evaluation of target detection and tracking. Proceedings of the SPIE, Automatic Target Recognition XIV, Orlando, FL, USA, 12 April 2004.
  55. Zelnio, E.G.; Garber, F.D. Shadow-based SAR ATR performance prediction. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery XVI, Orlando, FL, USA, 13 April 2009.
  56. Zelnio, E.G. Upper bound calculation of ATR performance for ladar sensors. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery V, Orlando, FL, USA, 13 April 1998.
  57. Appleby, R.; Wikner, D.A.; Trebits, R.; Kurtz, J.L. Performance modeling of vote-based object recognition. Proceedings of the SPIE, Passive Millimeter-Wave Imaging Technology VI and Radar Sensor Technology VII, Orlando, FL, USA, 23–24 April 2003.
  58. Kovalerchuk, B. Modeling ATR processes to predict their performance by using invariance, robustness and self-refusal approach. Proceedings of the 12th International Conference on Information Fusion (FUSION ′09), Seattle, WA, USA, 6–9 July 2009; pp. 1139–1146.
  59. Chiang, H.C.; Moses, R.L. ATR performance prediction using attributed scattering features. Proceedings of the SPIE, Conference on Algorithms for Synthetic Aperture Radar Imagery VI, Orlando, FL, USA, 5–9 April 1999; pp. 785–796.
  60. Schmalz, M.S. Automated analysis and prediction of accuracy and performance in ATR algorithms: I. Requirements, theory, and software implementation. Proceedings of the SPIE, Conference on Detection and Remediation Technologies for Mines and Minelike Targets II, Orlando, FL, USA, 21–24 April 1997.
  61. Schmalz, M.S. Automated analysis and prediction of accuracy and performance in ATR algorithms: II. Experimental results and system performance analysis. Proceedings of the SPIE, Conference on Detection and Remediation Technologies for Mines and Minelike Targets II, Orlando, FL, 21–24 April 1997.
  62. Horowitz, L.L.; Brendel, G.F. Fundamental SAR ATR performance predictions for design tradeoffs: 1-D HRR versus 2-D SAR versus 3-D SAR. Proceedings of the SPIE, Algorithms for Synthetic Aperture Radar Imagery VI, Orlando, FL, USA, 5 April 1999.
  63. Wang, H.; Waple, C.J.; Kawahara, T. Computer Assisted Language Learning system based on dynamic question generation and error prediction for automatic speech recognition. Speech Commun. 2009, 51, 995–1005. [Google Scholar]
  64. Sadjadi, F.A. Evaluation testbed for ATD performance prediction (ETAPP). Proceedings of the SPIE, Automatic Target Recognition XVII, Orlando, FL, USA, 9 April 2007.
  65. ATR algorithm performance evaluation based on the simulation image and real image. Proceedings of the SPIE, MIPPR 2013: Pattern Recognition and Computer Vision, Wuhan, China, 26 October 2013.
  66. Ratches, J.A.; Walters, C.P.; Buser, R.G.; Guenther, B.D. Aided and automatic target recognition based upon sensory inputs from image forming systems. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 1004–1019. [Google Scholar]
  67. Tou, J.T.; Gonzalez, R.C. Pattern Recognition Principles; Addison Wesley: Boston, MA, USA, 1974. [Google Scholar]
  68. Ugarte, M.D.; Militino, A.F.; Arnholt, A.T. Probability and Statistics with R; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar]
  69. Devore, J.L.; Berk, K.N. Modern Mathematical Statistics with Applications; Thomson Brooks/Cole: Pacific Grove, CA, USA, 2007. [Google Scholar]
  70. Greenberg, M.D. Advanced Engineering Mathematics; Publishing House of Electronics Industry: Beijing, China, 2011. [Google Scholar]
  71. Leon-Garcia, A. Probability, Statistics, and Random Processes for Electrical Engineering; Pearson Education Inc.: Upper Saddle River, NJ, USA, 2009. [Google Scholar]
  72. Cesa-Bianchi, N.; Lugosi, G. Prediction, Learning, and Games; Cambridge University Press: Cambridge, UK, 2006. [Google Scholar]
  73. Christopher, D.A.; Hinkley, D.V. Bootstrap Methods and Their Application; Cambridge University Press: London, UK, 1997. [Google Scholar]
  74. Zhou, J.; Shi, Z.; Cheng, X.; Fu, Q. Automatic target recognition of SAR images based on global scattering center model. IEEE Trans. Geosci. Remote Sen. 2011, 49, 3713–3729. [Google Scholar]
  75. Gao, G. An improved scheme for target discrimination in high-resolution SAR images. IEEE Trans. Geosci. Remote Sen. 2011, 49, 277–294. [Google Scholar]
  76. Liu, J.; Wang, C.; Sun, J. The detection and recognition of electrocardiogram's waveform based on sparse decomposition and neural network. Signal Process. 2011, 27, 843–850. (In Chinese). [Google Scholar]
  77. Li, S.; Zhou, W.; Cai, D.; Liu, K.; Zhao, J. EEG signal classification based on EMD and SVM. Biomed. Eng. 2011, 28, 891–894. (In Chinese). [Google Scholar]
  78. Wang, Z.; Luo, L.; Dong, W.; Zheng, S.; Wu, X. A infrared target recognition method based on classifier combination. J. Detect. Control. 2012, 34, 61–66. (In Chinese). [Google Scholar]
  79. Goldberger, A.L.; Amaral, L.A.N.; Hausdorff, G.L.J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank PhysioToolkit. PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 2000, 101, 215–220. [Google Scholar]
  80. EEG Database. Available online: http://archive.ics.uci.edu/ml/datasets/EEG+Database (accessed on 1 July 2012).
Figure 1. The time-line of the evolution in PE-ATR & PP-ATR.
Figure 1. The time-line of the evolution in PE-ATR & PP-ATR.
Sensors 14 11308f1 1024
Figure 2. The practical way to evaluate an ATR system's performance.
Figure 2. The practical way to evaluate an ATR system's performance.
Sensors 14 11308f2 1024
Figure 3. The difference among ATR, ATD and ATI.
Figure 3. The difference among ATR, ATD and ATI.
Sensors 14 11308f3 1024
Figure 4. The configuration steps of the proposed prediction algorithm.
Figure 4. The configuration steps of the proposed prediction algorithm.
Sensors 14 11308f4 1024
Figure 5. Performance evaluation results of the mentioned ATR systems with ROC approach. (a) The recognition rate of Sys1 when SNR is changing; (b) The recognition rate of Sys2 when false alarm setting is changing; (c) The recognition rate of Sys3A and Sys3B when false alarm setting is changing.
Figure 5. Performance evaluation results of the mentioned ATR systems with ROC approach. (a) The recognition rate of Sys1 when SNR is changing; (b) The recognition rate of Sys2 when false alarm setting is changing; (c) The recognition rate of Sys3A and Sys3B when false alarm setting is changing.
Sensors 14 11308f5 1024
Figure 6. Detailed performance evaluation results of Sys1 and Sys2 for 50 runs, with INR_Sys1 = 0.31, INR_Sys2 = 0.79. (a) Detailed results of I1; (b) Detailed results of h; (c) Detailed results of I4; (d) Detailed results of Q.
Figure 6. Detailed performance evaluation results of Sys1 and Sys2 for 50 runs, with INR_Sys1 = 0.31, INR_Sys2 = 0.79. (a) Detailed results of I1; (b) Detailed results of h; (c) Detailed results of I4; (d) Detailed results of Q.
Sensors 14 11308f6 1024
Figure 7. Detailed performance evaluation results of Sys1 and Sys2 for 50 runs, with INR_Sys1 = 0.77; INR_Sys2 = 0.56. (a) Detailed results of I1; (b) Detailed results of I3; (c) Detailed results of I4; (d) Detailed results of Q.
Figure 7. Detailed performance evaluation results of Sys1 and Sys2 for 50 runs, with INR_Sys1 = 0.77; INR_Sys2 = 0.56. (a) Detailed results of I1; (b) Detailed results of I3; (c) Detailed results of I4; (d) Detailed results of Q.
Sensors 14 11308f7 1024
Figure 8. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.31, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 8. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.31, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f8 1024
Figure 9. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.31, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 9. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.31, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f9 1024
Figure 10. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.31, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 10. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.31, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f10 1024
Figure 11. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.55, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 11. Detailed performance prediction results of Sys1 for 50 runs (INR = 0.55, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f11 1024
Figure 12. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.79, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 12. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.79, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f12 1024
Figure 13. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.79, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 13. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.79, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f13 1024
Figure 14. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.79, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 14. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.79, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f14 1024
Figure 15. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.63, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 15. Detailed performance prediction results of Sys2 for 50 runs (INR = 0.63, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f15 1024
Figure 16. Detailed performance evaluation results of Sys3A and Sys3B for 50 runs, with INR_Sys3A = 0.36, and INR_Sys3B = 0.36. (a) Detailed results of I1; (b) Detailed results of I3; (c) Detailed results of I4; (d) Detailed results of Q.
Figure 16. Detailed performance evaluation results of Sys3A and Sys3B for 50 runs, with INR_Sys3A = 0.36, and INR_Sys3B = 0.36. (a) Detailed results of I1; (b) Detailed results of I3; (c) Detailed results of I4; (d) Detailed results of Q.
Sensors 14 11308f16 1024
Figure 17. Detailed performance evaluation results of Sys3A and Sys3B for 50 runs, with INR_Sys3A=0.40, and INR_Sys3B=0.40. (a) Detailed results of I1; (b) Detailed results of I3; (c) Detailed results of I4; (d) Detailed results of Q.
Figure 17. Detailed performance evaluation results of Sys3A and Sys3B for 50 runs, with INR_Sys3A=0.40, and INR_Sys3B=0.40. (a) Detailed results of I1; (b) Detailed results of I3; (c) Detailed results of I4; (d) Detailed results of Q.
Sensors 14 11308f17 1024
Figure 18. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.36, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 18. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.36, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f18 1024
Figure 19. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.36, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 19. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.36, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f19 1024
Figure 20. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.36, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 20. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.36, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f20 1024
Figure 21. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.49, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 21. Detailed performance prediction results of Sys3A for 50 runs (INR = 0.49, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f21 1024
Figure 22. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.36, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 22. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.36, the number of prediction seeds is 12). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f22 1024
Figure 23. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.36, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 23. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.36, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f23 1024
Figure 24. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.36, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 24. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.36, the number of prediction seeds is 3). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f24 1024
Figure 25. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.49, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Figure 25. Detailed performance prediction results of Sys3B for 50 runs (INR = 0.49, the number of prediction seeds is 6). (a) Predicted and actual I1; (b) Predicted and actual I3; (c) Predicted and actual I4; (d) Predicted and actual Q.
Sensors 14 11308f25 1024
Table 1. For an automatic target recognition system, all the decision types and their names.
Table 1. For an automatic target recognition system, all the decision types and their names.
[Info.] in the Feeding SignalThe Final Decision of an ATR System
Target AOther Trained TargetUntrained TargetNo Target
Target ATrue [R]False [R]False [R]Omitted [R]
(n11)type I (n12)type II (n13)(n14)
An Untrained TargetFalse [D]False [D]True [D]Omitted [D]
type I (n21)type I (n22)(n23)(n24)
No TargetFalse [I]False [I]False [I]True [I]
type I (n31)type I (n32)type II (n33)(n34)
Table 2. Summary of the proposed methodologies.
Table 2. Summary of the proposed methodologies.
TechnologyThe Proposed Methodologies
SCRINR. INR is employed to scale the condition for recognition. It is built based on the knowledge of pattern recognition.
PE-ATRGARO. There are two types of GARO: n-GARO and c-GARO. GARO shows the capability on correct decisions of the ATR system. The related knowledge are statistics and engineering mathematics.
RRO. RRO reveals how close the distribution of the operating output is likely to the same distribution as that in the training course. RRO is developed based on statistics.
IRO. IRO estimates the independence of the recognition output to the operating condition. IRO is investigated based on statistics.
QF-ATR. QF-ATR estimates the comprehensive performance of an ATR system. It is empirically proposed according to the background of ATR and is resolved based on the proposed evaluation indexes.
PP-ATRCP CP forecasts the performance of an ATR system. CP is developed based on the knowledge of random processes and regression.
Table 3. Partial performance evaluation results for the mentioned ATR algorithms.
Table 3. Partial performance evaluation results for the mentioned ATR algorithms.
MetricsATR System

Sys1Sys2Sys3ASys3B
INR (d(i1))0.310.790.360.36
PE index (I1/I3/I4)6.14/0.82/0.8311.03/0.67/0.7510.06/0.81/0.667.02/0.75/0.78
QF-ATR (Q)13.996.9514.8511.33
RR0.710.820.850.83
PM-FI0.760.780.820.82
Table 4. Performance forecasting results and validation.
Table 4. Performance forecasting results and validation.
INR, PE Indexes and QF-ATRATR System

Sys1Sys2Sys3ASys3B
INR (d(i1))0.310.790.360.36

PE index (I1/I3/I4, ▷)6.16/0.68/0.8411.04/0.70/0.799.92/0.61/0.807.16/0.73/0.83
PE index (I1/I3/I4, ◁)6.14/0.65/0.8010.98/0.69/0.839.88/0.61/0.827.11/0.73/0.82

QF-ATR (Q, ▷)11.247.7614.1712.11
QF-ATR (Q, ◁)10.357.9713.8211.96

UB-RR (▷)0.750.850.900.90
LB-RR (◁)0.700.800.850.85
RR(▷)0.700.830.860.87

Legend: ▷: forecasted output. ◁: tested result. UB-RR: upper bound of recognition rate. LB-RR: lower bound of recognition rate.

Table 5. Performance evaluation results with “Confusion Matrix Method” for the mentioned ATR algorithms Sys1 and Sys2.
Table 5. Performance evaluation results with “Confusion Matrix Method” for the mentioned ATR algorithms Sys1 and Sys2.
Sys1Sys2


TargetOmitted [R]TlT2T3TargetOmitted [R]T6T7T9
Tl2443505T603521
T23204752T701334
T313611470T910136
Table 6. Performance evaluation results with “Confusion Matrix Method” for the mentioned ATR algorithms Sys3A and Sys3B.
Table 6. Performance evaluation results with “Confusion Matrix Method” for the mentioned ATR algorithms Sys3A and Sys3B.
SignalSys3ASys3B


Omitted [R]P PulseT PulseOmitted [R]P PulseT Pulse
P Pulse25945582
T Pulse32600461
Table 7. The primary setting of the performance evaluation experiments.
Table 7. The primary setting of the performance evaluation experiments.
MetricATR System

Sys1Sys2 Sys3ASys3B
INRFigure 6, 0.31Figure 6, 0.79Figure 16, 0.36
Figure 7, 0.77Figure 17, 0.40Figure 17, 0.40
Table 8. The primary setting of the performance prediction experiments.
Table 8. The primary setting of the performance prediction experiments.
Name of the ParameterATR System

Sys1Sys2Sys3ASys3B
INR0.310.790.360.36

Number of Prediction SeedsFigure 8, 12/Figure 9, 6/Figure 10, 3Figure 12, 12/Figure 13, 6/Figure 14, 3Figure 18, 12/Figure 19, 6/Figure 20, 3Figure 22, 12/Figure 23, 6/Figure 24, 3

INR0.550.630.490.49

Number of Prediction SeedsFigure 11, 3Figure 15, 3Figure 21, 3Figure 25, 3
Table 9. The comparison between the existing technologies and the proposed methodologies in performance evaluation for an ATR system.
Table 9. The comparison between the existing technologies and the proposed methodologies in performance evaluation for an ATR system.
AspectExisting TechnologiesNewly Proposed Methodologies


ROCCMRRPM-FIGARORROIROQF-ATR
LI×××Sensors 14 11308t1
L2Sensors 14 11308t2Sensors 14 11308t2Sensors 14 11308t2Sensors 14 11308t1Sensors 14 11308t2Sensors 14 11308t2Sensors 14 11308t3Sensors 14 11308t3
L3Sensors 14 11308t2Sensors 14 11308t2Sensors 14 11308t4Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3
L4Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t1Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3
L5Sensors 14 11308t3Sensors 14 11308t5Sensors 14 11308t5Sensors 14 11308t2Sensors 14 11308t5Sensors 14 11308t3Sensors 14 11308t3Sensors 14 11308t3

Legend: The following symbols are effective in all the tables in this work. Sensors 14 11308t5: high achievement; Sensors 14 11308t3: satisfactory; Sensors 14 11308t2: improvement needed; Sensors 14 11308t4 unsatisfactory; Sensors 14 11308t1: should be considered according to the corresponding method. ✓: yes; ×: no.

Share and Cite

MDPI and ACS Style

Li, Y.; Li, X.; Wang, H.; Chen, Y.; Zhuang, Z.; Cheng, Y.; Deng, B.; Wang, L.; Zeng, Y.; Gao, L. A Compact Methodology to Understand, Evaluate, and Predict the Performance of Automatic Target Recognition. Sensors 2014, 14, 11308-11350. https://doi.org/10.3390/s140711308

AMA Style

Li Y, Li X, Wang H, Chen Y, Zhuang Z, Cheng Y, Deng B, Wang L, Zeng Y, Gao L. A Compact Methodology to Understand, Evaluate, and Predict the Performance of Automatic Target Recognition. Sensors. 2014; 14(7):11308-11350. https://doi.org/10.3390/s140711308

Chicago/Turabian Style

Li, Yanpeng, Xiang Li, Hongqiang Wang, Yiping Chen, Zhaowen Zhuang, Yongqiang Cheng, Bin Deng, Liandong Wang, Yonghu Zeng, and Lei Gao. 2014. "A Compact Methodology to Understand, Evaluate, and Predict the Performance of Automatic Target Recognition" Sensors 14, no. 7: 11308-11350. https://doi.org/10.3390/s140711308

Article Metrics

Back to TopTop