Next Article in Journal
Editorial for the Special Issue on Industrial Machine Learning Applications
Previous Article in Journal
ViTSTR-Transducer: Cross-Attention-Free Vision Transformer Transducer for Scene Text Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Retrieval of Images with Irregular Patterns Using Morphological Image Analysis: Applications to Industrial and Healthcare Datasets

1
Department of Computer Science, School of Science, Loughborough University, Loughborough LE11 3TT, UK
2
Department of Physics, School of Science, Loughborough University, Loughborough LE11 3TT, UK
3
Railston & Co., Ltd., Nottingham NG7 2TU, UK
*
Authors to whom correspondence should be addressed.
J. Imaging 2023, 9(12), 277; https://doi.org/10.3390/jimaging9120277
Submission received: 10 October 2023 / Revised: 5 December 2023 / Accepted: 8 December 2023 / Published: 13 December 2023
(This article belongs to the Special Issue Advances and Challenges in Multimodal Machine Learning 2nd Edition)

Abstract

:
Image retrieval is the process of searching and retrieving images from a datastore based on their visual content and features. Recently, much attention has been directed towards the retrieval of irregular patterns within industrial or healthcare images by extracting features from the images, such as deep features, colour-based features, shape-based features, and local features. This has applications across a spectrum of industries, including fault inspection, disease diagnosis, and maintenance prediction. This paper proposes an image retrieval framework to search for images containing similar irregular patterns by extracting a set of morphological features (DefChars) from images. The datasets employed in this paper contain wind turbine blade images with defects, chest computerised tomography scans with COVID-19 infections, heatsink images with defects, and lake ice images. The proposed framework was evaluated with different feature extraction methods (DefChars, resized raw image, local binary pattern, and scale-invariant feature transforms) and distance metrics to determine the most efficient parameters in terms of retrieval performance across datasets. The retrieval results show that the proposed framework using the DefChars and the Manhattan distance metric achieves a mean average precision of 80% and a low standard deviation of ±0.09 across classes of irregular patterns, outperforming alternative feature–metric combinations across all datasets. Our proposed ImR framework performed better (by 8.71%) than Super Global, a state-of-the-art deep-learning-based image retrieval approach across all datasets.

1. Introduction

Image Retrieval (ImR) refers to the task of searching for and retrieving relevant images from a large collection or database of images. ImR systems can also enable the search and matching of images that contain similar irregular patterns [1]. ImR has attained extensive implementation for diverse tasks, including image search engines, image captioning, and image quality assessment. An irregular pattern is a region that differs in consistency or uniformity compared to surrounding areas or a typical example baseline. Irregular pattern retrieval and analysis have been used for the detection of industrial defects [2,3,4], chest infections in medical scans [5,6,7], and ice or snow on lakes [8,9], serving industry, healthcare, and environmental monitoring. An accurate ImR system can aid experts (e.g., manufacturing engineers, doctors, and quality inspectors) during decision-making.
Many research studies have explored the retrieval of images containing irregular patterns in industrial and healthcare datasets using different features and similarity metrics. Image-based similarity metrics (e.g., Mean Squared Error (MSE), Universal Image Quality Index (UIQ) [10], Spectral Angle Mapper (SAM) [11]) [12,13,14,15], which compare the similarities between image data, provide a simple and intuitive means of comparing two images in ImR tasks. However, the similarity values computed from these metrics are sensitive to image noise and quality. Feature extraction methods can extract the hidden features of irregular patterns within images and improve retrieval performance. These methods extract Local Binary Pattern (LBP) features [16,17,18,19], Scale-Invariant Feature Transform (SIFT) features [20,21,22], as well as colour and shape features [23,24] to conduct retrieval of images with irregular patterns. Distance-based similarity metrics (e.g., Manhattan, Jaccard, Euclidean, cosine, etc.) can be utilised to compute similarity values between two sets of features extracted from images with irregular patterns.
Feature extraction is pivotal in ImR, significantly influencing the performance of a retrieval system [25,26,27]. However, many existing feature extraction methods are dataset-specific in retrieving images containing irregular patterns, resulting in low retrieval performance for different datasets [25,28]. Zhang et al. [29] proposed a set of morphological features, known as Defect Characteristics (DefChars), to characterise images with irregular patterns in terms of colour, shape, and meta aspects. Zhang et al. [29,30] successfully utilised the DefChars to reason the outputs from different Artificial Intelligence (AI)-based defect detection and classification models across industrial and healthcare datasets. This paper extends the application of the DefChars proposed by Zhang et al. [29,30] to the task of ImR. In particular, the paper proposes a novel ImR framework that extracts DefChars from images that contain irregular patterns and retrieves images with similar irregular patterns by comparing their DefChars vectors using a feature-based similarity metric.
Four datasets are employed in this study: wind turbine blade images with defects [31], chest Computerised Tomography (CT) scans with COVID-19 infections [32], heatsink images with defects [33], and lake images with ice [34]. The proposed framework was evaluated with different feature–metric combinations, such as DefChar vectors and feature-based similarity metrics, resized raw images and image-based similarity metrics (MSE, SAM, UIQ), LBP and feature-based similarity metrics, SIFT and the Euclidean metric, and a state-of-the-art Deep Learning (DL)-based ImR approach (Super Global (SG)). The retrieval results demonstrate that using the combination of DefChars and the Manhattan metric within the proposed framework consistently achieves the highest mean Average Precision (mAP) (average 0.80) and also maintains the lowest Standard Deviation (std.) (average ± 0.09) across classes of irregular patterns and a fast retrieval time (average 0.14 s per query) across all datasets. Additionally, the retrieval results also indicate that using DefChars within the proposed framework achieves relatively high and balanced retrieval accuracy across classes despite dataset imbalances or small-sized datasets. The proposed ImR framework could be expanded to various industrial tasks in the future, including irregular pattern identification, classification, and deterioration monitoring.
The contributions of this paper are as follows.
  • A novel ImR framework that uses the morphological features to retrieve images containing irregular patterns.
  • Retrieval and time performance comparison of the proposed ImR framework using different features (i.e., the DefChars, LBP, SIFT, and resized images) across various datasets.
  • An empirical comparison of the retrieval and time performance between the proposed ImR framework and a state-of-the-art DL-based ImR approach.
There are six sections in this paper. Section 2 discusses related works concerning similarity metrics and feature extraction methods that can be employed in the ImR task. Section 3 presents the proposed ImR framework for retrieving irregular patterns in industrial and healthcare datasets. Section 4 describes the datasets, relevant feature extraction methods, similarity metrics, and the methodology used in this research, providing insight into the experimental setup. Section 5 evaluates and discusses the retrieval performance and execution time of the proposed ImR framework and compares it against a deep learning ImR framework. Section 6 summarises the key findings and conclusions drawn from the research conducted in this paper; moreover, some directions for future work are provided.

2. Related Works

2.1. Feature Extraction and Relevant Similarity Metrics for Retrieving Images

Recently, an increasing number of researchers have explored how effective feature extraction methods can enhance the performance of ImR. In 2019, Latif et al. [35] provided a comprehensive review of successful feature extraction methods used in Content-based Image Retrieval (CBIR) tasks. There are six major types of features that can be extracted from these methods, including colour-based features, shape-based features, texture-based features, spatial features, fusion features, and local features.
Colour-based features [36,37,38,39,40,41,42,43] offer fundamental visual information that is similar to human vision, and they are relatively robust against image transformations. Texture-based features [44,45,46,47,48,49,50] capture repeating patterns of local variance in image intensity; these features often hold more semantic meaning than colour-based features, though they can be susceptible to image noise. Shape-based features decode an object’s geometrical forms into machine-readable values and have been summarised into features that encompass contour, vertex angles, edges, polygons, spatial interrelation, moments, scale space, and shape transformation [35]. Spatial features [51,52,53,54,55,56,57] convey the location information of objects within the image space. Fusion features [58,59,60,61,62] combine basic features to form high-dimensional concatenated features—often, principal component analysis is applied to reduce dimensions. Local features [63,64,65,66,67,68,69,70] represent distinct structures and patches in an image, providing fine-grained details for ImR tasks.
Seetharaman and Sathiamoorthy [24] applied the Manhattan similarity metric along with colour-based and shape-based features to complete a medical ImR task. Their method achieved an average retrieval rate of 84.47% and a speed of 2.29 s. Petal et al. [23] extracted both colour-based and texture-based features from images and applied them to a CBIR task using various distance measures (e.g., Euclidean, cosine, Jaccard, Manhattan, etc.). Their approach achieved an accuracy of 87.2% in retrieving similar images. In 2022, Shamna et al. [71] employed the bag of visual words model as spatial features for retrieving medical images. Their method excelled in handling greyscale datasets, achieving a mAP of 69.70%; however, its performance on the colour datasets was lower than that of the greyscale datasets by 20.6%.

2.2. Recent Works to Retrieve Industrial and Healthcare Images with Irregular Patterns

In 2021, Boudani et al. [16] employed the wavelet-based LBP feature with the chi-square similarity metric to identify images containing surface defects on hot-rolled steel strips, achieving an mAP @ 10 score of 0.93; however, the performance was not stable between classes. Mo et al. [72] proposed a concentrated hashing method with neighbourhood embedding, utilising a Convolutional Neural Network (CNN) to extract hashing features for retrieving fabric and textile datasets in industrial applications. Their method outperformed other methods in four fabric datasets, with an average mAP of over 90%, but the precision dropped sharply, by 35%, when retrieving more than eight images. In 2022, Deep et al. [73] introduced a texture descriptor based on the concept of LBP for conducting ImR tasks on three biomedical datasets. The results showed that their proposed method reached an Average Precision (AP) rate of 91.5%; however, the AP rate of one of the datasets was very low (46%). Maintaining a consistent retrieval performance using the same feature and similarity metric for different datasets is challenging in the ImR domain. Boudani et al. [16] and Deep et al. [73] both applied LBP-based methods for ImR tasks, but their mAPs differed across their respective datasets.
Deep features, the high-level features extracted from the feature maps in DL models, have recently been used for ImR tasks. In 2022, Tan et al. [74] proposed a supervised ImR algorithm to retrieve integrated circuit images containing similar damage by matching deep features extracted from a pre-trained VGG16 model through dimension reduction and re-ranking methods. They conducted experiments to retrieve images for determining the integrated circuits’ levels (i.e., package and device) and damage classes within their levels. Their retrieval results reached a precision @ 5 of 97.3 when determining levels of the integrated circuits. However, the precision @ 5 dropped to 76.2% and 54.3% when determining the damage classes of package and device levels, respectively, resulting in an unbalanced retrieval performance of their proposed algorithms.
Agrawal et al. [5] proposed an ImR framework to retrieve chest X-ray images of lungs with COVID-19 infections. Their proposed framework extracted deep features from CNN models (i.e., VGG19 and ResNet50) and utilised distance-based metrics (i.e., chi-square, Euclidean, and cosine) to compute the similarity between images. Their retrieval performance achieved 50.4% in mAP across all classes using the ResNet50 model and cosine metric. Agrawal et al. [5] also highlighted that large-scale datasets and advanced CNN architectures can further improve retrieval performance. In 2023, Gassner et al. [75] proposed a saliency-enhanced CBIR algorithm to retrieve medical images containing skin lesions. Their proposed algorithm employed two sets of CNN classifiers to extract the deep features and saliency maps from images for the ImR tasks. The retrieval performance of their proposed algorithm improved by 0.13 points, from 0.69 to 0.82, when using saliency maps. However, the mAP of two lesion classes within the dataset was significantly lower than other classes, by 0.23 on average, due to the small number of images in these two classes.
In summary, despite the high retrieval performance achieved using deep features, some limitations still exist in conducting retrieval tasks [76]. For example, most DL-based ImR methods require training processes for extracting deep features and have expensive costs in computational resources (i.e., GPU and RAM), resulting in low efficiency when needed to update the database. In addition, there is a lack of research on DL-based ImR for industrial and healthcare datasets.

3. Proposed Image Retrieval Framework

This section introduces a proposed ImR framework, as shown in Figure 1, retrieving images with similar irregular patterns in industrial or healthcare datasets through the use of DefChars. The Python code for utilising our proposed ImR framework is provided in a GitHub repository (https://github.com/edgetrier/ImgRetreival-DefChars accessed on 12 December 2023). The proposed ImR framework consists of two main processes: the repository process and the retrieval process. The repository process employs two key modules, the DefChars extraction module and the indexing module, to construct a datastore. This datastore contains a DefChars matrix extracted from annotated images that include irregular patterns. The retrieval process, on the other hand, relies on three modules: the DefChars extraction module, the similarity computation module, and the ranking module. These modules are used to search for images with similar irregular patterns within the datastore by comparing the extracted DefChars vectors.

3.1. Repository Process

DefChars Extraction module. This module serves as a feature extraction component responsible for generating a DefChars matrix that extracts the colour-based, shape-based, and meta-based features of irregular patterns within images. The module takes as input raw images and extracts DefChars from these images. The module can also be modified to extract features using extraction methods other than DefChars, such as LBP or SIFT. The input to this module consists of a set of images and corresponding annotation matrices. Each annotation is represented as a mask-based matrix, outlining an irregular pattern’s region within an image. This matrix matches the size of the input image, with each value indicating whether the corresponding pixel in the input image falls inside or outside the irregular pattern’s region. To prepare the annotation matrices, image annotation tools like VIA [77], Labelme [78], or the drawContours function from the OpenCV package [79] can be employed if the dataset lacks annotation data. Subsequently, the module computes a DefChars vector (size: 38 × 1 ) for each image to represent the DefChars of the irregular pattern. This is achieved by analysing the pixel values within both the irregular pattern and background regions, as detailed in Section 4.2. The values in these vectors are then normalised to a range between 0 and 1. Finally, these DefChars vectors are aggregated into a DefChars matrix, which serves as the module’s output. The method for extracting DefChars is described in the paper by Zhang et al. [29]. Furthermore, Zhang et al. [30] also provided a toolkit for extracting the DefChars matrices from images.
Indexing module. This module associates each input image with its respective DefChars vector during the repository process. It takes the extracted DefChars matrix as input and assigns a unique index to each vector within the DefChars matrix. The index within a DefChars vector can assist users in locating the corresponding input image during the retrieval process. The module outputs an indexed DefChars matrix and stores it in a datastore. Additionally, this module can extend its indexing functionality when adding new images to the datastore.

3.2. Retrieval Process

DefChars Extraction module. This module is the same as the one described in Section 3.1. However, when utilised for retrieval, it extracts a single DefChars vector from an annotated query image. The input for this module consists of an image containing a single irregular pattern and an annotation matrix showing the region of the irregular pattern within the image. The annotation matrix for a query image can be collected by using an annotation tool. Moreover, preprocessing of the query image is not required in order to maintain consistency with the images in the datastore. The module generates a single DefChars vector that represents the features of the irregular pattern within the query image.
Similarity computation module. This module compares the DefChars vector extracted from the query image to each DefChars vector in the datastore using a feature-based similarity metric. In this paper, the Manhattan metric was employed as the feature-based similarity metric, although it can be substituted with any distance-based metric (e.g., cosine, Jaccard, Euclidean, etc.). The input to this module comprises the datastore (DefChars matrix) and the DefChars vector extracted from the query image. Subsequently, the module computes a set of similarity values using the selected metric to analyse the similarity between each DefChars vector in the datastore and the DefChars vector extracted from the query image. Finally, the set of similarity values is recorded and outputted in a similarity results table.
Ranking module. This module is the last step of the retrieval process, which ranks the retrieved results and presents the retrieved images. The module’s input is the similarity results table generated from the similarity computation module. Next, the similarity results table is ranked in order of the computed similarity values, and the module outputs the respective images according to their index.

4. Experiment Methodology

4.1. Datasets

Four datasets across industrial and healthcare domains were used to evaluate the retrieval performance of the proposed ImR framework: wind turbine blade defects [31], chest CT [32], heatsink defects [33] and Photi-LakeIce [34].
  • The wind turbine blade dataset, provided by our industrial partner Railston & Co., Ltd., Nottingham, United Kingdom, contains 191 images with 304 irregular patterns across four classes (crack, void, erosion, and other). The images of wind turbine blade defects were captured during inspection; mask annotations were gathered from Zhang et al.’s experiment [31].
  • The chest CT dataset was collected from Ter-Sarkisov’s [32] experiment and utilised to detect and classify the COVID-19 infection regions shown in chest CT scans. The chest CT dataset contains 750 images with 4665 irregular patterns across three classes (lung area, ground glass opacity, and consolidation), and the mask annotations were also provided in the dataset.
  • The heatsink dataset was collected from Yang et al.’s experiment [33] and used to detect defects on the surfaces of gold-plated tungsten–copper alloy heatsinks. The heatsink dataset contains 1000 images, captured by an industrial camera, with 7007 irregular patterns and corresponding mask annotations across two classes (scratch and stain).
  • The Photi-LakeIce (lake ice) dataset was collected from Prabha et al.’s [34] project and utilised to monitor the ice and snow on lakes by using AI techniques. The lake ice dataset contains 4017 images, captured by fixed-position webcams, with 5365 irregular patterns and corresponding mask annotations [34] across four classes (water, ice, snow, and clutter).
To conduct irregular pattern retrieval tasks on these datasets, each irregular pattern was cropped into an individual image based on the boundaries outlined in its respective mask annotation. The distribution of irregular pattern classes within each dataset is described in Table 1. Moreover, Figure 2 illustrates example cropped images for each class within each dataset.

4.2. Feature Extraction Method for Image Retrieval

Defect characteristics: This feature extraction methodology necessitates images with associated mask-based annotations, which are essential for calculating the DefChar values corresponding to each defect within the dataset. This is particularly important since a single image may encompass multiple defects. Table 2 provides a comprehensive list of the DefChar, their respective value ranges, and descriptions. The outcome of Zhang et al.’s method [29] manifests as a matrix with dimensions 38 × n , where n signifies the count of defects present within the dataset.
Scale-invariant feature transform: Lowe [80] introduced SIFT, a local feature extraction method that can maintain an image’s scale invariance. SIFT identifies a set of keypoints and descriptors, capturing distinctive points within images. This set of keypoints and descriptors can subsequently be employed to compute similarity between images by comparing their keypoints and descriptors using the Euclidean distance metric. The SIFT extraction method contains three key steps. The first step, scale-space extrema detection, utilises the Gaussian pyramid, images are progressively downsampled to identify a collection of potential keypoints. This is achieved by analysing the differences across each level of the Gaussian pyramid. The second step, orientation assignment, involves the elimination of low-contrast keypoints, enhancing the quality of the selected keypoints. The third step, keypoint descriptor, computes dominant orientations for individual keypoints, ensuring invariance to image rotation. Lowe [80] recommended using Euclidean similarity metric (explained in Section 4.3.2) to determine the similarity between the keypoints and descriptors of two images.
Local binary pattern: Ojala et al. [81] introduced a texture descriptor feature extraction method designed for ImR tasks. The LBP-based method extracts texture descriptors by comparing the intensity value of each pixel in an image to the intensity values of its neighbouring pixels. The LBP extraction method contains four sequential steps. The first step, neighbourhood definition, identifies the eight neighbouring pixels around each pixel within the image. The second step, binary comparison, calculates the binary intensity value of each neighbouring pixel relative to the centre pixel. The third step, binary pattern generation, concatenates all the binary intensity values into a singular vector, following either a clockwise or counter-clockwise order. The last step, decimal representation, converts the generated binary patterns into a decimal number, serving as a representation of the texture feature.

4.3. Similarity Metrics for Image Retrieval

There are two categories of similarity metrics to conduct an ImR task: image-based metrics for image data and feature-based metrics for extracted feature data. Image-based metrics compute the similarity or dissimilarity between images by directly comparing the image data. The MSE, SAM, and UIQ image-based metrics are used in this experiment. Feature-based metrics assess the similarity or dissimilarity between images by analysing the extracted features. The Euclidean, cosine, Jaccard, and Manhattan feature-based metrics are used in this experiment. In the descriptions that follow, let X be the query image; and let Y be one of the retrieved images.

4.3.1. Image-Based Similarity Metrics

Image similarity metrics play a crucial role in an ImR task by searching similar images within a database. Traditional image similarity metrics, such as MSE, SAM [11], UIQ [10], and Structural Similarity Index (SSIM) [13], allow for a direct comparison of pixel value differences between two images using mathematical equations.
Mean squared error (MSE) calculates the average squared difference in pixel values between two images. A higher MSE value signifies a greater dissimilarity between the two images.
MSE = 1 H W C i = 1 H j = 1 W k = 1 C ( x ( i , j , k ) y ( i , j , k ) ) 2
where H represents the height of the image; W represents the width of the image; C represents the number of channels (colour components) in each pixel; x ( i , j , k ) represents the pixel value of the ith row, jth column, and kth channel in the query image X; y ( i , j , k ) represents the pixel value of the ith row, jth column, and kth channel in the retrieving image Y.
Spectral angle mapper (SAM) calculates the angular disparity between two spectral signatures within a high-dimensional spectral space. A higher SAM value signifies a greater dissimilarity between the two images.
SAM = 1 C k = 1 C c o s 1 i = 1 H j = 1 W x ( i , j , k ) · y ( i , j , k ) i = 1 H j = 1 W x ( i , j , k ) 2 · i = 1 H j = 1 W y ( i , j , k ) 2
where C represents the number of channels (colour components) in each pixel; H represents the height of the image; W represents the width of the image; x ( i , j , k ) represents the pixel value of the ith row, jth column, and kth channel in the query image X; and y ( i , j , k ) represents the pixel value of the ith row, jth column, and kth channel in the retrieving image Y.
Universal image quality index (UIQ) takes into account the similarity between two images based on their correlation, luminance, and contrast. A higher UIQ value signifies a greater similarity between the two images. The maximum possible value of the UIQ is 1, indicating that the two images are exactly the same.
UIQ = 1 C k = 1 C σ x k y k σ x k σ y k · 2 x k ¯ y k ¯ ( x k ¯ ) 2 + ( y k ¯ ) 2 · 2 σ x k σ y k σ x k 2 + σ y k 2
where x k ¯ is the mean of the kth channel’s pixel values in the query image X; y k ¯ is the mean of the kth channel’s pixel values in the retrieving image Y; σ x k 2 is the variance of the kth channel’s pixel values in the query image X; σ y k 2 is the variance of the kth channel’s pixel values in the retrieving image Y; σ x k is the std. of the kth channel’s pixel values in the query image X; σ y k is the std. of the kth channel’s pixel values in the retrieving image Y; σ x k y k is the covariance of the kth channel’s pixel values between the query image X and the retrieving image Y.

4.3.2. Feature-Based Similarity Metrics

Euclidean distance calculates the direct straight-line distance between each point of two vectors. A higher value of the Euclidean distance signifies a greater dissimilarity between the two images.
Euclidean Distance = i = 1 N ( x i y i ) 2
where N represents the number of elements in the feature vector; x i represents the ith value of the extracted feature vector from the query image X; and y i represents the ith value of the extracted feature vector from the retrieving image Y.
Cosine distance calculates the cosine of the angles between two vectors. A higher value of the cosine distance indicates that the two images are more similar.
Cosine Distance = i = 1 N x i · y i i = 1 N x i 2 · i = 1 N y i 2
where N represents the number of elements in the feature vector; x i represents the ith value of the extracted feature vector from image X; and y i represents the ith value of the extracted feature vector from image Y.
Manhattan distance calculates the sum of absolute differences between corresponding elements of two vectors. A larger value of the Manhattan distance signifies that the two images are more dissimilar.
Manhattan Distance = i = 1 N | x i y i |
where N represents the number of elements in the feature vector; x i represents the ith value of the extracted feature vector from the query image X; and y i represents the ith value of the extracted feature vector from the retrieving image Y.
Jaccard distance calculates the ratio of the common elements between two feature vectors to the total number of elements present in the vectors. A higher value of the Jaccard distance suggests that the two images are more similar in terms of the shared features or values.
Jaccard Distance = x y x y
where x y represents the intersection of the sets of elements present in the feature vectors of the query image X and retrieving image Y, and x y represents the union of the sets of elements present in the feature vectors of the query image X and retrieving image Y.

4.4. Methodology

The proposed ImR framework (as described in Section 3) was utilised to conduct experiments for evaluating its performance when using different feature sets and similarity metrics across four different datasets. The experiments were conducted on a high-performance computer with an AMD Ryzen 9 CPU and 32 GB RAM. The utilisation of a graphics processing unit (GPU) was not necessary for the experiments.
  • In the initial step (DefChar extraction module) of the experiment, feature sets are extracted:
    Feature set 1 contains the raw images that have been compressed into four different sizes (i.e., 100 × 100 , 50 × 50 , 20 × 20 , 8 × 8 ) with the dual goals of normalisation and acceleration of the retrieval process. Hence, feature set 1 has four subsets of features each corresponding to a different size.
    Feature set 2 contains DefChars extracted from raw images. Raw images should be utilised when extracting DefChars, due to information loss caused by image resizing.
    Feature sets 3 and 4 contain LBP and SIFT features extracted from each of the feature subsets described in feature set 1. The parameters of LBP were set to radius = 1, sample points = 8, and method = uniform, following the recommendations by Rahillda et al. [82] based on their experimental results. The SIFT parameters used in this experiment were set according to the guidelines by Lowe [80]: nFeatures = max, nOctaveLayers = 3, contrastThreshold = 0.3, edgeThreshold = 10, and sigma = 1.6. The feature sets are separately stored (indexing module) in a datastore.
  • In the next step (similarity computation module), the similarity between a query image and images—represented as feature vectors—found in the datastore is computed. In this experiment, each image from the datasets is iteratively selected to be a query. To compute the similarity between images, feature- and image-based similarity metrics are applied. Feature-based similarity metrics (i.e., Euclidean, cosine, Manhattan, and Jaccard) are utilised for DefChar, SIFT, and LBP features; and image-based similarity metrics (i.e., MSE, SAM, and UIQ) are utilised for compressed raw images.
  • Then, the retrieved irregular patterns are ranked (ranking module) according to the computed similarity values. The metrics described in Section 4.5 are applied to evaluate the retrieval performance.

4.5. Evaluation Measures

This section introduces the evaluation measures employed within the scope of an ImR task carried out in this experiment. In the context of an ImR task, the primary objective revolves around searching for relevant images with similar irregular patterns within a datastore. A relevant image to an image query is one that belongs to the same class as the query image. Precision @ K is the ratio of relevant images with irregular patterns correctly retrieved among the top K retrieved images with irregular patterns, determining the retrieval performance for a single query. AP @ K is the average value of Precision @ K in all queries that are retrieved for one irregular pattern class within a dataset, determining the retrieval performance for an irregular pattern class in a dataset. Subsequently, mAP @ K is the average value of AP @ K across all irregular pattern classes within a dataset, determining the retrieval performance for a dataset. Furthermore, the std. values for AP and mAP were computed to evaluate the variations in the retrieval performance in classes and datasets, respectively.

5. Results and Discussion

This section discusses the retrieval and speed performance of the proposed ImR framework shown in Figure 1 (see Section 3) using different feature sets and similarity metrics. The experimental methodology is discussed in Section 4.4. The retrieval results when using different settings are presented in Appendix A, Appendix B, Appendix C, Appendix D. Furthermore, this section includes a comparison between the proposed ImR framework with a DL-based ImR approach.
In the rest of the paper, the following descriptions apply:
  • DefChar-based methods refer to the proposed ImR framework using DefChars within the DefChar extraction module.
  • Image-based methods refer to the proposed ImR framework using resized images instead of the DefChar extraction module.
  • LBP-based methods refer to the proposed ImR framework using LBP features instead of the DefChar extraction module.
  • SIFT-based methods refer to the proposed ImR framework using SIFT features instead of the DefChar extraction module.

5.1. Image Retrieval Performance When Using Different Feature Sets and Similarity Metrics across Datasets

5.1.1. Chest CT Dataset

Table A1 shows the mAPs along with std. values when using different feature sets and similarity metrics in the chest CT dataset; Table A2, Table A3, Table A4 show the APs with std. values for each class within the dataset.
ImR performance when using the proposed DefChar-based methods. There were two similarity metrics (i.e., cosine and Euclidean) that yielded the highest mAP and the lowest std., averaging 0.85 ± 0.06. In terms of retrieval performance for each class using the cosine or Euclidean metric, both metrics had the same mAP values and std. values for class 1. For class 2, the Euclidean metric achieved a slightly higher AP of 0.01 compared to the cosine metric at @ 1 and @ 20 . Conversely, the cosine metric outperformed the Euclidean metric by 0.01 in AP at @ 1 and @ 10 for class 3. Additionally, the Manhattan metric showed noteworthy retrieval performance for the chest CT dataset and relatively achieved the highest AP for classes 1 and 2. However, for class 3, the retrieval performance of the Manhattan metric was slightly lower than those of the cosine and Euclidean metrics, with an average difference of 0.01–0.02 at @ 5 , @ 10 , and @ 15 . This resulted in a higher std. in mAP, despite the metrics having the same mean mAP values.
ImR performance when using image-based methods. The best-performing image-based methods were the UIQ metric with a 20 × 20 image size, averaging 0.76 ± 0.17 in terms of mAP. The UIQ metrics with 50 × 50 and 100 × 100 image sizes showed similar mAP values across the range from @ 1 to @ 20 , but they had higher std. values compared to the UIQ metric with a 20 × 20 image size. The UIQ metric with a 20 × 20 image size showed a relatively small difference between the maximum and minimum AP in different classes (i.e., @ 1 : 0.94– @ 20 : 0.89 for class 1 and @ 1 : 0.59– @ 20 : 0.58 for class 3), although it may not achieve the highest AP.
ImR performance when using LBP-based methods. The performance of the LBP-based methods showed a correlation with the image size. The highest mAP was achieved, averaging 0.30 ± 0.22 between @ 1 and @ 20 , when using images of size 100 × 100 . However, it is worth noting that the retrieval performance of the best-performing LBP-based method for each class was not consistent.
ImR performance when using SIFT-based methods. The 20 × 20 image size proved to be the optimal setting for the SIFT-based method, with an average mAP of 0.52 ± 0.31. The std. of the SIFT-based methods were the highest among all methods; consequently, the SIFT-based methods struggled to maintain consistent retrieval performance across all classes within the chest CT dataset.

5.1.2. Heatsink Dataset

Table A5 shows the mAPs, along with std., when using different feature sets and similarity metrics in the heatsink dataset; Table A6 and Table A7 show the APs with std. for each class within the dataset.
ImR performance when using the proposed DefChar-based methods. All similarity metrics consistently achieved the highest mAP with an average of 0.97 ± 0.02, except for the Jaccard metric. When analysing the performance for each class, the Manhattan metric sometimes reached a slightly higher std., by ±0.01, compared to the cosine and Euclidean metrics. Additionally, the AP for the Manhattan metric was sometimes lower by 0.01. As a result, the performance of the Manhattan metric was slightly lower than that of the others, particularly at @ 10 and @ 15 .
ImR performance when using image-based methods. The average mAP between @ 1 and @ 20 reached its peak at 0.88 ± 0.08 when using 100 × 100 images with the UIQ metric, although its mAP @ 1 was slightly lower, at 0.86, compared to others. Additionally, the MSE with an 8 × 8 image size, demonstrated relatively high performance, with an average mAP of 0.87. When considering the performance for each class, the UIQ metric with a 100 × 100 image size and the MSE metric with an 8 × 8 image size consistently maintained a high AP across all classes. In contrast, other methods showed fluctuations in AP when applied to different classes.
ImR performance when using LBP-based methods. The LBP-based methods reached consistent mAP and std. values across all feature-based similarity metrics. The best-performing LBP-based methods achieved an mAP of 0.53 ± 0.18 on average when utilising 8 × 8 images. However, it is worth noting that the performance of the LBP-based methods varied significantly for each class. The LBP-based method with 8 × 8 images significantly outperformed the others, by more than 0.43 in AP for class 2; but, the performance of this metric fell behind the others for class 1, particularly when compared to the LBP-based method with a 20 × 20 image size.
ImR performance when using SIFT-based methods. The best retrieval performance was 0.54 ± 0.23 in terms of mAP when applying the 100 × 100 images to the SIFT-based methods. The performance of the best-performing SIFT-based methods across classes was not balanced; for instance, the AP ranged from @ 5 : 0.51 to @ 20 : 0.45 for class 1 and @ 5 : 0.59 to @ 20 : 0.64 for class 2.

5.1.3. Lake Ice Dataset

Table A8 shows the mAPs, along with std., when using different feature sets and similarity metrics in the lake ice dataset; Table A9, Table A10, Table A11, Table A12 show the APs with std. for each class within the dataset.
ImR performance when using the proposed DefChar-based methods. The highest mAP among the DefChar-based methods had an average of 0.90 ± 0.07 when using the Manhattan metric. Also, the retrieval performance was relatively balanced across all classes within the dataset; the AP for all classes exceeded 0.94 at @ 1 and 0.74 at @ 20 , which was higher than for other DefChar-based methods.
ImR performance when using image-based methods. All image-based methods reached similar performance; however, the SAM using 8 × 8 images outperformed the others with the highest mAP (0.86) and the lowest std. (±0.13). In terms of performance across different classes, there was no significant difference observed between all similarity metrics and image sizes. Nevertheless, the SAM with an 8 × 8 image size achieved a higher AP than other image-based methods by 0.04–0.06 for class 4, resulting in a higher mAP with a lower std. value.
ImR performance when using LBP-based methods. All LBP-based methods reached relatively low mAP values across all classes. The best-performing LBP-based method, when using any feature-based similarity metric with 100 × 100 images, only reached an average mAP of 0.26 ± 0.37. Furthermore, this best-performing LBP-based method achieved high performance for class 3 (from mAP @ 1 : 0.99 to mAP @ 20 : 0.68); for other classes, however, the mAP dropped significantly, falling below 0.23 and even reaching 0.00.
ImR performance when using SIFT-based methods. In all SIFT-based methods, using large-sized images (i.e., 100 × 100 ) achieved the highest mAP, averaging 0.64 ± 0.15. The performance of the SIFT-based method with a 100 × 100 image size was relatively balanced between all classes, except for class 4, where the AP was 40% lower than those for other classes.

5.1.4. Wind Turbine Blade Dataset

Table A13 shows the mAPs, along with std., when using different feature sets and similarity metrics in the wind turbine blade dataset; Table A14, Table A15, Table A16, Table A17 show the APs with std. for each class within the dataset.
ImR performance when using the proposed DefChar-based methods. In the DefChar-based methods, the Manhattan metric outperformed others with the highest mAP (0.62) and the lowest std. (±0.17). When considering the performance for each class, the retrieval performance using the Manhattan metric generally exceeded other metrics, except for AP @ 15 and AP @ 20 in class 2, and AP @ 5 , AP @ 10 , and AP @ 15 in class 4.
ImR performance when using image-based methods. In the image-based methods, two settings (i.e., MSE with an 8 × 8 image size and UIQ with a 20 × 20 image size) both achieved the highest mAP (0.44) and the lowest std. (±0.31). When evaluating the performance for each class using these best-performing image-based methods, they reached similar AP values across all classes. However, the AP of the MSE metric occasionally exceeded that of the UIQ by 0.03–0.08 at @ 1 for all classes except class 3.
ImR performance when using LBP-based methods. The feature-based similarity metric with a small-sized image (i.e., 8 × 8 ) outperformed other LBP-based methods and achieved an average mAP of 0.26 ± 0.17. However, all LBP-based methods, including the best-performing one, struggled to maintain a consistent performance across all classes. For instance, the AP of the best-performing method was lower than other LBP-based methods between @ 1 and @ 20 for classes 1 and 3.
ImR performance when using SIFT-based methods. The highest mAP was achieved at 0.35 ± 0.30 when utilising the Euclidean metric with a 100 × 100 image size in the SIFT-based method. When evaluating the performance differences of the SIFT-based methods across each class, the best-performing method was relatively more accurate than others for class 3 by over 0.10, although it did not outperform the others for the rest of the classes.

5.2. Overall Performance Comparisons for Image Retrieval Tasks

This section compares the retrieval performance of different feature extraction methods, utilising the settings that demonstrated the best results for each dataset. These settings were selected based on the analysis presented in Section 5.1. Additionally, this section delves into investigating the time completed for each method and dataset in the ImR task.
Comparison of retrieval performance across datasets: Figure 3 presents a line chart illustrating the mAP of the best-performing ImR methods, that were explained earlier, across all datasets. The DefChar-based methods consistently outperformed other methods across all datasets. When using DefChars, the performance of different similarity metrics was not significantly different; however, the Manhattan metric reached the highest mAP across all datasets. Image-based ImR methods were the second best performing across all datasets. However, the choice of similarity metrics and image sizes emerged as significant factors influencing performance for different datasets, and there is no image-based methods with a consistent setting that relatively maintains high retrieval performance. For instance, the UIQ metric performed relatively better with large-sized images in the chest CT and heatsink datasets, while the SAM and MSE metrics achieved higher mAP in the lake ice and wind turbine blade datasets when utilising small-sized images. The performance of LBP-based methods were the worst and did not show significant differences based on the choice of similarity metric. Furthermore, image sizes can easily affect the retrieval performance of LBP-based methods on different datasets. For instance, the chest CT and lake ice datasets reached higher mAP values with larger-sized images, whereas the heatsink and wind turbine blade datasets yielded better results with smaller-sized images. SIFT-based methods showed a better retrieval performance compared to the LBP-based methods. Using small-sized images (e.g., 20 × 20 ) within the SIFT-based methods reached higher retrieval performance than using large-sized images across all but the chest CT dataset.
Comparison of retrieval performance stability across classes and datasets: Figure 4 provides insight into the std. values encountered when calculating the mAP using the best performing methods that were explained in Section 5.1. DefChar-based methods overall achieved the lowest std. across all datasets. This indicates their relatively stable and reliable performance. In the wind turbine blade dataset, the std. of LBP-based methods was ±0.03–0.05 lower than that of the DefChar-based methods on average. However, the mAP achieved by LBP-based methods were the lowest among all methods. In contrast, the rest of the methods, such as image-based, SIFT-based, and LBP-based methods, reached higher std. values, often exceeding those of DefChar-based methods by more than ±0.1 across datasets. Moreover, these methods had varying std. values across different datasets. This implies that these methods might show less consistency when retrieving irregular patterns from different classes within a dataset. As noted in Table 1, the datasets are relatively imbalanced between each class. However, the ImR method utilising DefChars and the Manhattan metric showed the capability to maintain relatively high and balanced accuracy in retrieving similar irregular patterns. Additionally, the performance of DefChar-based methods did not show significant deterioration when applied to a small dataset (i.e., wind turbine blades).
Comparison of time performance across datasets: Figure 5 presents the time required to extract features and retrieve images with similar irregular patterns for each query across four datasets. Among these methods, LBP-based methods achieved the shortest total time, taking less than 0.062 s per query. On the other hand, UIQ consumed the most time, especially when using 100 × 100 images, resulting in query times exceeding 5 s for large datasets such as chest CT, heatsink, and lake ice. In the DefChar-based methods, the retrieval time ranged from 0.06 to 0.26 s, making it one of the faster approaches, except for in the wind turbine blade dataset. However, while the feature extraction times for DefChar-based methods were comparatively longer than other methods, their retrieval times were significantly shorter than those of other methods. The image-based approaches generally required more time for retrieval, although image resizing reduced the feature extraction time. The SIFT-based methods needed an additional 0.0003 to 0.0018 s for extracting the SIFT features after image resizing, but the retrieval time was relatively shorter than most image-based methods. When considering the time impact of the dataset or image size, a large dataset or image size typically results in a longer execution time. For instance, the ImR time for UIQ metrics when using 100 × 100 images was over three times longer than when using 20 × 20 images. On the other hand, the heatsink dataset, the largest dataset containing 7007 irregular patterns (see Table 1), used longer retrieval times across all datasets, because the query image needed to be compared to all images in the datastore. However, DefChar-based methods outperformed all other methods in terms of retrieval performance.

5.3. Image Retrieval Performance Comparisons between the Proposed ImR Framework and a DL-Based ImR Framework

This section empirically evaluates the retrieval performance of the proposed ImR framework when using its optimal settings (i.e., DefChars and Manhattan metric), compared to a state-of-the-art DL-based ImR approach (i.e., SG). In 2023, Shao et al. [83] proposed a supervised DL-based ImR retrieval framework, called SG, to conduct retrieval tasks for large-scale Google landmark image datasets: the revisited Oxford5k and Paris6k [84]. Their framework trained a ResNet101 model to extract deep features from images and employed a re-ranking method to search for similar images in the database. Their proposed framework achieved a mAP of 82.9% and outperformed other state-of-the-art DL-based ImR frameworks (i.e., CVNet [85], DOLG [86] and DELG [87]) in terms of retrieval and speed performance.
Motivated by these results, this paper evaluates the retrieval and speed performance of the ImR framework proposed by Shao et al. [83] using the four datasets described in Section 4.1. The pre-trained ResNet101 model from Shao et al. [83] is directly used to extract deep features. Then, these deep features are used to retrieve images containing similar irregular patterns using their proposed re-ranking method.
The retrieval and speed performance between SG and our proposed ImR framework for each dataset are shown in Table 3 and Table 4. SG performed slightly better than ours on the wind turbine blade and lake ice datasets, by 0.01–0.07 mAP between @ 1 and @ 20 . However, on the chest CT and heatsink datasets, the retrieval performance of SG was substantially lower than ours, by 0.14–0.19, in mAP. Furthermore, the std. of SG was also higher than ours across all except the lake ice dataset, demonstrating that SG was not able to achieve a high retrieval performance for all classes across datasets. Regarding speed performance, SG was significantly faster than ours, achieving 0.029–0.041 s per query across all datasets. The high retrieval speed achieved by SG can be attributed to its utilisation of a GPU to accelerate the feature extraction and retrieval processes. However, given that our proposed ImR framework runs on a CPU, the retrieval performance comparison between these two approaches in terms of speed is not directly comparable. Further benchmarking would be required with both methods running on the same hardware for an equal evaluation.
Overall, while SG has a faster retrieval speed and performs marginally better on two datasets, our proposed ImR framework achieves more robust retrieval performance across all datasets. Furthermore, our proposed ImR framework is unsupervised and does not need dataset-specific training. This highlights its capability to retrieve images with irregular patterns on different industrial and healthcare datasets without the need for customised feature learning.

6. Conclusions and Future Work

This paper proposed an ImR framework to retrieve images with irregular patterns and completed a comprehensive evaluation of retrieval performance using different feature extraction methods (i.e., DefChars, resized raw images, LBP, and SIFT) along with different similarity metrics across four datasets (chest CT, heatsink, lake ice, and wind turbine blade). The findings highlighted that the ImR framework, utilising DefChars and the Manhattan similarity metric, consistently demonstrated robust retrieval and speed performance across all datasets. Moreover, the proposed framework did not show a significant bias towards each class within the dataset, despite minor fluctuations in AP values possibly attributed to dataset complexity. Furthermore, the proposed ImR framework outperformed the SG, a DL-based approach, in retrieval performance across datasets.
Future work could involve comparing other DL-based ImR approaches; investigating retrieval performance when assigning weights to DefChars during the feature extraction; providing explanations for the retrieved results; and utilising object detection or segmentation techniques to automatically complete annotations of the irregular patterns for the ImR task.

Author Contributions

Conceptualisation, J.Z. and G.C.; methodology, J.Z.; software, J.Z.; validation, J.Z., S.B. and J.W.; formal analysis, J.Z.; investigation, J.Z.; resources, J.Z. and G.C.; data curation, J.Z. and J.W.; writing—original draft preparation, J.Z.; writing—review and editing, J.Z., G.C. and S.B.; visualisation, J.Z.; supervision, G.C., S.B. and J.W.; project administration, J.Z. and G.C.; funding acquisition, G.C. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded through joint funding by the School of Science at Loughborough University with industrial support from Railston & Co., Ltd.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Chest CT COVID-19 Dataset: Publicly available datasets were analysed in this study. These data can be found here: http://ncov-ai.big.ac.cn/download, accessed on 30 September 2023. Heatsink Defect Dataset: Publicly available datasets were analysed in this study. These data can be found here: https://www.kaggle.com/datasets/kaifengyang/heat-sink-surface-defect-dataset, accessed on 30 September 2023. Lake Ice Dataset: Publicly available datasets were analysed in this study. These data can be found here: https://github.com/prs-eth/photi-lakeice-dataset/, accessed on 30 September 2023. Wind Turbine Blade Defect Dataset: Not applicable.

Acknowledgments

The authors would like to acknowledge Jason Watkins, Chris Gibson, and Andrew Rattray from Railston & Co., Ltd., for providing the datasets and expert knowledge required during the project.

Conflicts of Interest

Author J.W. was employed by the company Railston & Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ImRImage Retrieval
AIArtificial Intelligence
CBIRContent-based Image Retrieval
CNNConvolutional Neural Network
DLDeep Learning
mAP Mean Average Precision
AP Average Precision
LBPLocal Binary Pattern
DefCharsDefect Characteristics
SIFTScale-Invariant Feature Transform
MSEMean Squared Error
SAMSpectral Angle Mapper
UIQUniversal Image Quality Index
SSIMStructural Similarity Index
CTComputerised Tomography
SGSuper Global
StdStandard Deviation

Appendix A. ImR Evaluation Results for the Chest CT Dataset

Table A1. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
Table A1. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size mAP @ 1 mAP @ 5 mAP @ 10 mAP @ 15 mAP @ 20 Average
DefCharsCosineRaw0.88 ± 0.060.86 ± 0.060.85 ± 0.060.84 ± 0.060.84 ± 0.060.85 ± 0.06
DefCharsEuclideanRaw0.88 ± 0.060.86 ± 0.060.85 ± 0.060.84 ± 0.060.84 ± 0.060.85 ± 0.06
DefCharsJaccardRaw0.51 ± 0.190.49 ± 0.200.47 ± 0.210.47 ± 0.210.47 ± 0.220.48 ± 0.21
DefCharsManhattanRaw0.88 ± 0.060.86 ± 0.070.85 ± 0.070.84 ± 0.070.84 ± 0.070.85 ± 0.07
ImageMSE80.76 ± 0.220.74 ± 0.220.73 ± 0.220.72 ± 0.220.71 ± 0.220.73 ± 0.22
ImageMSE200.75 ± 0.240.74 ± 0.230.73 ± 0.230.72 ± 0.230.71 ± 0.230.73 ± 0.23
ImageMSE500.75 ± 0.230.73 ± 0.230.73 ± 0.230.72 ± 0.230.71 ± 0.230.73 ± 0.23
ImageMSE1000.75 ± 0.230.73 ± 0.230.72 ± 0.230.72 ± 0.230.71 ± 0.230.73 ± 0.23
ImageSAM80.72 ± 0.200.69 ± 0.220.67 ± 0.220.66 ± 0.230.65 ± 0.230.68 ± 0.22
ImageSAM200.72 ± 0.210.69 ± 0.220.67 ± 0.230.66 ± 0.240.65 ± 0.240.68 ± 0.23
ImageSAM500.72 ± 0.210.69 ± 0.230.67 ± 0.240.66 ± 0.240.65 ± 0.250.68 ± 0.23
ImageSAM1000.72 ± 0.210.69 ± 0.220.67 ± 0.240.66 ± 0.240.65 ± 0.250.68 ± 0.23
ImageUIQ80.33 ± 0.580.33 ± 0.420.33 ± 0.420.33 ± 0.310.33 ± 0.230.33 ± 0.39
ImageUIQ200.77 ± 0.180.76 ± 0.170.75 ± 0.170.75 ± 0.160.75 ± 0.160.76 ± 0.17
ImageUIQ500.77 ± 0.200.76 ± 0.190.75 ± 0.190.75 ± 0.190.74 ± 0.190.75 ± 0.19
ImageUIQ1000.77 ± 0.210.76 ± 0.200.75 ± 0.210.74 ± 0.210.74 ± 0.210.75 ± 0.21
LBPCosine80.22 ± 0.160.22 ± 0.170.22 ± 0.180.22 ± 0.180.23 ± 0.190.22 ± 0.18
LBPCosine200.23 ± 0.280.21 ± 0.270.21 ± 0.280.21 ± 0.280.21 ± 0.280.21 ± 0.28
LBPCosine500.28 ± 0.280.29 ± 0.110.28 ± 0.120.28 ± 0.080.28 ± 0.080.28 ± 0.13
LBPCosine1000.30 ± 0.240.30 ± 0.280.30 ± 0.220.30 ± 0.190.31 ± 0.160.30 ± 0.22
LBPEuclidean80.22 ± 0.160.22 ± 0.170.22 ± 0.180.22 ± 0.180.23 ± 0.190.22 ± 0.18
LBPEuclidean200.23 ± 0.280.21 ± 0.270.21 ± 0.280.21 ± 0.280.21 ± 0.280.21 ± 0.28
LBPEuclidean500.28 ± 0.280.29 ± 0.110.28 ± 0.120.28 ± 0.080.28 ± 0.080.28 ± 0.13
LBPEuclidean1000.30 ± 0.240.30 ± 0.280.30 ± 0.220.30 ± 0.190.31 ± 0.160.30 ± 0.22
LBPJaccard80.22 ± 0.160.22 ± 0.170.22 ± 0.180.22 ± 0.180.23 ± 0.190.22 ± 0.18
LBPJaccard200.23 ± 0.280.21 ± 0.270.21 ± 0.280.21 ± 0.280.21 ± 0.280.21 ± 0.28
LBPJaccard500.28 ± 0.280.29 ± 0.110.28 ± 0.120.28 ± 0.080.28 ± 0.080.28 ± 0.13
LBPJaccard1000.30 ± 0.240.30 ± 0.280.30 ± 0.220.30 ± 0.190.31 ± 0.160.30 ± 0.22
LBPManhattan80.22 ± 0.160.22 ± 0.170.22 ± 0.180.22 ± 0.180.23 ± 0.190.22 ± 0.18
LBPManhattan200.23 ± 0.280.21 ± 0.270.21 ± 0.280.21 ± 0.280.21 ± 0.280.21 ± 0.28
LBPManhattan500.28 ± 0.280.29 ± 0.110.28 ± 0.120.28 ± 0.080.28 ± 0.080.28 ± 0.13
LBPManhattan1000.30 ± 0.240.30 ± 0.280.30 ± 0.220.30 ± 0.190.31 ± 0.160.30 ± 0.22
SIFTEuclidean80.40 ± 0.510.35 ± 0.410.35 ± 0.420.36 ± 0.330.36 ± 0.260.36 ± 0.39
SIFTEuclidean200.56 ± 0.290.52 ± 0.320.51 ± 0.330.51 ± 0.310.51 ± 0.300.52 ± 0.31
SIFTEuclidean500.43 ± 0.400.44 ± 0.390.45 ± 0.390.45 ± 0.390.44 ± 0.380.44 ± 0.39
SIFTEuclidean1000.38 ± 0.390.40 ± 0.350.42 ± 0.340.43 ± 0.340.43 ± 0.330.41 ± 0.35
Table A2. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
Table A2. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.93 ± 0.250.92 ± 0.210.91 ± 0.210.90 ± 0.210.90 ± 0.21
DefCharsEuclideanRaw0.93 ± 0.250.92 ± 0.210.91 ± 0.210.90 ± 0.210.90 ± 0.21
DefCharsJaccardRaw0.70 ± 0.460.69 ± 0.330.68 ± 0.300.68 ± 0.290.68 ± 0.28
DefCharsManhattanRaw0.94 ± 0.240.92 ± 0.210.92 ± 0.200.91 ± 0.200.91 ± 0.20
ImageMSE80.93 ± 0.250.92 ± 0.200.91 ± 0.200.90 ± 0.200.89 ± 0.20
ImageMSE200.94 ± 0.240.93 ± 0.190.92 ± 0.190.91 ± 0.190.91 ± 0.19
ImageMSE500.94 ± 0.230.93 ± 0.180.92 ± 0.180.92 ± 0.190.91 ± 0.19
ImageMSE1000.94 ± 0.240.93 ± 0.180.92 ± 0.180.92 ± 0.190.91 ± 0.19
ImageSAM80.94 ± 0.230.94 ± 0.170.93 ± 0.160.92 ± 0.160.92 ± 0.16
ImageSAM200.96 ± 0.200.95 ± 0.150.94 ± 0.150.93 ± 0.150.93 ± 0.15
ImageSAM500.96 ± 0.200.95 ± 0.150.94 ± 0.150.94 ± 0.140.93 ± 0.15
ImageSAM1000.96 ± 0.200.95 ± 0.150.94 ± 0.150.94 ± 0.150.93 ± 0.15
ImageUIQ80.00 ± 0.000.80 ± 0.000.80 ± 0.000.67 ± 0.000.55 ± 0.00
ImageUIQ200.94 ± 0.230.92 ± 0.190.91 ± 0.190.90 ± 0.190.89 ± 0.20
ImageUIQ500.95 ± 0.230.93 ± 0.180.92 ± 0.180.91 ± 0.180.91 ± 0.19
ImageUIQ1000.95 ± 0.220.94 ± 0.170.93 ± 0.170.92 ± 0.180.91 ± 0.18
LBPCosine80.40 ± 0.490.41 ± 0.390.43 ± 0.360.43 ± 0.350.43 ± 0.34
LBPCosine200.55 ± 0.500.53 ± 0.380.53 ± 0.360.53 ± 0.350.54 ± 0.34
LBPCosine500.04 ± 0.200.28 ± 0.200.27 ± 0.200.26 ± 0.200.27 ± 0.20
LBPCosine1000.03 ± 0.160.16 ± 0.170.23 ± 0.120.26 ± 0.120.27 ± 0.12
LBPEuclidean80.40 ± 0.490.41 ± 0.390.43 ± 0.360.43 ± 0.350.43 ± 0.34
LBPEuclidean200.55 ± 0.500.53 ± 0.380.53 ± 0.360.53 ± 0.350.54 ± 0.34
LBPEuclidean500.04 ± 0.200.28 ± 0.200.27 ± 0.200.26 ± 0.200.27 ± 0.20
LBPEuclidean1000.03 ± 0.160.16 ± 0.170.23 ± 0.120.26 ± 0.120.27 ± 0.12
LBPJaccard80.40 ± 0.490.41 ± 0.390.43 ± 0.360.43 ± 0.350.43 ± 0.34
LBPJaccard200.55 ± 0.500.53 ± 0.380.53 ± 0.360.53 ± 0.350.54 ± 0.34
LBPJaccard500.04 ± 0.200.28 ± 0.200.27 ± 0.200.26 ± 0.200.27 ± 0.20
LBPJaccard1000.03 ± 0.160.16 ± 0.170.23 ± 0.120.26 ± 0.120.27 ± 0.12
LBPManhattan80.40 ± 0.490.41 ± 0.390.43 ± 0.360.43 ± 0.350.43 ± 0.34
LBPManhattan200.55 ± 0.500.53 ± 0.380.53 ± 0.360.53 ± 0.350.54 ± 0.34
LBPManhattan500.04 ± 0.200.28 ± 0.200.27 ± 0.200.26 ± 0.200.27 ± 0.20
LBPManhattan1000.03 ± 0.160.16 ± 0.170.23 ± 0.120.26 ± 0.120.27 ± 0.12
SIFTEuclidean80.19 ± 0.400.82 ± 0.100.82 ± 0.090.72 ± 0.120.63 ± 0.16
SIFTEuclidean200.77 ± 0.420.86 ± 0.200.86 ± 0.180.84 ± 0.180.82 ± 0.19
SIFTEuclidean500.88 ± 0.320.89 ± 0.170.89 ± 0.150.88 ± 0.150.88 ± 0.15
SIFTEuclidean1000.83 ± 0.380.81 ± 0.210.80 ± 0.170.81 ± 0.160.81 ± 0.16
Table A3. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
Table A3. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.88 ± 0.320.86 ± 0.240.85 ± 0.230.84 ± 0.230.83 ± 0.23
DefCharsEuclideanRaw0.89 ± 0.310.86 ± 0.240.85 ± 0.220.84 ± 0.220.84 ± 0.22
DefCharsJaccardRaw0.50 ± 0.500.48 ± 0.250.47 ± 0.190.47 ± 0.170.47 ± 0.15
DefCharsManhattanRaw0.88 ± 0.330.86 ± 0.230.85 ± 0.220.85 ± 0.220.84 ± 0.21
ImageMSE80.82 ± 0.380.80 ± 0.280.79 ± 0.260.78 ± 0.250.78 ± 0.25
ImageMSE200.82 ± 0.390.79 ± 0.280.78 ± 0.260.78 ± 0.250.77 ± 0.25
ImageMSE500.81 ± 0.390.79 ± 0.290.78 ± 0.260.78 ± 0.250.77 ± 0.25
ImageMSE1000.81 ± 0.390.79 ± 0.290.78 ± 0.260.78 ± 0.250.77 ± 0.25
ImageSAM80.66 ± 0.470.61 ± 0.340.57 ± 0.320.55 ± 0.310.53 ± 0.30
ImageSAM200.63 ± 0.480.58 ± 0.360.55 ± 0.330.52 ± 0.320.51 ± 0.31
ImageSAM500.63 ± 0.480.58 ± 0.360.54 ± 0.330.52 ± 0.320.50 ± 0.31
ImageSAM1000.63 ± 0.480.58 ± 0.360.54 ± 0.330.52 ± 0.320.50 ± 0.31
ImageUIQ81.00 ± 0.020.20 ± 0.000.20 ± 0.000.27 ± 0.000.35 ± 0.00
ImageUIQ200.79 ± 0.410.78 ± 0.280.77 ± 0.260.77 ± 0.250.77 ± 0.24
ImageUIQ500.82 ± 0.390.80 ± 0.270.79 ± 0.260.79 ± 0.250.78 ± 0.25
ImageUIQ1000.82 ± 0.390.80 ± 0.270.79 ± 0.260.79 ± 0.250.78 ± 0.25
LBPCosine80.14 ± 0.350.15 ± 0.230.16 ± 0.210.17 ± 0.210.17 ± 0.21
LBPCosine200.09 ± 0.280.06 ± 0.170.07 ± 0.150.07 ± 0.150.07 ± 0.14
LBPCosine500.21 ± 0.410.41 ± 0.240.40 ± 0.230.36 ± 0.230.36 ± 0.24
LBPCosine1000.38 ± 0.490.62 ± 0.290.55 ± 0.260.51 ± 0.230.48 ± 0.23
LBPEuclidean80.14 ± 0.350.15 ± 0.230.16 ± 0.210.17 ± 0.210.17 ± 0.21
LBPEuclidean200.09 ± 0.280.06 ± 0.170.07 ± 0.150.07 ± 0.150.07 ± 0.14
LBPEuclidean500.21 ± 0.410.41 ± 0.240.40 ± 0.230.36 ± 0.230.36 ± 0.24
LBPEuclidean1000.38 ± 0.490.62 ± 0.290.55 ± 0.260.51 ± 0.230.48 ± 0.23
LBPJaccard80.14 ± 0.350.15 ± 0.230.16 ± 0.210.17 ± 0.210.17 ± 0.21
LBPJaccard200.09 ± 0.280.06 ± 0.170.07 ± 0.150.07 ± 0.150.07 ± 0.14
LBPJaccard500.21 ± 0.410.41 ± 0.240.40 ± 0.230.36 ± 0.230.36 ± 0.24
LBPJaccard1000.38 ± 0.490.62 ± 0.290.55 ± 0.260.51 ± 0.230.48 ± 0.23
LBPManhattan80.14 ± 0.350.15 ± 0.230.16 ± 0.210.17 ± 0.210.17 ± 0.21
LBPManhattan200.09 ± 0.280.06 ± 0.170.07 ± 0.150.07 ± 0.150.07 ± 0.14
LBPManhattan500.21 ± 0.410.41 ± 0.240.40 ± 0.230.36 ± 0.230.36 ± 0.24
LBPManhattan1000.38 ± 0.490.62 ± 0.290.55 ± 0.260.51 ± 0.230.48 ± 0.23
SIFTEuclidean80.98 ± 0.150.20 ± 0.050.20 ± 0.030.27 ± 0.020.35 ± 0.02
SIFTEuclidean200.67 ± 0.470.50 ± 0.300.48 ± 0.260.49 ± 0.240.50 ± 0.21
SIFTEuclidean500.25 ± 0.430.28 ± 0.240.30 ± 0.200.31 ± 0.180.31 ± 0.18
SIFTEuclidean1000.18 ± 0.380.25 ± 0.230.28 ± 0.200.30 ± 0.180.31 ± 0.17
Table A4. AP and std. values for class 3. The highest AP with a relatively lower std. is marked in bold text.
Table A4. AP and std. values for class 3. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.82 ± 0.390.81 ± 0.260.80 ± 0.240.78 ± 0.230.77 ± 0.22
DefCharsEuclideanRaw0.81 ± 0.390.81 ± 0.270.79 ± 0.240.78 ± 0.230.77 ± 0.23
DefCharsJaccardRaw0.33 ± 0.470.29 ± 0.210.26 ± 0.150.25 ± 0.120.25 ± 0.11
DefCharsManhattanRaw0.82 ± 0.380.79 ± 0.270.78 ± 0.240.77 ± 0.230.77 ± 0.22
ImageMSE80.52 ± 0.500.50 ± 0.320.48 ± 0.280.47 ± 0.260.46 ± 0.25
ImageMSE200.48 ± 0.500.48 ± 0.320.48 ± 0.290.47 ± 0.270.46 ± 0.26
ImageMSE500.49 ± 0.500.48 ± 0.320.47 ± 0.290.46 ± 0.270.46 ± 0.26
ImageMSE1000.49 ± 0.500.48 ± 0.320.47 ± 0.290.46 ± 0.270.46 ± 0.26
ImageSAM80.55 ± 0.500.52 ± 0.290.51 ± 0.250.50 ± 0.220.49 ± 0.21
ImageSAM200.57 ± 0.490.54 ± 0.310.53 ± 0.250.51 ± 0.240.51 ± 0.22
ImageSAM500.58 ± 0.490.54 ± 0.300.52 ± 0.260.51 ± 0.240.51 ± 0.22
ImageSAM1000.57 ± 0.490.54 ± 0.310.53 ± 0.260.51 ± 0.240.51 ± 0.22
ImageUIQ80.00 ± 0.000.00 ± 0.000.00 ± 0.000.07 ± 0.000.10 ± 0.00
ImageUIQ200.59 ± 0.490.58 ± 0.350.58 ± 0.320.58 ± 0.310.58 ± 0.30
ImageUIQ500.55 ± 0.500.55 ± 0.330.55 ± 0.300.54 ± 0.280.53 ± 0.27
ImageUIQ1000.54 ± 0.500.54 ± 0.330.52 ± 0.290.52 ± 0.270.51 ± 0.26
LBPCosine80.11 ± 0.320.08 ± 0.160.08 ± 0.130.08 ± 0.120.08 ± 0.11
LBPCosine200.05 ± 0.220.04 ± 0.140.04 ± 0.130.04 ± 0.120.03 ± 0.11
LBPCosine500.59 ± 0.490.18 ± 0.150.17 ± 0.150.21 ± 0.190.21 ± 0.19
LBPCosine1000.50 ± 0.500.11 ± 0.110.13 ± 0.130.14 ± 0.100.17 ± 0.13
LBPEuclidean80.11 ± 0.320.08 ± 0.160.08 ± 0.130.08 ± 0.120.08 ± 0.11
LBPEuclidean200.05 ± 0.220.04 ± 0.140.04 ± 0.130.04 ± 0.120.03 ± 0.11
LBPEuclidean500.59 ± 0.490.18 ± 0.150.17 ± 0.150.21 ± 0.190.21 ± 0.19
LBPEuclidean1000.50 ± 0.500.11 ± 0.110.13 ± 0.130.14 ± 0.100.17 ± 0.13
LBPJaccard80.11 ± 0.320.08 ± 0.160.08 ± 0.130.08 ± 0.120.08 ± 0.11
LBPJaccard200.05 ± 0.220.04 ± 0.140.04 ± 0.130.04 ± 0.120.03 ± 0.11
LBPJaccard500.59 ± 0.490.18 ± 0.150.17 ± 0.150.21 ± 0.190.21 ± 0.19
LBPJaccard1000.50 ± 0.500.11 ± 0.110.13 ± 0.130.14 ± 0.100.17 ± 0.13
LBPManhattan80.11 ± 0.320.08 ± 0.160.08 ± 0.130.08 ± 0.120.08 ± 0.11
LBPManhattan200.05 ± 0.220.04 ± 0.140.04 ± 0.130.04 ± 0.120.03 ± 0.11
LBPManhattan500.59 ± 0.490.18 ± 0.150.17 ± 0.150.21 ± 0.190.21 ± 0.19
LBPManhattan1000.50 ± 0.500.11 ± 0.110.13 ± 0.130.14 ± 0.100.17 ± 0.13
SIFTEuclidean80.03 ± 0.180.03 ± 0.140.03 ± 0.120.08 ± 0.080.11 ± 0.06
SIFTEuclidean200.23 ± 0.420.22 ± 0.210.20 ± 0.160.21 ± 0.140.21 ± 0.13
SIFTEuclidean500.15 ± 0.360.15 ± 0.180.15 ± 0.130.15 ± 0.110.14 ± 0.10
SIFTEuclidean1000.12 ± 0.330.16 ± 0.180.17 ± 0.140.17 ± 0.110.17 ± 0.10

Appendix B. ImR Evaluation Results for the Heatsink Dataset

Table A5. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
Table A5. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size mAP @ 1 mAP @ 5 mAP @ 10 mAP @ 15 mAP @ 20 Average
DefCharsCosineRaw0.98 ± 0.010.97 ± 0.020.97 ± 0.020.97 ± 0.030.96 ± 0.030.97 ± 0.02
DefCharsEuclideanRaw0.98 ± 0.010.97 ± 0.020.97 ± 0.020.97 ± 0.030.96 ± 0.030.97 ± 0.02
DefCharsJaccardRaw0.64 ± 0.410.63 ± 0.430.62 ± 0.430.61 ± 0.440.61 ± 0.440.62 ± 0.43
DefCharsManhattanRaw0.98 ± 0.010.97 ± 0.020.97 ± 0.030.96 ± 0.030.96 ± 0.030.97 ± 0.02
ImageMSE80.89 ± 0.110.88 ± 0.130.87 ± 0.130.87 ± 0.140.86 ± 0.140.87 ± 0.13
ImageMSE200.89 ± 0.130.87 ± 0.140.87 ± 0.140.86 ± 0.150.86 ± 0.150.87 ± 0.14
ImageMSE500.88 ± 0.130.87 ± 0.150.86 ± 0.150.86 ± 0.150.86 ± 0.150.87 ± 0.15
ImageMSE1000.88 ± 0.130.87 ± 0.150.86 ± 0.150.86 ± 0.150.86 ± 0.150.87 ± 0.15
ImageSAM80.84 ± 0.150.83 ± 0.150.82 ± 0.160.82 ± 0.170.81 ± 0.170.82 ± 0.16
ImageSAM200.81 ± 0.240.81 ± 0.220.82 ± 0.220.82 ± 0.210.82 ± 0.210.82 ± 0.22
ImageSAM500.77 ± 0.300.80 ± 0.240.81 ± 0.230.81 ± 0.230.81 ± 0.220.80 ± 0.24
ImageSAM1000.77 ± 0.290.80 ± 0.240.81 ± 0.230.81 ± 0.230.81 ± 0.230.80 ± 0.24
ImageUIQ80.50 ± 0.710.50 ± 0.140.50 ± 0.280.50 ± 0.240.50 ± 0.210.50 ± 0.32
ImageUIQ200.89 ± 0.100.88 ± 0.100.87 ± 0.100.87 ± 0.100.87 ± 0.100.88 ± 0.10
ImageUIQ500.87 ± 0.130.88 ± 0.090.89 ± 0.070.89 ± 0.060.89 ± 0.060.88 ± 0.08
ImageUIQ1000.86 ± 0.140.89 ± 0.080.89 ± 0.060.89 ± 0.050.89 ± 0.050.88 ± 0.08
LBPCosine80.55 ± 0.410.53 ± 0.180.52 ± 0.110.52 ± 0.110.52 ± 0.100.53 ± 0.18
LBPCosine200.39 ± 0.510.36 ± 0.310.35 ± 0.270.34 ± 0.250.33 ± 0.240.35 ± 0.32
LBPCosine500.18 ± 0.160.18 ± 0.160.19 ± 0.170.20 ± 0.180.21 ± 0.170.19 ± 0.17
LBPCosine1000.24 ± 0.290.24 ± 0.290.25 ± 0.310.27 ± 0.330.27 ± 0.340.25 ± 0.31
LBPEuclidean80.55 ± 0.410.53 ± 0.180.52 ± 0.110.52 ± 0.110.52 ± 0.100.53 ± 0.18
LBPEuclidean200.39 ± 0.510.36 ± 0.310.35 ± 0.270.34 ± 0.250.33 ± 0.240.35 ± 0.32
LBPEuclidean500.18 ± 0.160.18 ± 0.160.19 ± 0.170.20 ± 0.180.21 ± 0.170.19 ± 0.17
LBPEuclidean1000.24 ± 0.290.24 ± 0.290.25 ± 0.310.27 ± 0.330.27 ± 0.340.25 ± 0.31
LBPJaccard80.55 ± 0.410.53 ± 0.180.52 ± 0.110.52 ± 0.110.52 ± 0.100.53 ± 0.18
LBPJaccard200.39 ± 0.510.36 ± 0.310.35 ± 0.270.34 ± 0.250.33 ± 0.240.35 ± 0.32
LBPJaccard500.18 ± 0.160.18 ± 0.160.19 ± 0.170.20 ± 0.180.21 ± 0.170.19 ± 0.17
LBPJaccard1000.24 ± 0.290.24 ± 0.290.25 ± 0.310.27 ± 0.330.27 ± 0.340.25 ± 0.31
LBPManhattan80.55 ± 0.410.53 ± 0.180.52 ± 0.110.52 ± 0.110.52 ± 0.100.53 ± 0.18
LBPManhattan200.39 ± 0.510.36 ± 0.310.35 ± 0.270.34 ± 0.250.33 ± 0.240.35 ± 0.32
LBPManhattan500.18 ± 0.160.18 ± 0.160.19 ± 0.170.20 ± 0.180.21 ± 0.170.19 ± 0.17
LBPManhattan1000.24 ± 0.290.24 ± 0.290.25 ± 0.310.27 ± 0.330.27 ± 0.340.25 ± 0.31
SIFTEuclidean80.51 ± 0.690.50 ± 0.150.50 ± 0.290.50 ± 0.240.50 ± 0.210.50 ± 0.32
SIFTEuclidean200.50 ± 0.690.51 ± 0.130.51 ± 0.270.51 ± 0.230.51 ± 0.200.51 ± 0.30
SIFTEuclidean500.49 ± 0.650.53 ± 0.090.54 ± 0.220.53 ± 0.190.52 ± 0.170.52 ± 0.26
SIFTEuclidean1000.48 ± 0.640.55 ± 0.060.56 ± 0.180.55 ± 0.150.54 ± 0.140.54 ± 0.23
Table A6. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
Table A6. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.98 ± 0.150.96 ± 0.150.95 ± 0.150.95 ± 0.160.94 ± 0.16
DefCharsEuclideanRaw0.98 ± 0.150.96 ± 0.150.95 ± 0.150.95 ± 0.160.94 ± 0.16
DefCharsJaccardRaw0.35 ± 0.480.32 ± 0.260.31 ± 0.210.31 ± 0.180.30 ± 0.17
DefCharsManhattanRaw0.97 ± 0.170.96 ± 0.160.95 ± 0.160.94 ± 0.160.94 ± 0.17
ImageMSE80.81 ± 0.390.79 ± 0.310.77 ± 0.300.77 ± 0.290.76 ± 0.29
ImageMSE200.79 ± 0.400.77 ± 0.310.76 ± 0.300.76 ± 0.300.75 ± 0.29
ImageMSE500.79 ± 0.410.77 ± 0.320.76 ± 0.300.75 ± 0.300.75 ± 0.30
ImageMSE1000.79 ± 0.410.77 ± 0.320.76 ± 0.300.76 ± 0.300.75 ± 0.30
ImageSAM80.74 ± 0.440.72 ± 0.300.71 ± 0.280.70 ± 0.270.69 ± 0.26
ImageSAM200.64 ± 0.480.66 ± 0.300.66 ± 0.280.67 ± 0.280.67 ± 0.27
ImageSAM500.56 ± 0.500.64 ± 0.300.65 ± 0.290.65 ± 0.280.65 ± 0.28
ImageSAM1000.57 ± 0.500.63 ± 0.310.65 ± 0.290.65 ± 0.280.65 ± 0.28
ImageUIQ81.00 ± 0.020.40 ± 0.010.30 ± 0.000.33 ± 0.000.35 ± 0.00
ImageUIQ200.82 ± 0.390.81 ± 0.270.80 ± 0.260.80 ± 0.250.80 ± 0.25
ImageUIQ500.78 ± 0.410.82 ± 0.250.84 ± 0.230.84 ± 0.220.85 ± 0.22
ImageUIQ1000.76 ± 0.430.83 ± 0.240.85 ± 0.220.85 ± 0.210.86 ± 0.21
LBPCosine80.26 ± 0.440.40 ± 0.230.44 ± 0.170.45 ± 0.150.46 ± 0.13
LBPCosine200.75 ± 0.430.58 ± 0.260.54 ± 0.270.52 ± 0.260.50 ± 0.27
LBPCosine500.30 ± 0.460.29 ± 0.450.31 ± 0.430.33 ± 0.390.33 ± 0.37
LBPCosine1000.45 ± 0.500.44 ± 0.490.47 ± 0.460.50 ± 0.430.51 ± 0.43
LBPEuclidean80.26 ± 0.440.40 ± 0.230.44 ± 0.170.45 ± 0.150.46 ± 0.13
LBPEuclidean200.75 ± 0.430.58 ± 0.260.54 ± 0.270.52 ± 0.260.50 ± 0.27
LBPEuclidean500.30 ± 0.460.29 ± 0.450.31 ± 0.430.33 ± 0.390.33 ± 0.37
LBPEuclidean1000.45 ± 0.500.44 ± 0.490.47 ± 0.460.50 ± 0.430.51 ± 0.43
LBPJaccard80.26 ± 0.440.40 ± 0.230.44 ± 0.170.45 ± 0.150.46 ± 0.13
LBPJaccard200.75 ± 0.430.58 ± 0.260.54 ± 0.270.52 ± 0.260.50 ± 0.27
LBPJaccard500.30 ± 0.460.29 ± 0.450.31 ± 0.430.33 ± 0.390.33 ± 0.37
LBPJaccard1000.45 ± 0.500.44 ± 0.490.47 ± 0.460.50 ± 0.430.51 ± 0.43
LBPManhattan80.26 ± 0.440.40 ± 0.230.44 ± 0.170.45 ± 0.150.46 ± 0.13
LBPManhattan200.75 ± 0.430.58 ± 0.260.54 ± 0.270.52 ± 0.260.50 ± 0.27
LBPManhattan500.30 ± 0.460.29 ± 0.450.31 ± 0.430.33 ± 0.390.33 ± 0.37
LBPManhattan1000.45 ± 0.500.44 ± 0.490.47 ± 0.460.50 ± 0.430.51 ± 0.43
SIFTEuclidean81.00 ± 0.040.40 ± 0.010.30 ± 0.000.33 ± 0.000.35 ± 0.00
SIFTEuclidean200.99 ± 0.110.42 ± 0.100.32 ± 0.100.35 ± 0.080.36 ± 0.07
SIFTEuclidean500.95 ± 0.220.47 ± 0.170.38 ± 0.180.40 ± 0.150.40 ± 0.13
SIFTEuclidean1000.93 ± 0.260.51 ± 0.210.43 ± 0.220.45 ± 0.190.45 ± 0.17
Table A7. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
Table A7. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.99 ± 0.100.99 ± 0.070.99 ± 0.070.98 ± 0.080.98 ± 0.08
DefCharsEuclideanRaw0.99 ± 0.100.99 ± 0.070.99 ± 0.070.98 ± 0.080.98 ± 0.08
DefCharsJaccardRaw0.94 ± 0.250.93 ± 0.130.92 ± 0.100.92 ± 0.090.92 ± 0.09
DefCharsManhattanRaw0.99 ± 0.100.99 ± 0.070.98 ± 0.080.98 ± 0.080.98 ± 0.08
ImageMSE80.97 ± 0.170.97 ± 0.130.96 ± 0.130.96 ± 0.130.96 ± 0.13
ImageMSE200.98 ± 0.150.97 ± 0.120.97 ± 0.130.97 ± 0.130.97 ± 0.13
ImageMSE500.98 ± 0.150.97 ± 0.120.97 ± 0.120.97 ± 0.130.97 ± 0.13
ImageMSE1000.98 ± 0.150.97 ± 0.120.97 ± 0.120.97 ± 0.130.97 ± 0.13
ImageSAM80.95 ± 0.220.94 ± 0.160.94 ± 0.160.94 ± 0.150.94 ± 0.15
ImageSAM200.98 ± 0.150.97 ± 0.120.97 ± 0.120.97 ± 0.120.97 ± 0.12
ImageSAM500.98 ± 0.140.97 ± 0.110.97 ± 0.110.97 ± 0.110.97 ± 0.11
ImageSAM1000.98 ± 0.150.97 ± 0.110.97 ± 0.110.97 ± 0.110.97 ± 0.11
ImageUIQ80.00 ± 0.000.60 ± 0.000.70 ± 0.000.67 ± 0.000.65 ± 0.00
ImageUIQ200.96 ± 0.200.95 ± 0.170.94 ± 0.170.94 ± 0.170.94 ± 0.17
ImageUIQ500.96 ± 0.200.94 ± 0.180.94 ± 0.180.93 ± 0.180.93 ± 0.19
ImageUIQ1000.96 ± 0.200.94 ± 0.180.94 ± 0.180.93 ± 0.190.93 ± 0.19
LBPCosine80.84 ± 0.360.66 ± 0.220.60 ± 0.200.60 ± 0.170.59 ± 0.16
LBPCosine200.03 ± 0.170.14 ± 0.230.16 ± 0.230.16 ± 0.220.16 ± 0.22
LBPCosine500.07 ± 0.260.07 ± 0.250.07 ± 0.250.08 ± 0.240.09 ± 0.24
LBPCosine1000.03 ± 0.180.03 ± 0.180.03 ± 0.170.03 ± 0.160.03 ± 0.15
LBPEuclidean80.84 ± 0.360.66 ± 0.220.60 ± 0.200.60 ± 0.170.59 ± 0.16
LBPEuclidean200.03 ± 0.170.14 ± 0.230.16 ± 0.230.16 ± 0.220.16 ± 0.22
LBPEuclidean500.07 ± 0.260.07 ± 0.250.07 ± 0.250.08 ± 0.240.09 ± 0.24
LBPEuclidean1000.03 ± 0.180.03 ± 0.180.03 ± 0.170.03 ± 0.160.03 ± 0.15
LBPJaccard80.84 ± 0.360.66 ± 0.220.60 ± 0.200.60 ± 0.170.59 ± 0.16
LBPJaccard200.03 ± 0.170.14 ± 0.230.16 ± 0.230.16 ± 0.220.16 ± 0.22
LBPJaccard500.07 ± 0.260.07 ± 0.250.07 ± 0.250.08 ± 0.240.09 ± 0.24
LBPJaccard1000.03 ± 0.180.03 ± 0.180.03 ± 0.170.03 ± 0.160.03 ± 0.15
LBPManhattan80.84 ± 0.360.66 ± 0.220.60 ± 0.200.60 ± 0.170.59 ± 0.16
LBPManhattan200.03 ± 0.170.14 ± 0.230.16 ± 0.230.16 ± 0.220.16 ± 0.22
LBPManhattan500.07 ± 0.260.07 ± 0.250.07 ± 0.250.08 ± 0.240.09 ± 0.24
LBPManhattan1000.03 ± 0.180.03 ± 0.180.03 ± 0.170.03 ± 0.160.03 ± 0.15
SIFTEuclidean80.02 ± 0.140.61 ± 0.050.70 ± 0.030.67 ± 0.030.65 ± 0.03
SIFTEuclidean200.02 ± 0.140.60 ± 0.050.70 ± 0.040.67 ± 0.030.65 ± 0.03
SIFTEuclidean500.03 ± 0.160.60 ± 0.070.69 ± 0.060.66 ± 0.050.65 ± 0.04
SIFTEuclidean1000.02 ± 0.150.59 ± 0.080.69 ± 0.080.66 ± 0.060.64 ± 0.06

Appendix C. ImR Evaluation Results for the Lake Ice Dataset

Table A8. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
Table A8. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size mAP @ 1 mAP @ 5 mAP @ 10 mAP @ 15 mAP @ 20 Average
DefCharsCosineRaw0.95 ± 0.030.90 ± 0.060.87 ± 0.090.85 ± 0.100.84 ± 0.120.88 ± 0.08
DefCharsEuclideanRaw0.95 ± 0.030.90 ± 0.070.87 ± 0.090.85 ± 0.110.83 ± 0.130.88 ± 0.09
DefCharsJaccardRaw0.84 ± 0.120.80 ± 0.170.77 ± 0.200.74 ± 0.240.72 ± 0.270.77 ± 0.20
DefCharsManhattanRaw0.96 ± 0.030.92 ± 0.050.89 ± 0.070.87 ± 0.090.86 ± 0.110.90 ± 0.07
ImageMSE80.93 ± 0.080.87 ± 0.140.82 ± 0.170.80 ± 0.190.78 ± 0.210.84 ± 0.16
ImageMSE200.93 ± 0.090.86 ± 0.160.82 ± 0.190.79 ± 0.210.78 ± 0.230.84 ± 0.18
ImageMSE500.93 ± 0.100.86 ± 0.160.82 ± 0.190.79 ± 0.210.77 ± 0.230.83 ± 0.18
ImageMSE1000.93 ± 0.100.87 ± 0.160.82 ± 0.190.80 ± 0.210.78 ± 0.230.84 ± 0.18
ImageSAM80.94 ± 0.070.90 ± 0.090.85 ± 0.130.81 ± 0.160.79 ± 0.180.86 ± 0.13
ImageSAM200.94 ± 0.090.89 ± 0.130.84 ± 0.160.81 ± 0.190.78 ± 0.210.85 ± 0.16
ImageSAM500.93 ± 0.090.89 ± 0.120.85 ± 0.150.81 ± 0.180.79 ± 0.200.85 ± 0.15
ImageSAM1000.93 ± 0.090.89 ± 0.120.85 ± 0.150.81 ± 0.180.79 ± 0.200.85 ± 0.15
ImageUIQ80.25 ± 0.500.25 ± 0.190.25 ± 0.260.25 ± 0.260.25 ± 0.210.25 ± 0.28
ImageUIQ200.91 ± 0.120.84 ± 0.190.79 ± 0.230.76 ± 0.260.74 ± 0.280.81 ± 0.22
ImageUIQ500.92 ± 0.100.86 ± 0.170.81 ± 0.210.78 ± 0.230.76 ± 0.260.83 ± 0.19
ImageUIQ1000.93 ± 0.100.86 ± 0.160.82 ± 0.190.79 ± 0.220.77 ± 0.230.83 ± 0.18
LBPCosine80.13 ± 0.200.13 ± 0.180.13 ± 0.180.13 ± 0.170.13 ± 0.180.13 ± 0.18
LBPCosine200.10 ± 0.100.11 ± 0.120.11 ± 0.120.11 ± 0.120.12 ± 0.120.11 ± 0.12
LBPCosine500.24 ± 0.490.20 ± 0.380.15 ± 0.270.14 ± 0.230.13 ± 0.200.17 ± 0.31
LBPCosine1000.29 ± 0.470.26 ± 0.420.27 ± 0.340.25 ± 0.330.23 ± 0.310.26 ± 0.37
LBPEuclidean80.13 ± 0.200.13 ± 0.180.13 ± 0.180.13 ± 0.170.13 ± 0.180.13 ± 0.18
LBPEuclidean200.10 ± 0.100.11 ± 0.120.11 ± 0.120.11 ± 0.120.12 ± 0.120.11 ± 0.12
LBPEuclidean500.24 ± 0.490.20 ± 0.380.15 ± 0.270.14 ± 0.230.13 ± 0.200.17 ± 0.31
LBPEuclidean1000.29 ± 0.470.26 ± 0.420.27 ± 0.340.25 ± 0.330.23 ± 0.310.26 ± 0.37
LBPJaccard80.13 ± 0.200.13 ± 0.180.13 ± 0.180.13 ± 0.170.13 ± 0.180.13 ± 0.18
LBPJaccard200.10 ± 0.100.11 ± 0.120.11 ± 0.120.11 ± 0.120.12 ± 0.120.11 ± 0.12
LBPJaccard500.24 ± 0.490.20 ± 0.380.15 ± 0.270.14 ± 0.230.13 ± 0.200.17 ± 0.31
LBPJaccard1000.29 ± 0.470.26 ± 0.420.27 ± 0.340.25 ± 0.330.23 ± 0.310.26 ± 0.37
LBPManhattan80.13 ± 0.200.13 ± 0.180.13 ± 0.180.13 ± 0.170.13 ± 0.180.13 ± 0.18
LBPManhattan200.10 ± 0.100.11 ± 0.120.11 ± 0.120.11 ± 0.120.12 ± 0.120.11 ± 0.12
LBPManhattan500.24 ± 0.490.20 ± 0.380.15 ± 0.270.14 ± 0.230.13 ± 0.200.17 ± 0.31
LBPManhattan1000.29 ± 0.470.26 ± 0.420.27 ± 0.340.25 ± 0.330.23 ± 0.310.26 ± 0.37
SIFTEuclidean80.27 ± 0.490.26 ± 0.200.26 ± 0.270.26 ± 0.260.26 ± 0.220.26 ± 0.29
SIFTEuclidean200.48 ± 0.380.42 ± 0.260.42 ± 0.290.40 ± 0.290.40 ± 0.260.42 ± 0.30
SIFTEuclidean500.61 ± 0.140.56 ± 0.140.55 ± 0.170.53 ± 0.180.52 ± 0.180.55 ± 0.16
SIFTEuclidean1000.63 ± 0.160.65 ± 0.140.65 ± 0.150.64 ± 0.150.62 ± 0.160.64 ± 0.15
Table A9. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
Table A9. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.92 ± 0.280.87 ± 0.260.84 ± 0.270.83 ± 0.270.83 ± 0.28
DefCharsEuclideanRaw0.92 ± 0.270.86 ± 0.260.84 ± 0.270.83 ± 0.270.83 ± 0.27
DefCharsJaccardRaw0.80 ± 0.400.77 ± 0.270.75 ± 0.260.72 ± 0.250.72 ± 0.24
DefCharsManhattanRaw0.94 ± 0.240.88 ± 0.240.86 ± 0.250.85 ± 0.260.84 ± 0.26
ImageMSE80.94 ± 0.230.89 ± 0.240.87 ± 0.250.86 ± 0.260.85 ± 0.25
ImageMSE200.97 ± 0.180.91 ± 0.220.88 ± 0.250.87 ± 0.250.86 ± 0.25
ImageMSE500.97 ± 0.180.91 ± 0.220.88 ± 0.240.87 ± 0.250.86 ± 0.25
ImageMSE1000.96 ± 0.190.91 ± 0.220.88 ± 0.240.87 ± 0.250.86 ± 0.25
ImageSAM80.96 ± 0.210.90 ± 0.220.87 ± 0.260.85 ± 0.270.84 ± 0.28
ImageSAM200.97 ± 0.170.91 ± 0.210.88 ± 0.250.86 ± 0.260.85 ± 0.27
ImageSAM500.97 ± 0.180.91 ± 0.210.88 ± 0.250.86 ± 0.260.85 ± 0.27
ImageSAM1000.97 ± 0.170.92 ± 0.210.89 ± 0.240.87 ± 0.260.85 ± 0.27
ImageUIQ80.00 ± 0.000.20 ± 0.010.10 ± 0.000.07 ± 0.000.10 ± 0.00
ImageUIQ200.96 ± 0.200.89 ± 0.250.86 ± 0.260.85 ± 0.270.84 ± 0.27
ImageUIQ500.95 ± 0.210.89 ± 0.240.87 ± 0.260.85 ± 0.270.84 ± 0.27
ImageUIQ1000.95 ± 0.220.89 ± 0.240.87 ± 0.250.86 ± 0.260.85 ± 0.27
LBPCosine80.04 ± 0.200.06 ± 0.170.07 ± 0.160.07 ± 0.160.07 ± 0.16
LBPCosine200.01 ± 0.090.01 ± 0.080.01 ± 0.070.01 ± 0.070.01 ± 0.07
LBPCosine500.00 ± 0.000.00 ± 0.030.01 ± 0.040.01 ± 0.040.01 ± 0.05
LBPCosine1000.00 ± 0.000.01 ± 0.040.09 ± 0.030.06 ± 0.020.05 ± 0.01
LBPEuclidean80.04 ± 0.200.06 ± 0.170.07 ± 0.160.07 ± 0.160.07 ± 0.16
LBPEuclidean200.01 ± 0.090.01 ± 0.080.01 ± 0.070.01 ± 0.070.01 ± 0.07
LBPEuclidean500.00 ± 0.000.00 ± 0.030.01 ± 0.040.01 ± 0.040.01 ± 0.05
LBPEuclidean1000.00 ± 0.000.01 ± 0.040.09 ± 0.030.06 ± 0.020.05 ± 0.01
LBPJaccard80.04 ± 0.200.06 ± 0.170.07 ± 0.160.07 ± 0.160.07 ± 0.16
LBPJaccard200.01 ± 0.090.01 ± 0.080.01 ± 0.070.01 ± 0.070.01 ± 0.07
LBPJaccard500.00 ± 0.000.00 ± 0.030.01 ± 0.040.01 ± 0.040.01 ± 0.05
LBPJaccard1000.00 ± 0.000.01 ± 0.040.09 ± 0.030.06 ± 0.020.05 ± 0.01
LBPManhattan80.04 ± 0.200.06 ± 0.170.07 ± 0.160.07 ± 0.160.07 ± 0.16
LBPManhattan200.01 ± 0.090.01 ± 0.080.01 ± 0.070.01 ± 0.070.01 ± 0.07
LBPManhattan500.00 ± 0.000.00 ± 0.030.01 ± 0.040.01 ± 0.040.01 ± 0.05
LBPManhattan1000.00 ± 0.000.01 ± 0.040.09 ± 0.030.06 ± 0.020.05 ± 0.01
SIFTEuclidean80.00 ± 0.040.20 ± 0.010.10 ± 0.010.07 ± 0.010.10 ± 0.00
SIFTEuclidean200.17 ± 0.380.35 ± 0.260.29 ± 0.290.28 ± 0.310.31 ± 0.30
SIFTEuclidean500.56 ± 0.500.61 ± 0.320.60 ± 0.300.59 ± 0.290.58 ± 0.28
SIFTEuclidean1000.59 ± 0.490.73 ± 0.290.74 ± 0.270.74 ± 0.260.73 ± 0.25
Table A10. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
Table A10. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.97 ± 0.170.93 ± 0.200.91 ± 0.220.89 ± 0.230.88 ± 0.24
DefCharsEuclideanRaw0.97 ± 0.170.93 ± 0.190.91 ± 0.220.90 ± 0.230.89 ± 0.24
DefCharsJaccardRaw0.91 ± 0.280.91 ± 0.260.90 ± 0.270.90 ± 0.270.90 ± 0.28
DefCharsManhattanRaw0.98 ± 0.140.95 ± 0.170.92 ± 0.200.91 ± 0.210.90 ± 0.22
ImageMSE80.97 ± 0.180.92 ± 0.220.87 ± 0.250.83 ± 0.280.81 ± 0.29
ImageMSE200.98 ± 0.140.93 ± 0.200.88 ± 0.240.84 ± 0.270.82 ± 0.29
ImageMSE500.98 ± 0.150.93 ± 0.200.88 ± 0.240.84 ± 0.270.82 ± 0.29
ImageMSE1000.98 ± 0.150.93 ± 0.200.88 ± 0.240.85 ± 0.260.82 ± 0.28
ImageSAM80.98 ± 0.130.94 ± 0.190.89 ± 0.230.85 ± 0.260.82 ± 0.28
ImageSAM200.98 ± 0.130.95 ± 0.180.91 ± 0.220.87 ± 0.250.84 ± 0.28
ImageSAM500.98 ± 0.130.95 ± 0.170.91 ± 0.220.87 ± 0.250.84 ± 0.27
ImageSAM1000.98 ± 0.130.95 ± 0.170.91 ± 0.220.88 ± 0.250.84 ± 0.27
ImageUIQ80.00 ± 0.000.40 ± 0.010.30 ± 0.000.40 ± 0.000.35 ± 0.00
ImageUIQ200.97 ± 0.170.92 ± 0.210.87 ± 0.250.84 ± 0.270.81 ± 0.29
ImageUIQ500.98 ± 0.160.93 ± 0.200.88 ± 0.240.85 ± 0.270.82 ± 0.29
ImageUIQ1000.98 ± 0.160.93 ± 0.200.88 ± 0.240.84 ± 0.270.82 ± 0.29
LBPCosine80.02 ± 0.120.02 ± 0.090.03 ± 0.080.03 ± 0.080.04 ± 0.08
LBPCosine200.23 ± 0.420.11 ± 0.150.12 ± 0.130.14 ± 0.130.15 ± 0.13
LBPCosine500.00 ± 0.000.01 ± 0.040.06 ± 0.110.07 ± 0.130.08 ± 0.14
LBPCosine1000.18 ± 0.390.16 ± 0.180.23 ± 0.130.22 ± 0.160.21 ± 0.17
LBPEuclidean80.02 ± 0.120.02 ± 0.090.03 ± 0.080.03 ± 0.080.04 ± 0.08
LBPEuclidean200.23 ± 0.420.11 ± 0.150.12 ± 0.130.14 ± 0.130.15 ± 0.13
LBPEuclidean500.00 ± 0.000.01 ± 0.040.06 ± 0.110.07 ± 0.130.08 ± 0.14
LBPEuclidean1000.18 ± 0.390.16 ± 0.180.23 ± 0.130.22 ± 0.160.21 ± 0.17
LBPJaccard80.02 ± 0.120.02 ± 0.090.03 ± 0.080.03 ± 0.080.04 ± 0.08
LBPJaccard200.23 ± 0.420.11 ± 0.150.12 ± 0.130.14 ± 0.130.15 ± 0.13
LBPJaccard500.00 ± 0.000.01 ± 0.040.06 ± 0.110.07 ± 0.130.08 ± 0.14
LBPJaccard1000.18 ± 0.390.16 ± 0.180.23 ± 0.130.22 ± 0.160.21 ± 0.17
LBPManhattan80.02 ± 0.120.02 ± 0.090.03 ± 0.080.03 ± 0.080.04 ± 0.08
LBPManhattan200.23 ± 0.420.11 ± 0.150.12 ± 0.130.14 ± 0.130.15 ± 0.13
LBPManhattan500.00 ± 0.000.01 ± 0.040.06 ± 0.110.07 ± 0.130.08 ± 0.14
LBPManhattan1000.18 ± 0.390.16 ± 0.180.23 ± 0.130.22 ± 0.160.21 ± 0.17
SIFTEuclidean80.07 ± 0.260.44 ± 0.150.34 ± 0.160.43 ± 0.120.38 ± 0.11
SIFTEuclidean200.68 ± 0.470.72 ± 0.280.69 ± 0.300.69 ± 0.270.67 ± 0.28
SIFTEuclidean500.63 ± 0.480.60 ± 0.340.57 ± 0.320.57 ± 0.290.56 ± 0.28
SIFTEuclidean1000.66 ± 0.470.70 ± 0.300.70 ± 0.280.69 ± 0.270.68 ± 0.27
Table A11. AP and std. values for class 3. The highest AP with a relatively lower std. is marked in bold text.
Table A11. AP and std. values for class 3. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.99 ± 0.120.98 ± 0.110.97 ± 0.120.97 ± 0.120.96 ± 0.13
DefCharsEuclideanRaw0.99 ± 0.120.98 ± 0.110.97 ± 0.120.96 ± 0.130.96 ± 0.13
DefCharsJaccardRaw0.97 ± 0.180.95 ± 0.160.94 ± 0.160.93 ± 0.160.92 ± 0.16
DefCharsManhattanRaw0.99 ± 0.100.98 ± 0.090.98 ± 0.100.97 ± 0.110.97 ± 0.11
ImageMSE80.99 ± 0.100.98 ± 0.090.98 ± 0.100.98 ± 0.100.97 ± 0.10
ImageMSE200.99 ± 0.100.99 ± 0.090.98 ± 0.090.98 ± 0.100.98 ± 0.10
ImageMSE500.99 ± 0.090.98 ± 0.090.98 ± 0.090.98 ± 0.100.98 ± 0.10
ImageMSE1000.99 ± 0.100.98 ± 0.090.98 ± 0.090.98 ± 0.100.98 ± 0.10
ImageSAM80.99 ± 0.100.98 ± 0.100.97 ± 0.110.96 ± 0.120.96 ± 0.12
ImageSAM200.99 ± 0.100.98 ± 0.100.97 ± 0.110.97 ± 0.120.96 ± 0.12
ImageSAM500.99 ± 0.110.98 ± 0.110.97 ± 0.120.96 ± 0.130.96 ± 0.13
ImageSAM1000.99 ± 0.110.98 ± 0.110.97 ± 0.120.96 ± 0.130.96 ± 0.13
ImageUIQ81.00 ± 0.020.40 ± 0.000.60 ± 0.000.53 ± 0.000.50 ± 0.00
ImageUIQ200.99 ± 0.090.98 ± 0.090.98 ± 0.100.97 ± 0.100.97 ± 0.11
ImageUIQ500.99 ± 0.080.99 ± 0.070.98 ± 0.080.98 ± 0.080.98 ± 0.08
ImageUIQ1000.99 ± 0.070.99 ± 0.080.99 ± 0.080.98 ± 0.080.98 ± 0.08
LBPCosine80.43 ± 0.490.40 ± 0.310.39 ± 0.260.39 ± 0.240.39 ± 0.23
LBPCosine200.13 ± 0.330.27 ± 0.310.28 ± 0.250.27 ± 0.210.27 ± 0.19
LBPCosine500.97 ± 0.160.77 ± 0.170.56 ± 0.150.48 ± 0.160.42 ± 0.16
LBPCosine1000.99 ± 0.110.87 ± 0.190.76 ± 0.160.72 ± 0.140.68 ± 0.12
LBPEuclidean80.43 ± 0.490.40 ± 0.310.39 ± 0.260.39 ± 0.240.39 ± 0.23
LBPEuclidean200.13 ± 0.330.27 ± 0.310.28 ± 0.250.27 ± 0.210.27 ± 0.19
LBPEuclidean500.97 ± 0.160.77 ± 0.170.56 ± 0.150.48 ± 0.160.42 ± 0.16
LBPEuclidean1000.99 ± 0.110.87 ± 0.190.76 ± 0.160.72 ± 0.140.68 ± 0.12
LBPJaccard80.43 ± 0.490.40 ± 0.310.39 ± 0.260.39 ± 0.240.39 ± 0.23
LBPJaccard200.13 ± 0.330.27 ± 0.310.28 ± 0.250.27 ± 0.210.27 ± 0.19
LBPJaccard500.97 ± 0.160.77 ± 0.170.56 ± 0.150.48 ± 0.160.42 ± 0.16
LBPJaccard1000.99 ± 0.110.87 ± 0.190.76 ± 0.160.72 ± 0.140.68 ± 0.12
LBPManhattan80.43 ± 0.490.40 ± 0.310.39 ± 0.260.39 ± 0.240.39 ± 0.23
LBPManhattan200.13 ± 0.330.27 ± 0.310.28 ± 0.250.27 ± 0.210.27 ± 0.19
LBPManhattan500.97 ± 0.160.77 ± 0.170.56 ± 0.150.48 ± 0.160.42 ± 0.16
LBPManhattan1000.99 ± 0.110.87 ± 0.190.76 ± 0.160.72 ± 0.140.68 ± 0.12
SIFTEuclidean81.00 ± 0.050.40 ± 0.050.60 ± 0.030.54 ± 0.020.50 ± 0.02
SIFTEuclidean200.93 ± 0.260.50 ± 0.220.62 ± 0.150.57 ± 0.150.55 ± 0.15
SIFTEuclidean500.79 ± 0.410.68 ± 0.290.71 ± 0.230.70 ± 0.220.69 ± 0.22
SIFTEuclidean1000.83 ± 0.380.74 ± 0.270.73 ± 0.230.71 ± 0.220.70 ± 0.21
Table A12. AP and std. values for class 4. The highest AP with a relatively lower std. is marked in bold text.
Table A12. AP and std. values for class 4. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.93 ± 0.250.85 ± 0.260.77 ± 0.280.72 ± 0.290.68 ± 0.29
DefCharsEuclideanRaw0.92 ± 0.270.83 ± 0.270.76 ± 0.280.71 ± 0.290.66 ± 0.30
DefCharsJaccardRaw0.69 ± 0.460.58 ± 0.400.50 ± 0.380.40 ± 0.290.33 ± 0.23
DefCharsManhattanRaw0.94 ± 0.230.89 ± 0.230.81 ± 0.250.76 ± 0.270.72 ± 0.27
ImageMSE80.81 ± 0.400.67 ± 0.360.58 ± 0.350.52 ± 0.330.47 ± 0.31
ImageMSE200.79 ± 0.410.63 ± 0.390.54 ± 0.370.49 ± 0.340.45 ± 0.32
ImageMSE500.78 ± 0.410.63 ± 0.380.54 ± 0.370.49 ± 0.350.44 ± 0.33
ImageMSE1000.78 ± 0.410.63 ± 0.390.54 ± 0.370.49 ± 0.350.44 ± 0.33
ImageSAM80.84 ± 0.360.76 ± 0.350.66 ± 0.350.59 ± 0.330.53 ± 0.32
ImageSAM200.80 ± 0.400.70 ± 0.400.61 ± 0.390.54 ± 0.380.49 ± 0.37
ImageSAM500.80 ± 0.400.71 ± 0.400.62 ± 0.380.55 ± 0.370.50 ± 0.37
ImageSAM1000.80 ± 0.400.71 ± 0.400.62 ± 0.380.55 ± 0.370.50 ± 0.37
ImageUIQ80.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.05 ± 0.00
ImageUIQ200.74 ± 0.440.57 ± 0.400.45 ± 0.350.38 ± 0.300.33 ± 0.26
ImageUIQ500.77 ± 0.420.61 ± 0.400.51 ± 0.370.45 ± 0.340.39 ± 0.30
ImageUIQ1000.78 ± 0.410.63 ± 0.390.54 ± 0.370.48 ± 0.360.44 ± 0.33
LBPCosine80.03 ± 0.170.02 ± 0.080.02 ± 0.070.02 ± 0.060.02 ± 0.06
LBPCosine200.04 ± 0.200.04 ± 0.110.04 ± 0.100.04 ± 0.090.04 ± 0.09
LBPCosine500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.010.00 ± 0.01
LBPCosine1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPEuclidean80.03 ± 0.170.02 ± 0.080.02 ± 0.070.02 ± 0.060.02 ± 0.06
LBPEuclidean200.04 ± 0.200.04 ± 0.110.04 ± 0.100.04 ± 0.090.04 ± 0.09
LBPEuclidean500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.010.00 ± 0.01
LBPEuclidean1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPJaccard80.03 ± 0.170.02 ± 0.080.02 ± 0.070.02 ± 0.060.02 ± 0.06
LBPJaccard200.04 ± 0.200.04 ± 0.110.04 ± 0.100.04 ± 0.090.04 ± 0.09
LBPJaccard500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.010.00 ± 0.01
LBPJaccard1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPManhattan80.03 ± 0.170.02 ± 0.080.02 ± 0.070.02 ± 0.060.02 ± 0.06
LBPManhattan200.04 ± 0.200.04 ± 0.110.04 ± 0.100.04 ± 0.090.04 ± 0.09
LBPManhattan500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.010.00 ± 0.01
LBPManhattan1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
SIFTEuclidean80.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.05 ± 0.00
SIFTEuclidean200.15 ± 0.360.10 ± 0.200.07 ± 0.140.06 ± 0.120.09 ± 0.09
SIFTEuclidean500.46 ± 0.500.36 ± 0.320.31 ± 0.270.28 ± 0.250.27 ± 0.22
SIFTEuclidean1000.43 ± 0.500.45 ± 0.350.44 ± 0.310.41 ± 0.280.39 ± 0.26

Appendix D. ImR Evaluation Results for the Wind Turbine Blade Dataset

Table A13. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
Table A13. mAP and std. values. The highest mAP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size mAP @ 1 mAP @ 5 mAP @ 10 mAP @ 15 mAP @ 20 Average
DefCharsCosineRaw0.74 ± 0.060.62 ± 0.140.57 ± 0.160.54 ± 0.180.52 ± 0.190.60 ± 0.15
DefCharsEuclideanRaw0.75 ± 0.080.63 ± 0.150.58 ± 0.170.54 ± 0.190.52 ± 0.200.60 ± 0.16
DefCharsJaccardRaw0.41 ± 0.200.41 ± 0.170.37 ± 0.180.36 ± 0.170.36 ± 0.170.38 ± 0.18
DefCharsManhattanRaw0.77 ± 0.090.64 ± 0.170.59 ± 0.180.55 ± 0.200.53 ± 0.200.62 ± 0.17
ImageMSE80.52 ± 0.230.44 ± 0.330.42 ± 0.340.42 ± 0.340.40 ± 0.330.44 ± 0.31
ImageMSE200.54 ± 0.280.44 ± 0.350.43 ± 0.360.41 ± 0.350.40 ± 0.350.44 ± 0.34
ImageMSE500.54 ± 0.270.44 ± 0.360.43 ± 0.360.41 ± 0.350.40 ± 0.350.44 ± 0.34
ImageMSE1000.54 ± 0.270.44 ± 0.350.42 ± 0.360.41 ± 0.350.40 ± 0.350.44 ± 0.34
ImageSAM80.52 ± 0.200.41 ± 0.300.41 ± 0.320.39 ± 0.330.38 ± 0.320.42 ± 0.29
ImageSAM200.51 ± 0.220.43 ± 0.330.41 ± 0.340.39 ± 0.350.39 ± 0.350.43 ± 0.32
ImageSAM500.50 ± 0.250.42 ± 0.340.41 ± 0.350.39 ± 0.350.38 ± 0.350.42 ± 0.33
ImageSAM1000.51 ± 0.250.43 ± 0.330.41 ± 0.350.40 ± 0.350.38 ± 0.350.43 ± 0.33
ImageUIQ80.25 ± 0.500.25 ± 0.190.25 ± 0.100.25 ± 0.100.25 ± 0.070.25 ± 0.19
ImageUIQ200.50 ± 0.280.43 ± 0.330.42 ± 0.320.42 ± 0.320.41 ± 0.310.44 ± 0.31
ImageUIQ500.50 ± 0.270.44 ± 0.340.43 ± 0.340.42 ± 0.330.41 ± 0.320.44 ± 0.32
ImageUIQ1000.52 ± 0.250.45 ± 0.350.42 ± 0.350.42 ± 0.330.41 ± 0.320.44 ± 0.32
LBPCosine80.26 ± 0.230.28 ± 0.140.26 ± 0.160.26 ± 0.160.25 ± 0.150.26 ± 0.17
LBPCosine200.19 ± 0.250.20 ± 0.290.20 ± 0.280.21 ± 0.250.21 ± 0.210.20 ± 0.26
LBPCosine500.17 ± 0.280.17 ± 0.270.16 ± 0.250.17 ± 0.230.17 ± 0.240.17 ± 0.25
LBPCosine1000.16 ± 0.260.18 ± 0.290.17 ± 0.290.18 ± 0.270.18 ± 0.280.17 ± 0.28
LBPEuclidean80.26 ± 0.230.28 ± 0.140.26 ± 0.160.26 ± 0.160.25 ± 0.150.26 ± 0.17
LBPEuclidean200.19 ± 0.250.20 ± 0.290.20 ± 0.280.21 ± 0.250.21 ± 0.210.20 ± 0.26
LBPEuclidean500.17 ± 0.280.17 ± 0.270.16 ± 0.250.17 ± 0.230.17 ± 0.240.17 ± 0.25
LBPEuclidean1000.16 ± 0.260.18 ± 0.290.17 ± 0.290.18 ± 0.270.18 ± 0.280.17 ± 0.28
LBPJaccard80.26 ± 0.230.28 ± 0.140.26 ± 0.160.26 ± 0.160.25 ± 0.150.26 ± 0.17
LBPJaccard200.19 ± 0.250.20 ± 0.290.20 ± 0.280.21 ± 0.250.21 ± 0.210.20 ± 0.26
LBPJaccard500.17 ± 0.280.17 ± 0.270.16 ± 0.250.17 ± 0.230.17 ± 0.240.17 ± 0.25
LBPJaccard1000.16 ± 0.260.18 ± 0.290.17 ± 0.290.18 ± 0.270.18 ± 0.280.17 ± 0.28
LBPManhattan80.26 ± 0.230.28 ± 0.140.26 ± 0.160.26 ± 0.160.25 ± 0.150.26 ± 0.17
LBPManhattan200.19 ± 0.250.20 ± 0.290.20 ± 0.280.21 ± 0.250.21 ± 0.210.20 ± 0.26
LBPManhattan500.17 ± 0.280.17 ± 0.270.16 ± 0.250.17 ± 0.230.17 ± 0.240.17 ± 0.25
LBPManhattan1000.16 ± 0.260.18 ± 0.290.17 ± 0.290.18 ± 0.270.18 ± 0.280.17 ± 0.28
SIFTEuclidean80.24 ± 0.470.26 ± 0.210.26 ± 0.110.25 ± 0.110.25 ± 0.080.25 ± 0.20
SIFTEuclidean200.30 ± 0.360.30 ± 0.260.32 ± 0.220.31 ± 0.220.30 ± 0.210.31 ± 0.25
SIFTEuclidean500.38 ± 0.200.34 ± 0.240.34 ± 0.220.32 ± 0.220.32 ± 0.220.34 ± 0.22
SIFTEuclidean1000.40 ± 0.310.34 ± 0.330.33 ± 0.310.34 ± 0.270.33 ± 0.260.35 ± 0.30
Table A14. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
Table A14. AP and std. values for class 1. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.75 ± 0.430.64 ± 0.300.59 ± 0.280.55 ± 0.260.52 ± 0.25
DefCharsEuclideanRaw0.73 ± 0.450.64 ± 0.300.59 ± 0.260.55 ± 0.260.53 ± 0.24
DefCharsJaccardRaw0.31 ± 0.470.37 ± 0.230.31 ± 0.160.30 ± 0.140.29 ± 0.12
DefCharsManhattanRaw0.78 ± 0.420.64 ± 0.290.59 ± 0.270.55 ± 0.260.53 ± 0.24
ImageMSE80.71 ± 0.460.70 ± 0.360.68 ± 0.310.67 ± 0.290.65 ± 0.28
ImageMSE200.71 ± 0.460.71 ± 0.360.73 ± 0.300.70 ± 0.300.69 ± 0.28
ImageMSE500.70 ± 0.460.72 ± 0.360.73 ± 0.300.71 ± 0.290.69 ± 0.28
ImageMSE1000.70 ± 0.460.73 ± 0.350.73 ± 0.300.71 ± 0.280.69 ± 0.27
ImageSAM80.74 ± 0.440.69 ± 0.350.71 ± 0.310.70 ± 0.300.68 ± 0.27
ImageSAM200.71 ± 0.460.74 ± 0.350.75 ± 0.300.74 ± 0.290.74 ± 0.27
ImageSAM500.72 ± 0.450.74 ± 0.340.75 ± 0.300.74 ± 0.280.74 ± 0.26
ImageSAM1000.73 ± 0.450.74 ± 0.340.75 ± 0.300.74 ± 0.280.73 ± 0.26
ImageUIQ80.00 ± 0.000.00 ± 0.000.10 ± 0.000.20 ± 0.000.20 ± 0.01
ImageUIQ200.67 ± 0.470.69 ± 0.360.70 ± 0.320.71 ± 0.290.69 ± 0.26
ImageUIQ500.69 ± 0.470.73 ± 0.340.74 ± 0.300.73 ± 0.270.72 ± 0.25
ImageUIQ1000.69 ± 0.470.76 ± 0.330.76 ± 0.280.75 ± 0.260.73 ± 0.24
LBPCosine80.56 ± 0.500.31 ± 0.160.23 ± 0.100.22 ± 0.100.22 ± 0.10
LBPCosine200.52 ± 0.500.63 ± 0.380.62 ± 0.320.56 ± 0.280.49 ± 0.23
LBPCosine500.10 ± 0.300.12 ± 0.260.12 ± 0.240.18 ± 0.200.16 ± 0.17
LBPCosine1000.11 ± 0.320.10 ± 0.220.08 ± 0.170.13 ± 0.150.12 ± 0.15
LBPEuclidean80.56 ± 0.500.31 ± 0.160.23 ± 0.100.22 ± 0.100.22 ± 0.10
LBPEuclidean200.52 ± 0.500.63 ± 0.380.62 ± 0.320.56 ± 0.280.49 ± 0.23
LBPEuclidean500.10 ± 0.300.12 ± 0.260.12 ± 0.240.18 ± 0.200.16 ± 0.17
LBPEuclidean1000.11 ± 0.320.10 ± 0.220.08 ± 0.170.13 ± 0.150.12 ± 0.15
LBPJaccard80.56 ± 0.500.31 ± 0.160.23 ± 0.100.22 ± 0.100.22 ± 0.10
LBPJaccard200.52 ± 0.500.63 ± 0.380.62 ± 0.320.56 ± 0.280.49 ± 0.23
LBPJaccard500.10 ± 0.300.12 ± 0.260.12 ± 0.240.18 ± 0.200.16 ± 0.17
LBPJaccard1000.11 ± 0.320.10 ± 0.220.08 ± 0.170.13 ± 0.150.12 ± 0.15
LBPManhattan80.56 ± 0.500.31 ± 0.160.23 ± 0.100.22 ± 0.100.22 ± 0.10
LBPManhattan200.52 ± 0.500.63 ± 0.380.62 ± 0.320.56 ± 0.280.49 ± 0.23
LBPManhattan500.10 ± 0.300.12 ± 0.260.12 ± 0.240.18 ± 0.200.16 ± 0.17
LBPManhattan1000.11 ± 0.320.10 ± 0.220.08 ± 0.170.13 ± 0.150.12 ± 0.15
SIFTEuclidean80.00 ± 0.000.01 ± 0.040.10 ± 0.010.19 ± 0.030.20 ± 0.01
SIFTEuclidean200.21 ± 0.410.13 ± 0.180.16 ± 0.120.21 ± 0.090.22 ± 0.07
SIFTEuclidean500.35 ± 0.480.33 ± 0.260.30 ± 0.190.29 ± 0.160.28 ± 0.12
SIFTEuclidean1000.31 ± 0.470.22 ± 0.190.26 ± 0.160.27 ± 0.130.27 ± 0.11
Table A15. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
Table A15. AP and std. values for class 2. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.68 ± 0.470.65 ± 0.320.62 ± 0.270.62 ± 0.250.61 ± 0.24
DefCharsEuclideanRaw0.68 ± 0.470.64 ± 0.320.62 ± 0.280.59 ± 0.270.58 ± 0.26
DefCharsJaccardRaw0.45 ± 0.500.42 ± 0.280.41 ± 0.200.41 ± 0.180.41 ± 0.17
DefCharsManhattanRaw0.70 ± 0.460.66 ± 0.300.63 ± 0.290.60 ± 0.280.59 ± 0.27
ImageMSE80.41 ± 0.500.25 ± 0.210.23 ± 0.170.21 ± 0.130.20 ± 0.12
ImageMSE200.42 ± 0.500.23 ± 0.200.20 ± 0.160.19 ± 0.140.17 ± 0.12
ImageMSE500.44 ± 0.500.22 ± 0.200.20 ± 0.150.18 ± 0.130.17 ± 0.11
ImageMSE1000.44 ± 0.500.23 ± 0.200.20 ± 0.150.18 ± 0.130.17 ± 0.11
ImageSAM80.37 ± 0.490.21 ± 0.220.19 ± 0.190.16 ± 0.150.15 ± 0.12
ImageSAM200.30 ± 0.460.18 ± 0.230.15 ± 0.190.13 ± 0.160.13 ± 0.13
ImageSAM500.32 ± 0.470.17 ± 0.210.16 ± 0.190.14 ± 0.160.13 ± 0.12
ImageSAM1000.30 ± 0.460.18 ± 0.220.16 ± 0.190.15 ± 0.160.13 ± 0.12
ImageUIQ80.00 ± 0.000.39 ± 0.030.30 ± 0.020.20 ± 0.010.20 ± 0.01
ImageUIQ200.32 ± 0.470.24 ± 0.240.24 ± 0.200.22 ± 0.160.22 ± 0.14
ImageUIQ500.34 ± 0.480.25 ± 0.240.25 ± 0.190.23 ± 0.150.22 ± 0.14
ImageUIQ1000.42 ± 0.500.26 ± 0.240.23 ± 0.180.23 ± 0.160.22 ± 0.13
LBPCosine80.27 ± 0.450.36 ± 0.180.39 ± 0.160.37 ± 0.140.36 ± 0.14
LBPCosine200.00 ± 0.000.01 ± 0.050.04 ± 0.080.06 ± 0.080.10 ± 0.08
LBPCosine500.00 ± 0.000.00 ± 0.020.00 ± 0.020.00 ± 0.020.01 ± 0.04
LBPCosine1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPEuclidean80.27 ± 0.450.36 ± 0.180.39 ± 0.160.37 ± 0.140.36 ± 0.14
LBPEuclidean200.00 ± 0.000.01 ± 0.050.04 ± 0.080.06 ± 0.080.10 ± 0.08
LBPEuclidean500.00 ± 0.000.00 ± 0.020.00 ± 0.020.00 ± 0.020.01 ± 0.04
LBPEuclidean1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPJaccard80.27 ± 0.450.36 ± 0.180.39 ± 0.160.37 ± 0.140.36 ± 0.14
LBPJaccard200.00 ± 0.000.01 ± 0.050.04 ± 0.080.06 ± 0.080.10 ± 0.08
LBPJaccard500.00 ± 0.000.00 ± 0.020.00 ± 0.020.00 ± 0.020.01 ± 0.04
LBPJaccard1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPManhattan80.27 ± 0.450.36 ± 0.180.39 ± 0.160.37 ± 0.140.36 ± 0.14
LBPManhattan200.00 ± 0.000.01 ± 0.050.04 ± 0.080.06 ± 0.080.10 ± 0.08
LBPManhattan500.00 ± 0.000.00 ± 0.020.00 ± 0.020.00 ± 0.020.01 ± 0.04
LBPManhattan1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
SIFTEuclidean80.00 ± 0.000.39 ± 0.040.30 ± 0.020.20 ± 0.010.20 ± 0.01
SIFTEuclidean200.08 ± 0.280.29 ± 0.160.27 ± 0.110.21 ± 0.070.19 ± 0.05
SIFTEuclidean500.26 ± 0.440.21 ± 0.210.25 ± 0.150.23 ± 0.110.21 ± 0.08
SIFTEuclidean1000.21 ± 0.410.18 ± 0.210.19 ± 0.150.22 ± 0.150.22 ± 0.12
Table A16. AP and std. values for class 3. The highest AP with a relatively lower std. is marked in bold text.
Table A16. AP and std. values for class 3. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.82 ± 0.380.77 ± 0.270.73 ± 0.270.71 ± 0.270.69 ± 0.26
DefCharsEuclideanRaw0.87 ± 0.330.79 ± 0.260.75 ± 0.270.74 ± 0.260.72 ± 0.24
DefCharsJaccardRaw0.68 ± 0.470.64 ± 0.300.60 ± 0.230.58 ± 0.200.56 ± 0.19
DefCharsManhattanRaw0.90 ± 0.300.83 ± 0.250.78 ± 0.260.76 ± 0.250.74 ± 0.25
ImageMSE80.70 ± 0.460.74 ± 0.290.74 ± 0.270.73 ± 0.250.71 ± 0.25
ImageMSE200.82 ± 0.380.76 ± 0.290.74 ± 0.280.72 ± 0.260.71 ± 0.25
ImageMSE500.82 ± 0.380.76 ± 0.290.74 ± 0.280.72 ± 0.270.71 ± 0.26
ImageMSE1000.82 ± 0.380.75 ± 0.290.74 ± 0.280.72 ± 0.270.71 ± 0.26
ImageSAM80.64 ± 0.480.65 ± 0.310.64 ± 0.300.64 ± 0.280.62 ± 0.27
ImageSAM200.70 ± 0.460.67 ± 0.330.65 ± 0.310.64 ± 0.300.63 ± 0.30
ImageSAM500.71 ± 0.450.68 ± 0.330.66 ± 0.310.64 ± 0.310.63 ± 0.29
ImageSAM1000.71 ± 0.450.68 ± 0.340.65 ± 0.310.64 ± 0.310.63 ± 0.30
ImageUIQ81.00 ± 0.000.40 ± 0.000.30 ± 0.020.40 ± 0.010.35 ± 0.00
ImageUIQ200.81 ± 0.400.73 ± 0.300.69 ± 0.280.68 ± 0.270.66 ± 0.27
ImageUIQ500.78 ± 0.420.72 ± 0.310.68 ± 0.280.65 ± 0.280.64 ± 0.27
ImageUIQ1000.75 ± 0.440.71 ± 0.320.67 ± 0.300.65 ± 0.280.63 ± 0.27
LBPCosine80.22 ± 0.420.36 ± 0.270.37 ± 0.190.38 ± 0.160.38 ± 0.14
LBPCosine200.25 ± 0.440.18 ± 0.270.16 ± 0.200.20 ± 0.160.21 ± 0.14
LBPCosine500.58 ± 0.500.57 ± 0.360.53 ± 0.380.50 ± 0.350.50 ± 0.33
LBPCosine1000.54 ± 0.500.61 ± 0.390.61 ± 0.390.57 ± 0.350.59 ± 0.33
LBPEuclidean80.22 ± 0.420.36 ± 0.270.37 ± 0.190.38 ± 0.160.38 ± 0.14
LBPEuclidean200.25 ± 0.440.18 ± 0.270.16 ± 0.200.20 ± 0.160.21 ± 0.14
LBPEuclidean500.58 ± 0.500.57 ± 0.360.53 ± 0.380.50 ± 0.350.50 ± 0.33
LBPEuclidean1000.54 ± 0.500.61 ± 0.390.61 ± 0.390.57 ± 0.350.59 ± 0.33
LBPJaccard80.22 ± 0.420.36 ± 0.270.37 ± 0.190.38 ± 0.160.38 ± 0.14
LBPJaccard200.25 ± 0.440.18 ± 0.270.16 ± 0.200.20 ± 0.160.21 ± 0.14
LBPJaccard500.58 ± 0.500.57 ± 0.360.53 ± 0.380.50 ± 0.350.50 ± 0.33
LBPJaccard1000.54 ± 0.500.61 ± 0.390.61 ± 0.390.57 ± 0.350.59 ± 0.33
LBPManhattan80.22 ± 0.420.36 ± 0.270.37 ± 0.190.38 ± 0.160.38 ± 0.14
LBPManhattan200.25 ± 0.440.18 ± 0.270.16 ± 0.200.20 ± 0.160.21 ± 0.14
LBPManhattan500.58 ± 0.500.57 ± 0.360.53 ± 0.380.50 ± 0.350.50 ± 0.33
LBPManhattan1000.54 ± 0.500.61 ± 0.390.61 ± 0.390.57 ± 0.350.59 ± 0.33
SIFTEuclidean80.95 ± 0.220.46 ± 0.140.34 ± 0.100.41 ± 0.030.37 ± 0.04
SIFTEuclidean200.83 ± 0.380.67 ± 0.240.64 ± 0.260.64 ± 0.210.61 ± 0.22
SIFTEuclidean500.67 ± 0.470.69 ± 0.220.66 ± 0.190.64 ± 0.170.63 ± 0.15
SIFTEuclidean1000.86 ± 0.340.83 ± 0.210.78 ± 0.200.73 ± 0.190.70 ± 0.17
Table A17. AP and std. values for class 4. The highest AP with a relatively lower std. is marked in bold text.
Table A17. AP and std. values for class 4. The highest AP with a relatively lower std. is marked in bold text.
FeatureSimilarity MetricImage Size AP @ 1 AP @ 5 AP @ 10 AP @ 15 AP @ 20
DefCharsCosineRaw0.71 ± 0.460.43 ± 0.380.35 ± 0.280.30 ± 0.200.25 ± 0.16
DefCharsEuclideanRaw0.71 ± 0.460.43 ± 0.340.35 ± 0.270.29 ± 0.190.24 ± 0.15
DefCharsJaccardRaw0.21 ± 0.410.23 ± 0.220.17 ± 0.130.17 ± 0.110.17 ± 0.08
DefCharsManhattanRaw0.71 ± 0.460.42 ± 0.350.34 ± 0.240.29 ± 0.180.26 ± 0.14
ImageMSE80.25 ± 0.440.08 ± 0.130.05 ± 0.070.05 ± 0.050.04 ± 0.04
ImageMSE200.21 ± 0.410.06 ± 0.090.05 ± 0.070.04 ± 0.050.04 ± 0.04
ImageMSE500.21 ± 0.410.05 ± 0.090.03 ± 0.060.04 ± 0.050.04 ± 0.04
ImageMSE1000.21 ± 0.410.05 ± 0.090.03 ± 0.060.04 ± 0.050.04 ± 0.04
ImageSAM80.33 ± 0.480.11 ± 0.140.08 ± 0.090.06 ± 0.070.06 ± 0.06
ImageSAM200.33 ± 0.480.11 ± 0.200.08 ± 0.110.06 ± 0.070.05 ± 0.06
ImageSAM500.25 ± 0.440.10 ± 0.190.07 ± 0.110.05 ± 0.070.04 ± 0.06
ImageSAM1000.29 ± 0.460.10 ± 0.190.06 ± 0.110.05 ± 0.070.05 ± 0.06
ImageUIQ80.00 ± 0.000.19 ± 0.040.29 ± 0.030.19 ± 0.020.24 ± 0.02
ImageUIQ200.21 ± 0.410.08 ± 0.120.05 ± 0.080.06 ± 0.090.06 ± 0.07
ImageUIQ500.21 ± 0.410.05 ± 0.090.05 ± 0.060.05 ± 0.050.05 ± 0.04
ImageUIQ1000.21 ± 0.410.05 ± 0.090.03 ± 0.050.05 ± 0.050.05 ± 0.04
LBPCosine80.00 ± 0.000.07 ± 0.100.05 ± 0.070.05 ± 0.060.05 ± 0.05
LBPCosine200.00 ± 0.000.00 ± 0.000.00 ± 0.000.01 ± 0.030.01 ± 0.03
LBPCosine500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPCosine1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPEuclidean80.00 ± 0.000.07 ± 0.100.05 ± 0.070.05 ± 0.060.05 ± 0.05
LBPEuclidean200.00 ± 0.000.00 ± 0.000.00 ± 0.000.01 ± 0.030.01 ± 0.03
LBPEuclidean500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPEuclidean1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPJaccard80.00 ± 0.000.07 ± 0.100.05 ± 0.070.05 ± 0.060.05 ± 0.05
LBPJaccard200.00 ± 0.000.00 ± 0.000.00 ± 0.000.01 ± 0.030.01 ± 0.03
LBPJaccard500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPJaccard1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPManhattan80.00 ± 0.000.07 ± 0.100.05 ± 0.070.05 ± 0.060.05 ± 0.05
LBPManhattan200.00 ± 0.000.00 ± 0.000.00 ± 0.000.01 ± 0.030.01 ± 0.03
LBPManhattan500.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
LBPManhattan1000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
SIFTEuclidean80.00 ± 0.000.19 ± 0.040.29 ± 0.030.19 ± 0.020.24 ± 0.02
SIFTEuclidean200.08 ± 0.280.13 ± 0.130.21 ± 0.120.18 ± 0.070.18 ± 0.05
SIFTEuclidean500.25 ± 0.440.14 ± 0.160.15 ± 0.100.14 ± 0.070.14 ± 0.06
SIFTEuclidean1000.21 ± 0.410.13 ± 0.180.10 ± 0.120.13 ± 0.090.13 ± 0.07

References

  1. Halawani, A.; Teynor, A.; Setia, L.; Brunner, G.; Burkhardt, H. Fundamentals and Applications of Image Retrieval: An Overview. Datenbank-Spektrum 2006, 18, 14–23. [Google Scholar]
  2. Nakazawa, T.; Kulkarni, D.V. Wafer Map Defect Pattern Classification and Image Retrieval Using Convolutional Neural Network. IEEE Trans. Semicond. Manuf. 2018, 31, 309–314. [Google Scholar] [CrossRef]
  3. Hu, X.; Fu, M.; Zhu, Z.; Xiang, Z.; Qian, M.; Wang, J. Unsupervised defect detection algorithm for printed fabrics using content-based image retrieval techniques. Text. Res. J. 2021, 91, 2551–2566. [Google Scholar] [CrossRef]
  4. Liu, P.; El-Gohary, N. Semantic Image Retrieval and Clustering for Supporting Domain-Specific Bridge Component and Defect Classification. In Proceedings of the Construction Research Congress 2020, Tempe, AZ, USA, 8–10 March 2020; pp. 809–818. [Google Scholar] [CrossRef]
  5. Agrawal, S.; Chowdhary, A.; Agarwala, S.; Mayya, V.; Kamath, S.S. Content-based medical image retrieval system for lung diseases using deep CNNs. Int. J. Inf. Technol. 2022, 14, 3619–3627. [Google Scholar] [CrossRef] [PubMed]
  6. Xie, B.; Zhuang, Y.; Jiang, N.; Liu, J. An effective and efficient framework of content-based similarity retrieval of large CT image sequences based on WSLEN model. Multimed. Tools Appl. 2023. [CrossRef]
  7. Choe, J.; Hwang, H.J.; Seo, J.B.; Lee, S.M.; Yun, J.; Kim, M.J.; Jeong, J.; Lee, Y.; Jin, K.; Park, R.; et al. Content-based Image Retrieval by Using Deep Learning for Interstitial Lung Disease Diagnosis with Chest CT. Radiology 2022, 302, 187–197. [Google Scholar] [CrossRef]
  8. Scott, K.A.; Xu, L.; Pour, H.K. Retrieval of ice/water observations from synthetic aperture radar imagery for use in lake ice data assimilation. J. Great Lakes Res. 2020, 46, 1521–1532. [Google Scholar] [CrossRef]
  9. Stonevicius, E.; Uselis, G.; Grendaite, D. Ice Detection with Sentinel-1 SAR Backscatter Threshold in Long Sections of Temperate Climate Rivers. Remote Sens. 2022, 14, 1627. [Google Scholar] [CrossRef]
  10. Wang, Z.; Bovik, A. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  11. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Proceedings of the JPL, Summaries of the Third Annual JPL Airborne Geoscience Workshop, Volume 1: AVIRIS Workshop, Pasadena, CA, USA, 1–5 June 1992. [Google Scholar]
  12. VenkatNarayanaRao, T.; Govardhan, A. Assessment of Diverse Quality Metrics for Medical Images Including Mammography. Int. J. Comput. Appl. 2013, 83, 42–47. [Google Scholar] [CrossRef]
  13. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  14. Rajith, B.; Srivastava, M.; Agarwal, S. Edge Preserved De-noising Method for Medical X-Ray Images Using Wavelet Packet Transformation. In Emerging Research in Computing, Information, Communication and Applications; Shetty, N.R., Prasad, N., Nalini, N., Eds.; Springer India: New Delhi, India, 2016; pp. 449–467. [Google Scholar]
  15. Zhang, Y. Methods for image fusion quality assessment—A review, comparison and analysis. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 1101–1109. [Google Scholar]
  16. Boudani, F.Z.; Nacereddine, N.; Laiche, N. Content-Based Image Retrieval for Surface Defects of Hot Rolled Steel Strip Using Wavelet-Based LBP. In Progress in Artificial Intelligence and Pattern Recognition; Hernández Heredia, Y., Milián Núñez, V., Ruiz Shulcloper, J., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 404–413. [Google Scholar]
  17. Zhang, L.; Liu, X.; Lu, Z.; Liu, F.; Hong, R. Lace Fabric Image Retrieval Based on Multi-Scale and Rotation Invariant LBP. In Proceedings of the 7th International Conference on Internet Multimedia Computing and Service, ICIMCS ’15, Zhangjiajie, China, 19–21 August 2015; Association for Computing Machinery: New York, NY, USA, 2015. [Google Scholar] [CrossRef]
  18. Khan, A.; Rajvee, M.H.; Deekshatulu, B.L.; Pratap Reddy, L. A Fused LBP Texture Descriptor-Based Image Retrieval System. In Advances in Signal Processing, Embedded Systems and IoT; Chakravarthy, V., Bhateja, V., Flores Fuentes, W., Anguera, J., Vasavi, K.P., Eds.; Springer Nature Singapore: Singapore, 2023; pp. 145–154. [Google Scholar]
  19. Lai, W.C.; Srividhya, S.R. A Modified LBP Operator-Based Optimized Fuzzy Art Map Medical Image Retrieval System for Disease Diagnosis and Prediction. Biomedicines 2022, 10, 2438. [Google Scholar] [CrossRef]
  20. Zhi, L.J.; Zhang, S.M.; Zhao, D.Z.; Zhao, H.; Lin, S.K. Medical Image Retrieval Using SIFT Feature. In Proceedings of the 2009 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 October 2009; pp. 1–4. [Google Scholar] [CrossRef]
  21. Cruz, B.F.; de Assis, J.T.; Estrela, V.V.; Khelassi, A. A Compact Sift-Based Strategy for Visual Information Retrieval in Large Image Databases: Array. Med. Technol. J. 2019, 3, 402–412. [Google Scholar]
  22. Srinivas, M.; Naidu, R.R.; Sastry, C.; Mohan, C.K. Content based medical image retrieval using dictionary learning. Neurocomputing 2015, 168, 880–895. [Google Scholar] [CrossRef]
  23. Patel, B.; Yadav, k.; Ghosh, D. State-of-Art: Similarity Assessment for Content Based Image Retrieval System. In Proceedings of the 2020 IEEE International Symposium on Sustainable Energy, Signal Processing and Cyber Security (iSSSC), Gunupur Odisha, India, 16–17 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  24. Seetharaman, K.; Sathiamoorthy, S. A unified learning framework for content based medical image retrieval using a statistical model. J. King Saud Univ.-Comput. Inf. Sci. 2016, 28, 110–124. [Google Scholar] [CrossRef]
  25. Schettini, R.; Ciocca, G.; Gagliardi, I. Feature Extraction for Content-Based Image Retrieval. In Encyclopedia of Database Systems; Springer: Boston, MA, USA, 2009; pp. 1115–1119. [Google Scholar] [CrossRef]
  26. Yuan, Z.W.; Zhang, J. Feature extraction and image retrieval based on AlexNet. In Proceedings of the Eighth International Conference on Digital Image Processing (ICDIP 2016), Chengu, China, 20–22 May 2016; Falco, C.M., Jiang, X., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2016; Volume 10033, p. 100330. [Google Scholar] [CrossRef]
  27. Ali, A.; Sharma, S. Content based image retrieval using feature extraction with machine learning. In Proceedings of the 2017 International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 15–16 June 2017; pp. 1048–1053. [Google Scholar] [CrossRef]
  28. Piras, L.; Giacinto, G. Information fusion in content based image retrieval: A comprehensive overview. Inf. Fusion 2017, 37, 50–60. [Google Scholar] [CrossRef]
  29. Zhang, J.; Cosma, G.; Bugby, S.; Finke, A.; Watkins, J. Morphological Image Analysis and Feature Extraction for Reasoning with AI-based Defect Detection and Classification Models. arXiv 2023, arXiv:2307.11643. [Google Scholar] [CrossRef]
  30. Zhang, J.; Cosma, G.; Bugby, S.; Watkins, J. ForestMonkey: Toolkit for Reasoning with AI-based Defect Detection and Classification Models. arXiv 2023, arXiv:2307.13815. [Google Scholar]
  31. Zhang, J.; Cosma, G.; Watkins, J. Image Enhanced Mask R-CNN: A Deep Learning Pipeline with New Evaluation Measures for Wind Turbine Blade Defect Detection and Classification. J. Imaging 2021, 7, 46. [Google Scholar] [CrossRef]
  32. Ter-Sarkisov, A. COVID-CT-Mask-Net: Prediction of COVID-19 from CT Scans Using Regional Features. Appl. Intell. 2022, 52, 9664–9675. [Google Scholar] [CrossRef]
  33. Yang, K.; Liu, Y.; Zhang, S.; Cao, J. Surface Defect Detection of Heat Sink Based on Lightweight Fully Convolutional Network. IEEE Trans. Instrum. Meas. 2022, 71, 2512912. [Google Scholar] [CrossRef]
  34. Prabha, R.; Tom, M.; Rothermel, M.; Baltsavias, E.; Leal-Taixe, L.; Schindler, K. Lake ice monitoring with webcams and crowd-sourced images. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, V-2-2020, 549–556. [Google Scholar] [CrossRef]
  35. Latif, A.; Rasheed, A.; Sajid, U.; Ahmed, J.; Ali, N.; Ratyal, N.I.; Zafar, B.; Dar, S.H.; Sajid, M.; Khalil, T. Content-Based Image Retrieval and Feature Extraction: A Comprehensive Review. Math. Probl. Eng. 2019, 2019, 1–21. [Google Scholar] [CrossRef]
  36. Shao, H.; Wu, Y.; Cui, W.; Zhang, J. Image Retrieval Based on MPEG-7 Dominant Color Descriptor. In Proceedings of the 2008 The 9th International Conference for Young Computer Scientists, Hunan, China, 18–21 November 2008; pp. 753–757. [Google Scholar] [CrossRef]
  37. Duanmu, X. Image Retrieval Using Color Moment Invariant. In Proceedings of the 2010 Seventh International Conference on Information Technology: New Generations, Las Vegas, NV, USA, 12–14 April 2010; pp. 200–203. [Google Scholar] [CrossRef]
  38. Wang, X.Y.; Zhang, B.B.; Yang, H.Y. Content-based image retrieval by integrating color and texture features. Multimed. Tools Appl. 2014, 68, 545–569. [Google Scholar] [CrossRef]
  39. Liu, Y.; Zhang, D.; Lu, G. Region-based image retrieval with high-level semantics using decision tree learning. Pattern Recognit. 2008, 41, 2554–2570. [Google Scholar] [CrossRef]
  40. Zhang, H.; Dong, Z.; Shu, H. Object recognition by a complete set of pseudo-Zernike moment invariants. In Proceedings of the 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 930–933. [Google Scholar] [CrossRef]
  41. Guo, J.M.; Prasetyo, H.; Chen, J.H. Content-Based Image Retrieval Using Error Diffusion Block Truncation Coding Features. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 466–481. [Google Scholar] [CrossRef]
  42. Jiexian, Z.; Xiupeng, L.; Yu, F. Multiscale Distance Coherence Vector Algorithm for Content-Based Image Retrieval. Sci. World J. 2014, 2014, 615973. [Google Scholar] [CrossRef]
  43. Islam, M.M.; Zhang, D.; Lu, G. Automatic Categorization of Image Regions Using Dominant Color Based Vector Quantization. In Proceedings of the 2008 Digital Image Computing: Techniques and Applications, Canberra, Australia, 1–3 December 2008; pp. 191–198. [Google Scholar] [CrossRef]
  44. Papakostas, G.; Koulouriotis, D.; Tourassis, V. Feature Extraction Based on Wavelet Moments and Moment Invariants in Machine Vision Systems. In Human-Centric Machine Vision; Chessa, M., Solari, F., Sabatini, S.P., Eds.; IntechOpen: Rijeka, Croatia, 2012; Chapter 2. [Google Scholar] [CrossRef]
  45. Liu, G.H.; Li, Z.Y.; Zhang, L.; Xu, Y. Image retrieval based on micro-structure descriptor. Pattern Recognit. 2011, 44, 2123–2133. [Google Scholar] [CrossRef]
  46. Wang, X.Y.; Chen, Z.F.; Yun, J.J. An effective method for color image retrieval based on texture. Comput. Stand. Interfaces 2012, 34, 31–35. [Google Scholar] [CrossRef]
  47. Ashraf, R.; Bashir, K.; Irtaza, A.; Mahmood, M.T. Content Based Image Retrieval Using Embedded Neural Networks with Bandletized Regions. Entropy 2015, 17, 3552–3580. [Google Scholar] [CrossRef]
  48. Irtaza, A.; Jaffar, M.A. Categorical image retrieval through genetically optimized support vector machines (GOSVM) and hybrid texture features. Signal Image Video Process. 2015, 9, 1503–1519. [Google Scholar] [CrossRef]
  49. Fadaei, S.; Amirfattahi, R.; Ahmadzadeh, M.R. Local derivative radial patterns: A new texture descriptor for content-based image retrieval. Signal Process. 2017, 137, 274–286. [Google Scholar] [CrossRef]
  50. Wang, X.; Wang, Z. A novel method for image retrieval based on structure elements’ descriptor. J. Vis. Commun. Image Represent. 2013, 24, 63–74. [Google Scholar] [CrossRef]
  51. Ali, N.; Bajwa, K.B.; Sablatnig, R.; Chatzichristofis, S.A.; Iqbal, Z.; Rashid, M.; Habib, H.A. A Novel Image Retrieval Based on Visual Words Integration of SIFT and SURF. PLoS ONE 2016, 11, e0157428. [Google Scholar] [CrossRef] [PubMed]
  52. Lazebnik, S.; Schmid, C.; Ponce, J. Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 2169–2178. [Google Scholar] [CrossRef]
  53. Mehmood, Z.; Anwar, S.M.; Ali, N.; Habib, H.A.; Rashid, M. A Novel Image Retrieval Based on a Combination of Local and Global Histograms of Visual Words. Math. Probl. Eng. 2016, 2016, 8217250. [Google Scholar] [CrossRef]
  54. Naeem, M.; Ashraf, R.; Ali, N.; Ahmad, M.; Habib, M.A. Bottom up Approach for Better Requirements Elicitation. In Proceedings of the International Conference on Future Networks and Distributed Systems (ICFNDS ’17), Cambridge, UK, 19–20 July 2017; Association for Computing Machinery: New York, NY, USA, 2017. [Google Scholar] [CrossRef]
  55. Zafar, B.; Ashraf, R.; Ali, N.; Iqbal, M.; Sajid, M.; Dar, S.; Ratyal, N. A Novel Discriminating and Relative Global Spatial Image Representation with Applications in CBIR. Appl. Sci. 2018, 8, 2242. [Google Scholar] [CrossRef]
  56. Anwar, H.; Zambanini, S.; Kampel, M. A rotation-invariant bag of visual words model for symbols based ancient coin classification. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 5257–5261. [Google Scholar] [CrossRef]
  57. Khan, R.; Barat, C.; Muselet, D.; Ducottet, C. Spatial histograms of soft pairwise similar patches to improve the bag-of-visual-words model. Comput. Vis. Image Underst. 2015, 132, 102–112. [Google Scholar] [CrossRef]
  58. Ashraf, R.; Ahmed, M.; Ahmad, U.; Habib, M.A.; Jabbar, S.; Naseer, K. MDCBIR-MF: Multimedia data for content-based image retrieval by using multiple features. Multimed. Tools Appl. 2020, 79, 8553–8579. [Google Scholar] [CrossRef]
  59. Mistry, Y.; Ingole, D.; Ingole, M. Content based image retrieval using hybrid features and various distance metric. J. Electr. Syst. Inf. Technol. 2018, 5, 874–888. [Google Scholar] [CrossRef]
  60. Ahmed, K.T.; Ummesafi, S.; Iqbal, A. Content based image retrieval using image features information fusion. Inf. Fusion 2019, 51, 76–99. [Google Scholar] [CrossRef]
  61. Liu, P.; Guo, J.M.; Chamnongthai, K.; Prasetyo, H. Fusion of color histogram and LBP-based features for texture image retrieval and classification. Inf. Sci. 2017, 390, 95–111. [Google Scholar] [CrossRef]
  62. Nazir, A.; Ashraf, R.; Hamdani, T.; Ali, N. Content based image retrieval system by using HSV color histogram, discrete wavelet transform and edge histogram descriptor. In Proceedings of the 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 3–4 March 2018; pp. 1–6. [Google Scholar] [CrossRef]
  63. Kang, L.W.; Hsu, C.Y.; Chen, H.W.; Lu, C.S.; Lin, C.Y.; Pei, S.C. Feature-Based Sparse Representation for Image Similarity Assessment. IEEE Trans. Multimed. 2011, 13, 1019–1030. [Google Scholar] [CrossRef]
  64. Zhao, Z.Q.; Glotin, H.; Xie, Z.; Gao, J.; Wu, X. Cooperative Sparse Representation in Two Opposite Directions for Semi-Supervised Image Annotation. IEEE Trans. Image Process. 2012, 21, 4218–4231. [Google Scholar] [CrossRef] [PubMed]
  65. Thiagarajan, J.J.; Natesan Ramamurthy, K.; Sattigeri, P.; Spanias, A. Supervised local sparse coding of sub-image features for image retrieval. In Proceedings of the 2012 19th IEEE International Conference on Image Processing, Lake Buena Vista, FL, USA, 30 September–3 October 2012; pp. 3117–3120. [Google Scholar] [CrossRef]
  66. Wang, D.; Hoi, S.C.; He, Y.; Zhu, J. Retrieval-Based Face Annotation by Weak Label Regularized Local Coordinate Coding. In Proceedings of the 19th ACM International Conference on Multimedia (MM ’11), Scottsdale, AZ, USA, 28 November–1 December 2011; Association for Computing Machinery: New York, NY, USA, 2011; pp. 353–362. [Google Scholar] [CrossRef]
  67. Hong, C.; Zhu, J. Hypergraph-based multi-example ranking with sparse representation for transductive learning image retrieval. Neurocomputing 2013, 101, 94–103. [Google Scholar] [CrossRef]
  68. Mohamadzadeh, S.; Farsi, H. Content-based image retrieval system via sparse representation. IET Comput. Vis. 2016, 10, 95–102. [Google Scholar] [CrossRef]
  69. Li, Q.; Han, Y.; Dang, J. Sketch4Image: A novel framework for sketch-based image retrieval based on product quantization with coding residuals. Multimed. Tools Appl. 2016, 75, 2419–2434. [Google Scholar] [CrossRef]
  70. Duan, Y.; Lu, J.; Feng, J.; Zhou, J. Context-Aware Local Binary Feature Learning for Face Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1139–1153. [Google Scholar] [CrossRef]
  71. Shamna, P.; Govindan, V.; Abdul Nazeer, K. Content-based medical image retrieval by spatial matching of visual words. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 58–71. [Google Scholar] [CrossRef]
  72. Mo, D.; Wong, W.K.; Liu, X.; Ge, Y. Concentrated hashing with neighborhood embedding for image retrieval and classification. Int. J. Mach. Learn. Cybern. 2022, 13, 1571–1587. [Google Scholar] [CrossRef]
  73. Deep, G.; Kaur, J.; Singh, S.P.; Nayak, S.R.; Kumar, M.; Kautish, S. MeQryEP: A Texture Based Descriptor for Biomedical Image Retrieval. J. Healthc. Eng. 2022, 2022, 9505229. [Google Scholar] [CrossRef] [PubMed]
  74. Tan, R.Z.; Venkatarayalu, N.; Atmosukarto, I.; Premkumar, A.B.; Teh, T.E.; Thinn, K.K.; Xue, M. Supervised Image Retrieval and Ranking Technique for Lock-in Thermography Images. In Proceedings of the 2022 IEEE International Symposium on the Physical and Failure Analysis of Integrated Circuits (IPFA), Singapore, 18–20 July 2022; pp. 1–5. [Google Scholar] [CrossRef]
  75. Gassner, M.; Barranco Garcia, J.; Tanadini-Lang, S.; Bertoldo, F.; Fröhlich, F.; Guckenberger, M.; Haueis, S.; Pelzer, C.; Reyes, M.; Schmithausen, P.; et al. Saliency-Enhanced Content-Based Image Retrieval for Diagnosis Support in Dermatology Consultation: Reader Study. JMIR Dermatol. 2023, 6, e42129. [Google Scholar] [CrossRef] [PubMed]
  76. Chen, W.; Liu, Y.; Wang, W.; Bakker, E.M.; Georgiou, T.; Fieguth, P.; Liu, L.; Lew, M.S. Deep Learning for Instance Retrieval: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 7270–7292. [Google Scholar] [CrossRef] [PubMed]
  77. Dutta, A.; Zisserman, A. The VIA Annotation Software for Images, Audio and Video. In Proceedings of the 27th ACM International Conference on Multimedia (MM ’19), Nice, France, 21–25 October 2019; ACM: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  78. Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. LabelMe: A Database and Web-Based Tool for Image Annotation. Int. J. Comput. Vis. 2008, 77, 157–173. [Google Scholar] [CrossRef]
  79. Bradski, G. The OpenCV Library. Dr. Dobb’s J. Softw. Tools 2000, 25, 120–123. [Google Scholar]
  80. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  81. Ojala, T.; Pietikainen, M.; Harwood, D. Performance evaluation of texture measures with classification based on Kullback discrimination of distributions. In Proceedings of the 12th International Conference on Pattern Recognition, Jerusalem, Israel, 9–13 October 1994; Volume 1, pp. 582–585. [Google Scholar] [CrossRef]
  82. Rahillda Nadhirah, N.R.; Hashim, U.R.; Salahuddin, L.; Kanchymalay, K.; Aji, P.W.; Teo, H.C. Local Texture Representation for Timber Defect Recognition based on Variation of LBP. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 443–448. [Google Scholar]
  83. Shao, S.; Chen, K.; Karpur, A.; Cui, Q.; Araujo, A.; Cao, B. Global Features are All You Need for Image Retrieval and Reranking. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 11036–11046. [Google Scholar]
  84. Radenović, F.; Iscen, A.; Tolias, G.; Avrithis, Y.; Chum, O. Revisiting Oxford and Paris: Large-Scale Image Retrieval Benchmarking. In Proceedings of the CVPR, 2018, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  85. Lee, S.; Seong, H.; Lee, S.; Kim, E. Correlation Verification for Image Retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5374–5384. [Google Scholar]
  86. Yang, M.; He, D.; Fan, M.; Shi, B.; Xue, X.; Li, F.; Ding, E.; Huang, J. DOLG: Single-Stage Image Retrieval with Deep Orthogonal Fusion of Local and Global Features. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 11772–11781. [Google Scholar]
  87. Cao, B.; Araujo, A.; Sim, J. Unifying Deep Local and Global Features for Image Search. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 726–743. [Google Scholar]
Figure 1. Proposed image retrieval framework.
Figure 1. Proposed image retrieval framework.
Jimaging 09 00277 g001
Figure 2. Example images of irregular patterns in each class of every dataset.
Figure 2. Example images of irregular patterns in each class of every dataset.
Jimaging 09 00277 g002
Figure 3. mAP of the best performing ImR methods using different features, similarity metrics, and image sizes (as shown in the legend). The blue-based, orange-based, green-based, and purple-based lines, respectively, represent the ImR methods which used the DefChars, resized image, LBP, and SIFT features. The image size in the legend is the size of the image that was input into the ImR approach. For example, number 8 indicates that an image of size 8 × 8 was used.
Figure 3. mAP of the best performing ImR methods using different features, similarity metrics, and image sizes (as shown in the legend). The blue-based, orange-based, green-based, and purple-based lines, respectively, represent the ImR methods which used the DefChars, resized image, LBP, and SIFT features. The image size in the legend is the size of the image that was input into the ImR approach. For example, number 8 indicates that an image of size 8 × 8 was used.
Jimaging 09 00277 g003
Figure 4. std. of the best-performing ImR methods using different features, similarity metrics, and image sizes (as shown in the legend). The blue-based, orange-based, green-based, and purple-based lines, respectively, represent the ImR methods which used DefChars, resized images, LBPs, and SIFT features. Image size in the legend is the size of the image that was input into the ImR approach. For example, number 8 indicates that an image of size 8 × 8 was used.
Figure 4. std. of the best-performing ImR methods using different features, similarity metrics, and image sizes (as shown in the legend). The blue-based, orange-based, green-based, and purple-based lines, respectively, represent the ImR methods which used DefChars, resized images, LBPs, and SIFT features. Image size in the legend is the size of the image that was input into the ImR approach. For example, number 8 indicates that an image of size 8 × 8 was used.
Jimaging 09 00277 g004
Figure 5. Average time (s) required to extract features (blue bars) and average retrieval time (s) (orange bars) of images across queries across four datasets. The value at the end of each bar is the total of the feature extraction and retrieval times. The image size column shows the size of the image that was input into the ImR approach. For example, number 8 indicates that an image of size 8 × 8 was used.
Figure 5. Average time (s) required to extract features (blue bars) and average retrieval time (s) (orange bars) of images across queries across four datasets. The value at the end of each bar is the total of the feature extraction and retrieval times. The image size column shows the size of the image that was input into the ImR approach. For example, number 8 indicates that an image of size 8 × 8 was used.
Jimaging 09 00277 g005
Table 1. Class distribution of irregular patterns in each dataset. A ‘–’ denotes that there is no such class in the dataset.
Table 1. Class distribution of irregular patterns in each dataset. A ‘–’ denotes that there is no such class in the dataset.
Number of Irregular Patterns
DatasetNumber of ImagesClass 1Class 2Class 3Class 4Total Irregular Patterns
Wind Turbine Blade Defect191897311824304
Lake Ice [34]4017606120732373155365
Chest CT [32]750231716686804665
Heatsink Defect [33]1000216049277007
Table 2. Descriptions and value ranges of DefChars introduced by Zhang et al. [29].
Table 2. Descriptions and value ranges of DefChars introduced by Zhang et al. [29].
Colour Information Extracted and Stored Separately for Irregular Patterns and Background Areas
DefChar NameValue RangeDescription
Average Hue { 0 , 1 , , 359 } Average hue value
Mode of Hue { 0 , 1 , , 359 } Most frequent hue value
Unique Number of Hue values { 1 , 2 , , 360 } Number of unique hue values
Hue Range { 0 , 1 , , 180 } Difference between maximum and minimum hue value
Average Saturation { 0 , 1 , , 254 } Average saturation value
Mode of Saturation { 0 , 1 , , 254 } Most frequent saturation value
Unique Number of Saturation { 1 , 2 , , 255 } Number of unique saturation values
Saturation Range { 0 , 1 , , 254 } Difference between maximum and minimum saturation values
Average Brightness { 0 , 1 , , 254 } Average brightness value
Mode of Brightness { 0 , 1 , , 254 } Most frequent brightness value
Unique Number of Brightness { 1 , 2 , , 255 } Unique brightness values
Brightness Range { 0 , 1 , , 254 } Difference between maximum and minimum brightness value
Colour Complexity
DefChar NameValue RangeDescription
Hue Difference [ 0 , 1 ] Hue frequency distribution difference between the defect and background areas
Saturation Difference [ 0 , 1 ] Saturation frequency distribution difference between the defect and background areas
Brightness Difference [ 0 , 1 ] Brightness frequency distribution difference between the defect and background areas
Shape Information
DefChar NameValue RangeDescription
Number of Edges { 3 , 4 , } Number of edges of the defect polygon areas
Coverage [ 0 , 1 ] Percentage of the defect polygon area covered by its bounding box
Aspect Ratio [ 0 , 1 ] Ratio between the width and height of defect bounding box
Average Turning Angles { 1 , 2 , , 180 } Average value of vertex angles of the defect polygon area
Mode of Turning Angle { 1 , 2 , , 180 } Value of vertex angles that appears the most often in the defect polygon
Shape Complexity
Description
Edge Ratio [ 0 , 1 ] Average length ratio between two adjacent edges in the defect polygon area
Followed Turns [ 0 , 1 ] Proportion of two adjacent vertices which turn to the same direction in the defect polygon area
Small Turns [ 0 , 1 ] Percentage of vertices which are smaller than 90 in the defect polygon area
Reversed Turns [ 0 , 1 ] Proportion of two adjacent vertices which turn to a different direction in the defect polygon area
Meta Information
DefChar NameValue RangeDescription
Defect Size { 1 , 2 , } Number of pixels in the defect polygon area
Neighbour Distance { 0 , 1 , 2 } Categorised distances to the nearest neighbour, 0→Short (≤100 px); 1→Long; 2→No Neighbour.
Table 3. Retrieval performance of SG and our proposed ImR framework for each dataset. The highest mAP values with a relatively lower std. are marked in bold text.
Table 3. Retrieval performance of SG and our proposed ImR framework for each dataset. The highest mAP values with a relatively lower std. are marked in bold text.
Approach mAP @ 1 mAP @ 5 mAP @ 10 mAP @ 15 mAP @ 20
Wind Turbine Blade
SG0.80 ± 0.090.66 ± 0.180.63 ± 0.190.60 ± 0.200.58 ± 0.21
Ours0.77 ± 0.090.64 ± 0.170.59 ± 0.180.55 ± 0.200.53 ± 0.20
Chest CT
SG0.69 ± 0.170.67 ± 0.180.66 ± 0.180.65 ± 0.180.65 ± 0.18
Ours0.88 ± 0.060.86 ± 0.070.85 ± 0.070.84 ± 0.070.84 ± 0.07
Heatsink
SG0.84 ± 0.050.83 ± 0.050.82 ± 0.050.81 ± 0.050.81 ± 0.05
Ours0.98 ± 0.010.97 ± 0.020.97 ± 0.030.96 ± 0.030.96 ± 0.03
Lake Ice
SG0.97 ± 0.020.96 ± 0.030.95 ± 0.040.94 ± 0.050.93 ± 0.06
Ours0.96 ± 0.030.92 ± 0.050.89 ± 0.070.87 ± 0.090.86 ± 0.11
Table 4. Speed performance (feature extraction and retrieval) of SG and our proposed ImR framework for each dataset. The second column shows the average feature extraction time in seconds (s) across queries. The third column shows the average retrieval time in seconds (s) for retrieving the results of a query; and the last column, total time, is the sum of columns 2 and 3.
Table 4. Speed performance (feature extraction and retrieval) of SG and our proposed ImR framework for each dataset. The second column shows the average feature extraction time in seconds (s) across queries. The third column shows the average retrieval time in seconds (s) for retrieving the results of a query; and the last column, total time, is the sum of columns 2 and 3.
ApproachFeature Extraction Time (s)Average Retrieval Time (s)Total Time (s)
Wind Turbine Blade
SG0.0390.00000320.039
Ours0.2550.0040.259
Chest CT
SG0.0290.000000190.029
Ours0.0100.0460.056
Heatsink
SG0.0290.000000190.029
Ours0.0060.0700.076
Lake Ice
SG0.0410.000000190.041
Ours0.1120.0530.165
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Cosma, G.; Bugby, S.; Watkins, J. Efficient Retrieval of Images with Irregular Patterns Using Morphological Image Analysis: Applications to Industrial and Healthcare Datasets. J. Imaging 2023, 9, 277. https://doi.org/10.3390/jimaging9120277

AMA Style

Zhang J, Cosma G, Bugby S, Watkins J. Efficient Retrieval of Images with Irregular Patterns Using Morphological Image Analysis: Applications to Industrial and Healthcare Datasets. Journal of Imaging. 2023; 9(12):277. https://doi.org/10.3390/jimaging9120277

Chicago/Turabian Style

Zhang, Jiajun, Georgina Cosma, Sarah Bugby, and Jason Watkins. 2023. "Efficient Retrieval of Images with Irregular Patterns Using Morphological Image Analysis: Applications to Industrial and Healthcare Datasets" Journal of Imaging 9, no. 12: 277. https://doi.org/10.3390/jimaging9120277

APA Style

Zhang, J., Cosma, G., Bugby, S., & Watkins, J. (2023). Efficient Retrieval of Images with Irregular Patterns Using Morphological Image Analysis: Applications to Industrial and Healthcare Datasets. Journal of Imaging, 9(12), 277. https://doi.org/10.3390/jimaging9120277

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop