Next Article in Journal
Predictive Factors of Canine Malignant Hepatic Diseases with Multifocal Hepatic Lesions Using Clinicopathology, Ultrasonography, and Hepatobiliary Ultrasound Scores
Previous Article in Journal
Diagnostic Utility of Canine C-Reactive Protein, Haptoglobin, and 25-Hydroxyvitamin-D in Dogs with Nasal Cavity Disease
Previous Article in Special Issue
Improved Re-Parameterized Convolution for Wildlife Detection in Neighboring Regions of Southwest China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CECS-CLIP: Fusing Domain Knowledge for Rare Wildlife Detection Model

1
School of Information Science and Technology, Beijing Forestry University, Beijing 100083, China
2
Engineering Research Center for Forestry-Oriented Intelligent Information Processing, National Forestry and Grassland Administration, Beijing 100083, China
3
State Key Laboratory of Efficient Production of Forest Resources, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Animals 2024, 14(19), 2909; https://doi.org/10.3390/ani14192909
Submission received: 30 August 2024 / Revised: 20 September 2024 / Accepted: 3 October 2024 / Published: 9 October 2024
(This article belongs to the Special Issue Recent Advances and Innovation in Wildlife Population Estimation)

Abstract

:

Simple Summary

Accurate detection of wildlife, particularly small and hidden animals, is crucial for conservation efforts. Traditional image-based methods often struggle in complex environments. This study introduces a novel approach that combines image and text data to improve detection accuracy. By incorporating textual information about animal characteristics and leveraging a Concept Enhancement Module (CEM), our model can better understand and locate animals, even in challenging conditions. Experimental results demonstrate a significant improvement in detection accuracy, achieving an average precision of 95.8% on a challenging wildlife dataset. Compared to existing multimodal target detection algorithms, this model achieved at least a 25% improvement in AP and excelled in detecting small targets of certain species, significantly surpassing existing multimodal target detection model benchmarks. This represents a substantial improvement compared to existing state-of-the-art methods. Our multimodal approach offers a promising solution for enhancing wildlife monitoring and conservation efforts.

Abstract

Accurate and efficient wildlife monitoring is essential for conservation efforts. Traditional image-based methods often struggle to detect small, occluded, or camouflaged animals due to the challenges posed by complex natural environments. To overcome these limitations, an innovative multimodal target detection framework is proposed in this study, which integrates textual information from an animal knowledge base as supplementary features to enhance detection performance. First, a concept enhancement module was developed, employing a cross-attention mechanism to fuse features based on the correlation between textual and image features, thereby obtaining enhanced image features. Secondly, a feature normalization module was developed, amplifying cosine similarity and introducing learnable parameters to continuously weight and transform image features, further enhancing their expressive power in the feature space. Rigorous experimental validation on a specialized dataset provided by the research team at Northwest A&F University demonstrates that our multimodal model achieved a 0.3% improvement in precision over single-modal methods. Compared to existing multimodal target detection algorithms, this model achieved at least a 25% improvement in AP and excelled in detecting small targets of certain species, significantly surpassing existing multimodal target detection model benchmarks. This study offers a multimodal target detection model integrating textual and image information for the conservation of rare and endangered wildlife, providing strong evidence and new perspectives for research in this field.

1. Introduction

The precipitous decline in global biodiversity, primarily attributable to habitat loss, climate change, and anthropogenic pressures, necessitates urgent and evidence-based conservation strategies [1,2]. Wildlife monitoring emerges as a foundational component of these efforts, providing critical data on population dynamics, distribution, and behavior to inform targeted conservation interventions [3,4,5,6].
Traditional wildlife monitoring has been predominantly reliant on labor-intensive methods such as direct observation [7], transect surveys [8], and mark–recapture [9]. These approaches are often constrained by limitations in data accuracy and reliability due to observer bias and potential disturbance to wildlife behavior [10]. The integration of computer vision algorithms with unmanned aerial vehicle (UAV) and satellite imagery has ushered in a new paradigm for wildlife monitoring [11]. This technological advancement enables automated species detection, identification, and tracking across expansive geographic areas [12], providing unprecedented insights into population dynamics, behavior, and habitat use [13].
Conventional computer vision algorithms for wildlife monitoring primarily rely on single-image modalities, employing large-scale annotated datasets to train one-stage (YOLO, SSD) [14,15,16] or two-stage (Faster R-CNN) detectors [17,18]. Akito Takeki et al. [19] combined deep features from FCNs and DeepLab for semantic segmentation, employing a support vector machine to effectively detect objects of varying scales in large-scale imagery. Mirka et al. [20] leveraged thermal infrared (TIR) imagery captured by drones, incorporating contextual information such as thermal contrast and repeated site imaging (RSI) to accurately monitor arboreal monkey populations. Povlsen et al. [21] utilized drone-based thermal imaging video data, incorporating image enhancement techniques and negative samples to improve the detection of wildlife species such as hares and deer. Vega et al. [22] optimized the performance of ladybug target detection by experimenting with different distance metrics in the WHD loss function, combined with the U-Net network architecture and hyperparameter tuning.
While effective in controlled environments, these methods often struggle in real-world scenarios due to challenges [23] such as small object sizes, occlusions, and varying environmental conditions commonly encountered in remote sensing imagery [24,25]. The paucity of discriminative information within images of small, occluded, or camouflaged targets poses significant challenges for accurate wildlife monitoring; for instance, detecting small birds in dense foliage due to the lack of distinctive visual features. Moreover, environmental factors such as varying lighting conditions, atmospheric disturbances, and complex backgrounds can further exacerbate these challenges. Consequently, relying solely on visual data limits the effectiveness of traditional image-based detection methods, necessitating the exploration of complementary data sources to enhance detection accuracy and robustness [26].
Textual data, containing semantic information about animal attributes, behaviors, and habitats, can aid in target localization and identification [27]. This is especially beneficial in complex natural environments where wildlife exhibits high diversity and complexity, as textual data provide valuable prior knowledge that can guide the model towards the target [28,29]. The alignment of image and text modalities remains a challenging problem in multimodal learning due to the semantic gap between them [30]. Developing robust alignment mechanisms is essential for integrating information from different modalities.
To address the limitations of traditional image-based object detection methods in detecting small, occluded, or camouflaged wildlife, this study proposes a novel multimodal framework that leverages the complementary strengths of visual and textual information. By incorporating semantic information about animal attributes, behaviors, and habitats, the framework enhances the model’s semantic understanding of the visual scene. To effectively align visual and textual modalities, a Concept Enhancement Module (CEM) is integrated, employing cross-attention to correlate image and text features. Additionally, a feature normalization module is introduced to amplify subtle visual cues, enabling the model to better discriminate between target and background objects. This multimodal approach, coupled with the enhanced feature representation, offers a robust solution for wildlife conservation and surveillance in challenging scenarios.
Our framework incorporates several key innovations: (1) We propose a novel detection framework that significantly enhances the model’s understanding of complex visual scenes by integrating textual and visual information. Notably, our model excels at detecting objects that are small in size or partially occluded, a challenge that traditional methods struggle to address. (2) We introduce a continuous feature scaling method along with a learnable contrast parameter, applying continuous weighted transformations to image features to further strengthen their representational power in the feature space, enhancing the recognition accuracy of targets that are difficult to confirm solely through visual information due to their small size or partial occlusion. (3) We introduce an innovative Concept Enhancement Module (CEM) that significantly boosts the model’s semantic understanding capabilities. The CEM uses cross-attention to effectively align visual and textual features, enabling the model to capture more discriminative representations and improve object localization and recognition. (4) Compared to previous methods, we achieved a precision of 95.2% on the rare wildlife dataset, representing a 0.3% improvement over the original methods. Compared to existing multimodal object detection algorithms, this model achieved at least a 25% improvement in average precision.

2. Related Work

Deep learning is currently a hot topic. Zong et al. proposed an intelligent framework based on deep learning, which includes a Dynamic Attention Generative Adversarial Network (DATGAN) and a Parallel Spatial–Temporal Transformer (PSTTransformer) for missing data imputation and short-term prediction of traffic flow, reducing the mean square error by 5% in the data imputation task [31]. Liu et al. presented an automatic segmentation method combining a mask R-CNN and DBSCAN clustering algorithm for segmenting overlapped poplar seedling leaves under heavy metal stress [32]. Chen et al. introduced a multi-objective reinforcement learning approach named MORL-Trip for trip recommendation, aiming to provide personalized itinerary planning services for tourists in unfamiliar cities, constructing a series of ordered Points of Interest (POIs) to maximize travel experience with temporal and spatial constraints, as well as dynamic user preferences [33]. Nan et al. proposed a real-time lightweight object detection framework called HyperYOLO for processing multimodal remote sensing imagery, addressing the challenge of real-time processing of multimodal data on resource-constrained devices [34]. Yang et al. presented a fine-grained plant recognition network named PlantNet based on transfer learning and bilinear convolutional neural networks for high-throughput plant species identification in phenotyping analysis [35].

2.1. Single-Modal Object Detection

Single-modal object detection algorithms are primarily categorized into two-stage and one-stage detectors. Two-stage methods, such as R-CNNs and their variants, generate region proposals before classification and bounding box refinement [36]. While precise, these methods are computationally demanding [37,38,39]. One-stage detectors, exemplified by SSD and YOLO, directly predict object classes and locations in a single network pass, offering faster inference speeds at the potential cost of accuracy [40,41,42]. The latter’s speed is advantageous for applications like wildlife monitoring, where real-time detection is often crucial.

2.2. Multi-Modal Object Detection

Recent advancements in object detection have explored the potential of multimodal information [43]. Xu et al. introduced MQ-Det, a framework combining text and image modalities for enhanced object detection [44]. By leveraging the complementary strengths of both modalities, MQ-Det achieves improved detection accuracy for real-world targets. Zhong et al. proposed RegionCLIP, aligning region and text features through a pre-trained model, enabling precise object localization based on textual descriptions [45]. Mengjun et al. introduced ViSTA, a cross-modal retrieval algorithm that fuses image and text features at both local and global levels [46]. Finally, the GLIP model aligns regional and linguistic features through dot product operations, achieving effective object detection and phrase localization [47]. These studies collectively demonstrate the power of multimodal fusion for object detection, particularly in challenging scenarios such as wildlife monitoring.

2.3. Wildlife Monitoring

Recent advancements in object detection have yielded innovative approaches for wildlife monitoring. Roy et al. introduced WilDect-YOLO, a real-time model specifically designed for endangered species detection [48]. This work demonstrated the potential of deep learning for accurate wildlife identification in complex environments. Verma and Gupta focused on wildlife detection from camera trap image sequences, addressing challenges posed by dynamic backgrounds [49]. Eikelboom et al. applied deep learning to detect elephants, giraffes, and zebras in aerial imagery, showcasing the method’s applicability to large-scale monitoring [50]. These studies collectively highlight the growing importance of deep learning in addressing wildlife conservation challenges.

3. Materials and Methods

3.1. Dataset

In this study, we utilized a dataset focused on the detection of endangered wildlife to assess and experiment with the performance of our system. The dataset encompass the Wolong National Nature Reserve in the southwestern region of China’s Sichuan province, as well as the adjacent Mabian Dafengding Nature Reserve. It includes 28,000 images of 11 endangered wildlife species, such as the giant panda and the yellow-throated marten. These images capture a variety of scenes across different ecological seasons under fluctuating weather conditions and within diverse habitats. The dataset not only documents the appearance of the same species at various life stages but also the changes in imagery at the same location over different time points. The dataset was divided into 70% for the training set, 15% for the validation set, and 15% for the test set, ensuring sufficient data for model training and validity for evaluation. Additionally, a small dataset constructed from images collected from the internet was utilized to train and validate the target detection model for the same species to assess its performance in practical application scenarios. Figure 1 illustrates some data from the dataset.

3.2. CECS-CLIP Network

This study introduces a novel multi-modal object detection framework, CECS-CLIP, tailored for the challenging task of rare wildlife monitoring. To address the limitations of traditional image-based methods in detecting small, occluded, or camouflaged animals, we integrate textual information to enrich the model’s understanding of the visual scene. Our framework leverages an advanced Vision-and-Language (ViL) model, incorporating a concept enhancement module to effectively fuse visual and textual features, as shown in Figure 2.
The proposed framework consists of two primary components: visual and textual representations. For textual representation, species names are extracted from the dataset and input into a pre-trained GPT model to generate detailed textual descriptions. These descriptions are then cleaned and transformed into a structured knowledge base using feature extraction techniques. The resulting knowledge base serves as input to the text encoder. Simultaneously, image data are processed through a pre-trained visual encoder and a Region Proposal Network (RPN). The RPN generates bounding box proposals, and the visual encoder aligns region features with corresponding textual features. Region of Interest (RoI) pooling is then applied to extract classification and localization information.
To further enhance the performance of the model, a Concept Enhancement Module (CEM) is introduced. The CEM employs cross-attention to weight text features based on their relevance to the image, effectively integrating visual and semantic information. Additionally, a temperature parameter is introduced to amplify the results of cosine similarity; the learnability of this parameter enables continuous weighting transformations of image features, further enhancing their representational power in the feature space. This strengthens the role of image features within the model. By combining the advantages of visual and textual data, our proposed framework provides a powerful and effective solution for wildlife conservation and monitoring, especially in challenging scenarios. This section describes the detailed structure of CECS-CLIP.

3.2.1. Baseline Model

Our framework is implemented on the basis of the popular two-stage detector, faster R-CNN. The object detector first generates candidate regions using a Region Proposal Network (RPN), followed by further classification and regression of these candidate regions. Specifically, in the first stage, a series of candidate regions likely containing objects are extracted from the input image by the RPN. These candidate regions are then projected onto the feature map, and high-level features are extracted using a visual encoder. In the second stage, these features are fed into the ROI heads for target classification and bounding box regression, thereby precisely determining the target’s category and position. Through this two-stage process, more accurate detection and recognition of target objects in rare wildlife images are achieved.

3.2.2. Visual Representation

Pre-training of the visual encoder: Direct application of the CLIP model to object detection within image regions yields suboptimal results due to its inherent focus on image–text alignment at the holistic level rather than at a granular region-level. To address this limitation, we propose an enhanced region–text alignment approach, extending CLIP’s capabilities to learn region-level visual features and establish fine-grained correspondences between image regions and textual concepts.The pretraining of the visual encoder is shown in Figure 3.
Specifically, the pre-trained CLIP model was directly utilized as a teacher model, encompassing the teacher’s visual encoder V t and language encoder L. The visual encoder V t extracted the feature representation of the image I as f I = V t ( I ) , while the language encoder L extracted the feature representation of the text j as g j = L ( j ) . The similarity s i j was computed as the dot product of the image and text features:
s i j = f I · g j f I g j
while minimizing the similarity between mismatched pairs. The softmax function was employed to transform the similarity into a probability, indicative of the match between the image region and text:
P ( f I , g j ) = exp s i j τ k = 1 N exp s i k τ
where τ is the temperature parameter, and N is the batch size. The contrastive learning loss function comprised two parts: one representing the probability of a match from image regions to text, and the other representing the probability of a match from text to image regions. Maximizing these probabilities facilitates improved alignment of image regions with text by the model. The loss function is expressed as follows:
L contrastive = 1 N i = 1 N log exp s i i τ j = 1 N exp s i j τ + log exp s i i τ j = 1 N exp s j i τ
The objective of the distillation loss is to harmonize the output of the student model with that of the teacher model, facilitating the transfer of knowledge throughout the pre-training phase. The pre-trained CLIP model, serving as the teacher model, computes the similarity s i teacher between image regions and text pairs, whereas the student model, represented by the image encoder in RegionCLIP, computes the similarity s i student for the same pairs. The distillation loss function is expressed as L distillation = i = 1 N s i teacher s i student 2 . By minimizing the discrepancy between the teacher and student model outputs, the student model acquires the knowledge of the teacher model, enhancing its ability to achieve a region-level image encoder V.
Feature extraction from regions: The set of image regions R = { r 1 , r 2 , , r n } was produced by the RPN, where each r n is a preprocessed image region. After feature extraction by the pre-trained image encoder V, the feature representations of all image regions were aggregated, forming the feature representation set, V = { v 1 , v 2 , , v n } .

3.2.3. Textual Representation

Knowledge base construction for experts: To facilitate precise description and differentiation of object classes within photographs, the advanced language model GPT-3 is utilized to autonomously construct and refine the suite of visual feature descriptors for each category. These descriptors encapsulate a range of visual attributes pertinent to object classes, including color and shape, constituting critical components in the object detection process. Our method first utilized the generative capabilities of the large language model to produce a set of detailed visual feature descriptions for each class. These descriptions also extend to more subtle characteristics such as distinctive features of specific species, like the upright tail of the Tibetan macaque. By enriching and refining the descriptors for each class, the model is ensured to capture the subtle differences necessary to distinguish between different object classes. For example, in the case of the Tibetan macaque, not only are its fur color and body size described, but special emphasis is also placed on its upright tail: a unique biological trait that aids in improving detection accuracy and distinction in complex field environments. Additionally, our method includes an attribute association step that links the generated descriptors with the detected object instances in the images. This step further improves the accuracy of object detection by enhancing the model’s ability to recognize features of specific object classes.
Text encoding: To extract features from the target objects, the text encoder of the CLIP model was employed. The CLIP model, pre-trained on large-scale image–text pairs, was capable of capturing rich semantic information. In this study, a dataset containing N classes was defined, and using expert knowledge or species encyclopedias, the appearance descriptions for each species were extracted and encoded into description vectors S N = { s 1 , s 2 , , s N } . Specifically, for the nth species, a description vector s n was constructed, which included descriptions of features such as fur color and body size. Simultaneously, species label vectors C N = { c 1 , c 2 , , c N } were defined, where each c n is a binary vector representing the class label of the nth species.
To integrate the description information and species labels, a knowledge vector E K was constructed by combining the description vectors and species label vectors. The knowledge vector was constructed using simple vector addition as follows: E N = S N + C N = { e 1 , e 2 , , e N } . Here, each e n = s n + c n is the comprehensive feature representation for the nth species, combining appearance descriptions and class label information.
Next, the knowledge vectors E N were input into the CLIP text encoder L to obtain the high-level semantic feature representations for each species, T = L ( E N ) . Thus, a feature set T was obtained, encompassing the semantic features of all species in the dataset, which could then be utilized for the object detection task.

3.2.4. Inter-Modal Interactions

Concept enhancement: A cross-modal attention-based deformable block was incorporated into the backbone network, with its core function being to significantly enhance image feature recognition accuracy through the fusion of cross-modal information. By integrating textual information, this module not only augments the semantic expression of image embeddings but also effectively considers the semantic relationships between different concepts and their interactions with image embeddings. During the textual feature weighting process, texts more closely related to the image features are assigned higher weights, and the weighted textual features are then fused with the image features to obtain enhanced image features. Through the cross-attention mechanism, the deep integration of image and textual features is achieved, resulting in an enhanced image feature representation. This fusion ensures that the image features not only retain the original visual information but also incorporate the semantic information from the textual features.The specific network architecture is shown in Figure 4.
The textual feature T represents the original textual information acquired by the model, capturing the intrinsic semantics of the text data. The image feature V represents the features extracted from images, which encompass visual information that can complement textual information. The cross-attention mechanism CA serves as a pivotal component used to fuse features from disparate modalities. It dynamically modifies feature weights by computing the correlation between image features V and textual features T, thereby enhancing the text parts closely related to the image. The feedforward network FFN is a module composed of two linear layers, in accordance with the design principle of residual connections. It learns the residuals between input and output, aiding the network in better capturing complex patterns and reducing the risk of overfitting. In this way, the CEM not only enhances image features but also improves the model’s representational and generalization capabilities through the FFN.
V aug , i = V + CA ( W q V , W k T i , W v T i )
[ V aug , i bas , V aug , i cap ] = V aug , i + FFN ( V aug , i )
In our approach, the original image feature V is first enhanced through a cross-modal attention module. This module leverages the interaction between textual features T i and image features V, accomplished through the weighted sum of query W q V , key W k T i , and value W v T i . The enhanced image feature V aug , i is composed of the sum of the original feature V and the cross-modal attention output. Subsequently, V aug , i is processed by a feedforward neural network (FFN) for further nonlinear transformation, resulting in a new set of features V aug , i cap . These features are amalgamated with the original enhanced features V aug , i bas to form the final multimodal fused feature representation.
Continuous feature scaling: During the training phase, the visual encoder employs a contrastive learning strategy to align positive visual encodings with textual encodings within the same feature space without imposing any constraints on vector length. Therefore, to avoid errors in similarity computation due to varying feature vector lengths, it is necessary to perform scale normalization before calculating their similarity. In this case, the L2 norm is used for standardization, mapping the feature vectors F I and F T onto a unit hypersphere to standardize the scales of different feature vectors. The specific process is as follows:
V = { v 1 , v 2 , , v n }
V norm = v 1 V 2 , v 2 V 2 , , v n V 2
where
V 2 = i = 1 n v i 2
The same normalization process is applied to the textual features. The normalized image and textual features are then used to compute cosine similarity, which can be simplified as an inner product similarity computation:
sim c o s ( V , I ) = V · I V 2 I 2 = V n o r m · I n o r m
Typically, contrastive learning-based methods use the softmax function to convert image–text similarity predictions into class probabilities:
p ( y = i | x ) = e cos ( t i , f ) τ j = 1 K e cos ( t j , f ) τ
P i = { p ( y = 1 | x ) , p ( y = 2 | x ) , , p ( y = N | x ) }
where the hyperparameter is a temperature coefficient with a value range of ( 0 , 1 ) , used to amplify the cosine similarity results and enhance the separation of prediction outcomes. The use of a distillation model increases the operational flexibility of this process. To further enhance the accuracy of the prediction results, a learnable parameter, Scale_factor (denoted as α ), is introduced to replace it, allowing for continuous weighted transformation of image features to further strengthen their representation in the feature space:
α α η L α
V scaled = 1 α · V norm

3.2.5. Object Detector

Object detection: The candidate regions R = { r 1 , r 2 , , r n } generated by the RPN are projected onto the feature map, resulting in feature vectors f i for each region. These feature vectors are further processed by a pre-trained region-level visual encoder, generating higher-level feature representations V scaled = { v scaled 1 , v scaled 2 , , v scaled n } . The features V are then fed into ROI heads, which consist of two subnetworks: a classification subnetwork and a regression subnetwork. The classification subnetwork outputs the class probability distribution P = { p 1 , p 2 , , p n } for each region, while the regression subnetwork outputs the bounding box regression parameters B = { b 1 , b 2 , , b n } for each region. This process extracts the target’s class and location from the input image, accomplishing the object detection task.
Loss function: In our approach, the overall loss function comprises three components, which can be expressed as follows:
L loss = L contrastive + L cls + L reg .
The image encoder incorporates a contrastive learning loss, while the object detector includes a classification loss and a bounding box regression loss. To ensure accurate bounding box prediction, a bounding box regression loss is employed. This can be mathematically expressed as follows:
L contrastive = i , j log e sim ( f i , g j ) τ k e sim ( f i , g k ) τ
where f i denotes the feature representation of image region i, g j represents the feature representation of text description j, sim ( · , · ) denotes the similarity function between features, and τ is the temperature parameter controlling the smoothness of the softmax distribution.
The classification loss uses a cross-entropy loss function to compute the difference between each ROI’s class prediction and the ground-truth labels. The cross-entropy loss function is formulated as follows:
L cls = c = 1 C y c log ( p c )
The classification loss, denoted as L c l s , is computed using the cross-entropy loss function. Here, C represents the total number of classes, and y c is the one-hot encoded true label. If a sample belongs to class c, then y c = 1 ; otherwise, y c = 0 . The model’s predicted probability that the sample belongs to class c is denoted by p c . If the predicted probability p c is close to the true label y c , the loss will be small. Conversely, if the prediction probability deviates from the true label, the loss will be larger, incentivizing the model to adjust its parameters to reduce this error. The cross-entropy loss function is applied to each Region of Interest (ROI), and the total classification loss for the entire batch is obtained by summing the classification losses across all ROIs. For the regression loss, we employ the loss function, which combines the benefits of both L1 and L2 loss functions. It is formulated as follows:
L reg = i { x , y , w , h } | Δ i | 2 2 · ( | Δ i | < 1 ) + | Δ i | · ( | Δ i | 1 )
where Δ i represents the difference between the predicted and ground truth bounding boxes in terms of the center coordinates ( x , y ) and the width and height ( w , h ) .

4. Experiments

4.1. Evaluation Metrics

We employed a robust set of evaluation metrics to assess the proposed model. The Average Precision (AP) is calculated at a confidence threshold of 0.5, which measures the model’s ability to identify objects at a moderate confidence level. Additionally, the mean Average Precision (mAP) is computed across an IoU threshold range from 0.5 to 0.95 with an increment of 0.05, allowing us to meticulously evaluate the model’s overall performance under varying strictness levels of the IoU criteria. To further refine our assessment, we defined “small”, “medium”, and “large” objects based on the area or pixel proportion they occupy in the image. Consequently, we introduce size-specific average precisions for small objects (APs), medium-sized objects (APm), and large objects (APl) to gain a more precise understanding of the model’s strengths and limitations when dealing with objects of different sizes.
In our evaluation, True Positives (TPs) represent the number of samples correctly predicted as the positive class, while False Positives (FPs) denote the number of samples incorrectly predicted as the positive class. False Negatives (FNs) are the samples that are actually positive but were incorrectly predicted as the negative class. Precision (P) is the ratio of TPs to the sum of TPs and FPs, and Average Precision (AP) is the mean of the precision values at different recall levels, which measures the area under the Precision–Recall (PR) curve. The AP is calculated using the following formula:
P = T P T P + F P
R = T P T P + F N
A P = k = 1 n ( R k R k 1 ) · P k
where R k is the recall at the k-th, threshold and P k is the precision at the k-th threshold. The mean Average Precision (mAP) at an IoU threshold of 0.5 for all categories is denoted as mAP, and it is calculated using the following formula:
m A P = 1 N i = 1 N A P i
where N is the number of categories. The mAP50:95 requires the model to maintain a high accuracy and reliability under stricter matching criteria, making it more challenging to achieve than mAP at a single IoU threshold.
For the YOLO series of single-stage object detection models, several metrics are included, such as precision (P), detection rate (DR), false positive rate (FPR), miss rate (MR), mAP50 (mean average precision at an IoU of 0.5), and mAP50:90 (mean precision at IoU thresholds ranging from 0.5 to 0.95), to more finely reflect the model’s performance at different IoU thresholds, particularly in terms of localization accuracy.

4.2. Model Performance Evaluation

The experimental framework extends beyond standalone testing of the proposed methodology to include a comparative analysis with state-of-the-art object detection algorithms. The comparative trials involved renowned single-stage object detection algorithms from the YOLO family and the esteemed two-stage fast R-CNN algorithm, both demonstrating exceptional performance across various public datasets. Additionally, experimentation with multimodal object detection models sharing the same backbone architecture—RegionCLIP, Detic, DetPro, and PromptDet—was conducted, with nearly identical parameter settings achieved in all cases.
The experimental results indicate that, on the dataset of rare and endangered wildlife, the multimodal object detection algorithm achieved a precision of 95.2%, representing an approximate improvement of 0.3% over the baseline models of the YOLO series. Compared to existing multimodal object detection algorithms, the model showed at least a 25% improvement in AP and a 15% improvement in AP50. On the more challenging web dataset, our method’s precision was comparable to the highest values of the YOLO series’ baseline models, achieving at least a 14% improvement in AP and 3.8% in AP50 over existing multimodal object detection algorithms.
To present these comparative results clearly, the AP and AP50 metrics of each model on both datasets are detailed in Table 1. Synthesizing these data, the multimodal object detection algorithm not only exhibits theoretical innovation but also demonstrates significant performance advantages in practical applications.
Based on the comparative analysis of Table 2 and Table 3, our method achieved a detection precision (P) of 95.2%, representing a 0.3% improvement over the YOLOv7s model, the most accurate in the YOLO series. This improvement signifies the superiority of our method in object detection tasks. In terms of the three key performance indicators—Detection Recall (DR), False Positive Rate (FPR), and Miss Rate (MR)—although there was a slight increase in our method’s Miss Rate (MR), it achieved the lowest False Positive Rate (FPR), indicating that our method is more effective in reducing false detections. This feature is crucial for object detection systems, especially in scenarios that demand high accuracy. Furthermore, on the more challenging web dataset, our method outperformed the others across all evaluation metrics. Considering all factors, our method demonstrates a clear advantage in improving detection accuracy and reducing false positives.
We conducted a detailed performance evaluation of object detection for 11 species of wild animals, including the Giant Panda (Ailuropoda melanoleuca), Red Panda (Ailurus fulgens), and Yellow-throated Marten (Martes flavigula). To visually present the detection accuracy of these species, the corresponding Average Precision (AP) values are plotted in Figure 5 and Figure 6. The experimental results indicate that, on the wild dataset, the method achieved significantly high precision in detecting species such as the red panda, yellow-throated marten, Tibetan macaque (Macaca thibetana), and golden snub-nosed monkey (Rhinopithecus roxellana). Specifically, the precision for the yellow-throated marten reached 98.2%, showing an improvement of at least 3.3% over the YOLO series models. A comparative analysis revealed that the method significantly outperformed the YOLO series models in AP values for these species, demonstrating superior performance under complex backgrounds and low-light conditions.
On the more challenging web dataset, the method also showed high precision in detecting species such as the Tibetan macaque, golden snub-nosed monkey, and porcupine (Erethizon dorsatum). For instance, the method achieved an AP of 95.7% for porcupine detection, representing a 2.2% improvement over the YOLO series models. These results further validate the accuracy of the algorithm across different datasets and environmental conditions. Although the detection accuracy for some species in the multimodal object detection model did not reach the highest value, overall, the method made significant progress in improving the detection accuracy of species that were poorly detected by the baseline models.
The detailed experimental results and comparative analysis are presented in Figure 6, which displays the AP values of different species in the two datasets and compares them with the YOLO series models. These results not only demonstrate the effectiveness of the multimodal object detection algorithm but also provide important references for future applications in wildlife monitoring.
In this study, CECS-CLIP was employed to detect multiple object categories in images. As shown in Figure 7, the confusion matrix provides a comprehensive understanding of the model’s performance across different categories.
From the matrix, it can be observed that the model performed excellently in detecting categories such as giant panda, red panda, and porcupine, achieving accuracies of 97%, 97%, and 95%, respectively. The detection accuracy for Sambar was slightly lower, at only 76%. This difference may be attributed to the lower frequency of Sambar appearances in the images, resulting in insufficient samples during model training. Additionally, the recall rate for Chinese serow was relatively low, indicating that the model missed a significant number of true samples in this category. This could be due to the similarity of Chinese serow features to those of other categories, making it difficult for the model to distinguish between them. Despite a few misclassifications in minority categories, the overall performance of the model in classifying the primary categories remained commendable, demonstrating its exceptional capability in recognizing and distinguishing between key species. This superior performance is not only reflected in the high accuracy but also in the model’s ability to effectively handle background noise, enhancing its reliability and effectiveness in practical applications.
Comparative experiments against classic models such as the YOLO series demonstrate a significant advantage of the method in detecting small targets. On the wild dataset, the approach achieved an accuracy of 84% in the mAP50:90 metric, surpassing the YOLO series models and highlighting the algorithm’s efficiency in handling small-sized targets. On the web dataset, despite facing more complex backgrounds and target sizes, the method achieved a mAP50:90 accuracy of 79%, demonstrating superior performance compared to the comparative models. These results confirm the effectiveness of the method in small target detection and indicate its potential for practical applications. Particularly in rare wildlife monitoring, the method significantly enhances the detection accuracy of small-sized targets, providing more reliable technical support for wildlife conservation and research. Figure 8 presents a detailed comparison of the mAP50:90 metric between the method and the YOLO series models on the two datasets, further confirming the superiority of the method in small target detection.
Further in-depth analysis of the performance in small target detection focuses on the following key metrics: average precision on small targets (APs) and average precision on medium targets (APm). These metrics provide a nuanced reflection of the model’s detection capabilities across different target sizes, which is crucial for monitoring rare wildlife where accurate identification of small targets is essential.
The experimental results indicate that, in terms of the APs metric, the method achieved a precision of 28.7%, and for the APm metric, it achieved a precision of 52.7%. These figures highlight the superiority of the method in small target detection and demonstrate its high accuracy in detecting medium-sized targets. Compared to existing models, the method shows significant performance improvements in these metrics, further proving the effectiveness and robustness of the algorithm in handling small-sized targets. Detailed experimental results are presented in Table 4.
This study aimed to perform an in-depth analysis of the contributions of each key component in the proposed multimodal object detection algorithm. To achieve this, a series of ablation experiments were designed and implemented. Using the baseline model (RegionCLIP) as a control, models were constructed to include only the Concept Enhancement (CE) module, only the Continuous Feature Scaling (CS) module, and the CECS-CLIP model proposed in this paper for comparative analysis. The core purpose of these experiments was to systematically evaluate the specific impact of key technologies, such as the concept enhancement module and the continuous weighted feature smoothing module, on the overall performance of the model. The improved models were compared with the baseline model on the wild dataset. The evaluation metrics included Average Precision(AP), Average Precision on small targets (APs), Average Precision on medium targets (APm), and Average Precision on large targets (Apl).
The experimental results indicate that the introduction of both the concept enhancement module and the continuous weighted feature smoothing module significantly enhanced the model’s detection performance. The model with only the CE module demonstrated significant improvements in metrics such as AP, AP50, and APm. This indicates that the CE module plays a key role in enhancing the model’s detection capabilities for small targets. The model with only the CS module also demonstrated enhancements in metrics such as AP, AP50, and APm. The CS module, by normalizing features, enhanced recognition capabilities for medium-sized targets. The CECS-CLIP model, which combines the CE and CS modules, achieved optimal performance across all the evaluation metrics. Notably, its AP, AP50, APm, and Apl metrics were significantly higher than those of the variants with a single module, indicating that the synergistic effect of the two modules greatly enhances overall detection performance. To more intuitively demonstrate these improvements, each improvement method and its corresponding experimental data are listed in Table 5.
Furthermore, experimental explorations were conducted to determine the impact of different scaling factors on the model’s Average Precision (AP) and AP50 precision within the feature smoothing (CS) module. The study of the generalization capability of multimodal object detection models revealed that RegionCLIP could be effectively applied to new domains with only minor adjustments. Fine-tuning experiments on the RegionCLIP model revealed the significant influence of the number of iterations on model performance. Specifically, within the range of 10,000 to 30,000 iterations, the model’s accuracy continued to improve. After 30,000 iterations of training, the model’s average precision (AP) reached 55.86%, and the AP50 metric, i.e., under the condition of a recall rate exceeding 50%, increased significantly to 85.73%.
Further research revealed that adjusting the model’s scaling factor significantly impacts performance. The specific results are displayed in Table 6. The experiments showed that when the scaling factor was set to 0.9, compared to 0.5, the model’s AP value increased by approximately 28%, and the AP50 increased by about 35%. Based on this finding, an improved method was proposed that, by optimizing the scaling factor, achieved a model performance comparable to that of the RegionCLIP model after 30,000 iterations, even with only 10,000 iterations in some cases. Based on the experimental results, the scaling factor was set to 0.9 and adopted as the standard setting for subsequent experiments.
To thoroughly analyze the impact of training iteration counts on model performance and duration, a series of controlled experiments were designed and executed, as detailed in Table 7. The experimental results are listed in the table. The table reports the time required for model training under different iteration settings, as well as the values for average precision (AP) and AP50. Detailed analysis of the tabular data reveals that within 30,000 iterations, the average precision of the model showed a steady upward trend with increasing iteration counts. However, for single-modal and multimodal object detection tasks, achieving optimal performance requires careful selection of the number of training iterations based on task characteristics and requirements. This suggests that in practical applications, balancing model performance with training costs is necessary to achieve an optimal combination of efficiency and effectiveness.
In the practice of wildlife monitoring, the use of trap cameras often results in empty shots, where the captured images do not contain any animals. To assess the model’s performance in handling such images, we specifically selected a batch of empty images for testing. These images were submitted to the model for analysis, and subsequently, we employed manual review to verify the model’s recognition results. The experimental results are presented in Table 8.
Following a meticulous manual assessment of our collected image dataset, comprising 249 targets across 10 distinct categories, our model successfully identified 214 targets, yielding an impressive accuracy rate of 85.9%. Notably, the recognition accuracy for the sambar was relatively low, at 53.3%, whereas it achieved a remarkable 100% accuracy for the giant panda. Encouragingly, all other species exhibited accuracy rates surpassing the 70% threshold. Moreover, our model demonstrated an exceptional capability in filtering out blank images, with a success rate of 92% out of the 50 blank photos inspected. The experimental results show that the model can effectively identify empty images, with an `object absence detection’ accuracy rate of 92%. This result highlights the model’s efficient filtering capability when facing scenes without targets. The model did not mistakenly identify empty images as containing objects, which not only demonstrates its excellent robustness but also proves its reliability in practical applications. This capability is crucial for reducing false positives and improving monitoring efficiency, especially in wildlife reserves where it is necessary to distinguish between real animal activity and empty shots.

5. Discussion

In this study, we focus on leveraging multimodal object detection models to enhance the neural network’s recognition capabilities on the GPNP image dataset LoTE-Animal. This model integrates image data with textual information to supplement the context of the target objects, highlighting important areas by applying continuous weighted feature smoothing to feature maps, thereby enhancing the expressiveness of the feature maps and improving detection accuracy. To comprehensively evaluate the model’s performance, we conducted extensive experiments and compared them with state-of-the-art technologies in the current field. In this section, we will discuss in detail the experimental results, including the model’s performance on the LoTE-Animal dataset and its comparison with existing technologies. Additionally, we will explore the challenges the model may face in practical application scenarios, such as handling target objects under different environmental conditions and the robustness of detection when targets are small in size or partially occluded.
The recognition results displayed in Figure 9 highlight the performance of the CECS-CLIP model under various lighting conditions, background complexities, and target sizes. These image examples include a variety of species of different sizes from the dataset, showing the model’s generalization capability in diverse scenarios. Based on these visual results, we can see the recognition accuracy of the CECS-CLIP model in actual national park environments, which is of significant practical importance for wildlife conservation and monitoring efforts.
To comprehensively evaluate the performance of our proposed multimodal object detection method (CECS-CLIP), we selected advanced multimodal object detection models that also utilize textual information, such as RegionCLIP, Detic, DetPro, and PromptDet, as baselines. These models have been proven to have high accuracy and robustness in the field of object detection. In the experiments, all models used nearly identical parameter configurations to ensure the fairness and accuracy of the comparison. Based on a careful analysis of the data presented in Figure 10, our method demonstrates significant superiority in the key evaluation metrics of Average Precision (AP) and AP50. The results of the comparative experiments are summarized in Figure 11. Notably, on the rare and endangered wildlife dataset, our method achieved an average accuracy of 95.8%, which an approximately 24% improvement over existing multimodal object detection techniques. This significant performance enhancement not only confirms the potential application of our model in the wild environment but also highlights its efficiency in detecting small targets in complex scenarios. We believe that this significant performance improvement is mainly attributed to the innovative Concept Enhancement Module (CEM) in our model and the Continuous Feature Scaling (CS) with an optimized scaling factor configuration. Through these technical improvements, our model can more effectively capture the visual content of images and achieve more accurate target localization in the feature space. The CEM enhances the model’s ability to recognize target categories by introducing textual information, while CS improves the model’s sensitivity to target shape and size by optimizing feature representation.
Furthermore, after incorporating the Continuous Feature Scaling (CS) algorithm into our experimental model, we observed improvements in both Average Precision (AP) and AP50. This study thoroughly explored the impact of different scaling factors on object detection performance and visually demonstrated the specific effects of various weight adjustments on detection accuracy and recall rates, as shown in Figure 12. This enhancement not only increased the training efficiency but also reduced the consumption of computational resources. The augmentation of cosine similarity outcomes significantly bolstered the model’s representational efficacy and discriminative capacity within the feature space. An analysis of the experimental data indicates that there are significant differences in the impacts of different parameters on model performance.
In this study, to gain an in-depth understanding of the contributions of various key components in our proposed multimodal object detection algorithm, we designed and conducted a series of ablation experiments. These experiments specifically focused on the impact of the Concept Enhancement Module (CEM) and the Continuous Feature Scaling (CS) module on model performance. The results are summarized in Figure 12, illustrating the precision performance of the different variants.
Through meticulous data analysis, we found that by simply introducing animal morphological features and environmental information, the addition of the Concept Enhancement Module (CEM) increased the model’s precision by 11%. This significant improvement indicates that integrating textual information greatly enhances the model’s ability to recognize targets. Furthermore, by incorporating the Continuous Feature Scaling (CS) module, we increased the model’s precision by an additional 7%. This result confirms the importance of introducing learnable parameters for the continuous weighted transformation of image features; a transformation that enhances the model’s representational power in the feature space and significantly improves its sensitivity to the morphological attributes and sizes of target entities.
However, based on observations from Figure 13, we also noticed that in some cases, adding only the CS module may lead to inaccurate localization of the target detection box. This may be due to the process of amplifying cosine similarity results during feature normalization, where excessive scaling can lead to feature distortion or overemphasis on certain features, thereby interfering with the model’s correct understanding of the target’s shape and size, resulting in inaccurate bounding box localization and deviation from the true position of the target.
Our CECS-CLIP model, which combines the CE and CS modules, achieved the best performance across all the evaluation metrics. These results not only confirm the effectiveness of each component in our model but also reveal their synergistic effects. This synergy not only improves the model’s overall detection accuracy but also specifically enhances its ability to detect small targets. In complex scenes, the concept enhancement module assists the CS module in capturing target features more accurately by providing rich semantic information, thereby optimizing the localization of the detection box.

6. Conclusions

In this study, the primary challenge we faced was enhancing the recognition accuracy of small targets in images of rare and endangered wildlife, especially under adverse conditions such as environmental noise, complex backgrounds, and occlusions. To address these issues, we proposed a multimodal object detection algorithm that integrates textual information. This algorithm not only utilizes image data but also incorporates rich textual information, including category names, feature descriptions, and the living environments of the targets, thereby enhancing the model’s ability to recognize the objects. Through a series of experiments, we validated the effectiveness of the proposed algorithm. The experimental results indicate that, compared to existing single-modal object detection methods, our multimodal algorithm achieved significant performance improvements in evaluation metrics such as Average Precision (AP) and AP50. These improvements are primarily attributed to the introduction of textual information, which provides additional semantic guidance for the model, and the incorporation of the Continuous Feature Scaling (CS) method, which aids in locating and identifying targets that are difficult to confirm solely based on visual information due to their small size or partial occlusion.
Furthermore, we introduced the Continuous Feature Scaling (CS) algorithm, which includes trainable parameters to enhance the discriminative power of our predictive results. This algorithm applies a series of adaptive weighted transformations to image features, significantly enhancing their representational efficacy within the feature space. This contributes to the stability and robustness of the feature maps. We integrated the Concept Enhancement Module (CEM), which utilizes cross-attention to augment the original image features, further boosting the model’s capability for object detection. Our model achieved an AP of 95.8% and an AP50 of 97.6% on the LoTE-Animal dataset. Comparative experiments with classic YOLO models and multimodal object detection models showed that our model improved the precision value by 0.3% over the most accurate YOLOv7 and increased the mAP50:90 metric by 4.3%, indicating a significant enhancement in recognition accuracy. Compared with other multimodal object detection models, the APm and Apl metrics were improved by 25% and 0.8%, respectively.
Although our method performed well in multimodal object detection tasks, certain challenges and limitations were noted during the experiments. For example, under extreme conditions such as low lighting or high background noise, the model’s detection accuracy may be impacted. As some species’ detection accuracy is influenced by their occurrence frequency in images, future work will focus on increasing data sample diversity and optimizing the model structure. By collecting and integrating a more diverse set of data samples, the model’s performance across all categories will be enhanced. Future work will also focus on further optimizing the model to improve its performance under such complex conditions. In addition, further research will be conducted to improve the model’s generalization capabilities and explore its application to broader scenarios, such as drone surveillance and real-time video analysis.

Author Contributions

F.Y. authored the manuscript. C.H. and A.L. assisted in the writing of the manuscript. S.W. and Y.S. provided constructive feedback and assisted in revising the manuscript. F.X. acquired the funding. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key R&D Program of China, grant number 2022YFF1302700; The Emergency Open Competition Project of National Forestry and Grassland Administration, grant number 202303; and the Outstanding Youth Team Project of Central Universities, grant number QNTD202308.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors do not have permission to share the data.

Acknowledgments

The authors would like to thank the three anonymous reviewers for their valuable and helpful comments that substantially improved the quality of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Roy, A.M.; Bhaduri, J.; Kumar, T.; Raj, K. A Computer Vision-Based Object Localization Model for Endangered Wildlife Detection. Ecological Economics, Forthcoming. 2022. Available online: https://ssrn.com/abstract=4315295 (accessed on 5 September 2022).
  2. Chalmers, C.; Fergus, P.; Curbelo Montanez, C.A.; Longmore, S.N.; Wich, S.A. Video analysis for the detection of animals using convolutional neural networks and consumer-grade drones. J. Unmanned Veh. Syst. 2021, 9, 112–127. [Google Scholar] [CrossRef]
  3. Pollock, K.H.; Nichols, J.D.; Simons, T.R.; Farnsworth, G.L.; Bailey, L.L.; Sauer, J.R. Large scale wildlife monitoring studies: Statistical methods for design and analysis. Environmetrics Off. J. Int. Environmetrics Soc. 2002, 13, 105–119. [Google Scholar] [CrossRef]
  4. Baratchi, M.; Meratnia, N.; Havinga, P.J.M.; Skidmore, A.K.; Toxopeus, B.A.G. Sensing Solutions for Collecting Spatio-Temporal Data for Wildlife Monitoring Applications: A Review. Sensors 2013, 13, 6054–6088. [Google Scholar] [CrossRef] [PubMed]
  5. Cristescu, R.H.; Foley, E.; Markula, A.; Jackson, G.; Jones, D.; Frere, C. Accuracy and efficiency of detection dogs: A powerful new tool for koala conservation and management. Sci. Rep. 2015, 5, 8349. [Google Scholar] [CrossRef]
  6. Tuia, D.; Kellenberger, B.; Beery, S.; Costelloe, B.R.; Zuffi, S.; Risse, B.; Mathis, A.; Mathis, M.W.; Van Langevelde, F.; Burghardt, T.; et al. Perspectives in machine learning for wildlife conservation. Nat. Commun. 2022, 13, 792. [Google Scholar] [CrossRef]
  7. Eggert, L.S.; Eggert, J.A.; Woodruff, D.S. Estimating population sizes for elusive animals: The forest elephants of Kakum National Park, Ghana. Mol. Ecol. 2003, 12, 1389–1402. [Google Scholar] [CrossRef]
  8. Silveira, L.; Jácomo, A.T.; Diniz-Filho, J.A.F. Camera trap, line transect census and track surveys: A comparative evaluation. Biol. Conserv. 2003, 114, 351–355. [Google Scholar] [CrossRef]
  9. Buckland, S.T.; Goudie, I.B.J.; Borchers, D.L. Wildlife Population Assessment: Past Developments and Future Directions. Biometrics 2000, 56, 1–12. [Google Scholar] [CrossRef]
  10. McEvoy, J.F.; Hall, G.P.; McDonald, P.G. Evaluation of unmanned aerial vehicle shape, flight path and camera type for waterfowl surveys: Disturbance effects and species recognition. PeerJ 2016, 4, e1831. [Google Scholar] [CrossRef]
  11. Gonzalez, L.F.; Montes, G.A.; Puig, E.; Johnson, S.; Mengersen, K.; Gaston, K.J. Unmanned Aerial Vehicles (UAVs) and Artificial Intelligence Revolutionizing Wildlife Monitoring and Conservation. Sensors 2016, 16, 97. [Google Scholar] [CrossRef]
  12. Kabir, R.H.; Lee, K. Wildlife Monitoring Using a Multi-UAV System with Optimal Transport Theory. Appl. Sci. 2021, 11, 4070. [Google Scholar] [CrossRef]
  13. Lee, S.; Song, Y.; Kil, S.H. Feasibility Analyses of Real-Time Detection of Wildlife Using UAV-Derived Thermal and RGB Images. Remote Sens. 2021, 13, 2169. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Li, X.; Wang, F.; Wei, B.; Li, L. A Comprehensive Review of One-stage Networks for Object Detection. In Proceedings of the 2021 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Xi’an, China, 17–19 August 2021; pp. 1–6. [Google Scholar] [CrossRef]
  15. Chen, C.; Zheng, Z.; Xu, T.; Guo, S.; Feng, S.; Yao, W.; Lan, Y. YOLO-Based UAV Technology: A Review of the Research and Its Applications. Drones 2023, 7, 190. [Google Scholar] [CrossRef]
  16. Allard, P.; Fairbrother, A.; Hope, B.; Hull, R.; Johnson, M.; Kapustka, L.; Mann, G.; McDonald, B.; Sample, B. Recommendations for the development and application of wildlife toxicity reference values. Integr. Environ. Assess. Manag. 2009, 6, 28–37. [Google Scholar] [CrossRef] [PubMed]
  17. Duan, K.; Xie, L.; Qi, H.; Bai, S.; Huang, Q.; Tian, Q. Corner Proposal Network for Anchor-Free, Two-Stage Object Detection. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer: Cham, Switzerland, 2020; pp. 399–416. [Google Scholar]
  18. Ibraheam, M.; Li, K.F.; Gebali, F.; Sielecki, L.E. A Performance Comparison and Enhancement of Animal Species Detection in Images with Various R-CNN Models. AI 2021, 2, 552–577. [Google Scholar] [CrossRef]
  19. Takeki, A.; Trinh, T.T.; Yoshihashi, R.; Kawakami, R.; Iida, M.; Naemura, T. Combining deep features for object detection at various scales: Finding small birds in landscape images. IPSJ Trans. Comput. Vis. Appl. 2016, 8, 1–7. [Google Scholar] [CrossRef]
  20. Mirka, B.; Stow, D.A.; Paulus, G.; Loerch, A.C.; Coulter, L.L.; An, L.; Lewison, R.L.; Pflüger, L.S. Evaluation of thermal infrared imaging from uninhabited aerial vehicles for arboreal wildlife surveillance. Environ. Monit. Assess. 2022, 194, 512. [Google Scholar] [CrossRef]
  21. Povlsen, P.; Bruhn, D.; Durdevic, P.; Arroyo, D.O.; Pertoldi, C. Using YOLO Object Detection to Identify Hare and Roe Deer in Thermal Aerial Video Footage—Possible Future Applications in Real-Time Automatic Drone Surveillance and Wildlife Monitoring. Drones 2024, 8, 2. [Google Scholar] [CrossRef]
  22. Vega, M.; Benítez, D.S.; Pérez, N.; Riofrío, D.; Ramón, G.; Cisneros-Heredia, D. Weighted Hausdorff Distance Loss as a Function of Different Metrics in Convolutional Neural Networks for Ladybird Beetle Detection. In Applications of Computational Intelligence; Orjuela-Cañón, A.D., Lopez, J.A., Arias-Londoño, J.D., Figueroa-García, J.C., Eds.; Springer: Cham, Switzerland, 2022; pp. 65–77. [Google Scholar]
  23. Zhu, C.; Chen, L. A Survey on Open-Vocabulary Detection and Segmentation: Past, Present, and Future. IEEE Trans. Pattern Anal. Mach. Intell. 2024; early access. [Google Scholar] [CrossRef]
  24. Su, X.; Zhang, J.; Ma, Z.; Dong, Y.; Zi, J.; Xu, N.; Zhang, H.; Xu, F.; Chen, F. Identification of Rare Wildlife in the Field Environment Based on the Improved YOLOv5 Model. Remote Sens. 2024, 16, 1535. [Google Scholar] [CrossRef]
  25. Zhang, R.; Xu, L.; Yu, Z.; Shi, Y.; Mu, C.; Xu, M. Deep-IRTarget: An Automatic Target Detector in Infrared Imagery Using Dual-Domain Feature Extraction and Allocation. IEEE Trans. Multimed. 2022, 24, 1735–1749. [Google Scholar] [CrossRef]
  26. Sharma, M.; Dhanaraj, M.; Karnam, S.; Chachlakis, D.G.; Ptucha, R.; Markopoulos, P.P.; Saber, E. YOLOrs: Object Detection in Multimodal Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1497–1508. [Google Scholar] [CrossRef]
  27. Menon, S.; Vondrick, C. Visual Classification via Description from Large Language Models. In Proceedings of the The Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  28. Chen, Y.C.; Li, L.; Yu, L.; Kholy, A.E.; Ahmed, F.; Gan, Z.; Cheng, Y.; Liu, J. UNITER: UNiversal Image-TExt Representation Learning. arXiv 2020, arXiv:1909.11740. [Google Scholar] [CrossRef]
  29. Li, X.; Yin, X.; Li, C.; Zhang, P.; Hu, X.; Zhang, L.; Wang, L.; Hu, H.; Dong, L.; Wei, F.; et al. Oscar: Object-semantics aligned pre-training for vision-language tasks. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXX 16. Springer: Cham, Switzerland, 2020; pp. 121–137. [Google Scholar]
  30. Chen, P.; Sheng, K.; Zhang, M.; Lin, M.; Shen, Y.; Lin, S.; Ren, B.; Li, K. Open Vocabulary Object Detection with Proposal Mining and Prediction Equalization. arXiv 2022, arXiv:2206.11134. [Google Scholar]
  31. Zong, X.; Qi, Y.; Yan, H.; Ye, Q. An Intelligent Deep Learning Framework for Traffic Flow Imputation and Short-term Prediction Based on Dynamic Features. Knowl.-Based Syst. 2024, 300, 112178. [Google Scholar] [CrossRef]
  32. Liu, X.; Hu, C.; Li, P. Automatic segmentation of overlapped poplar seedling leaves combining Mask R-CNN and DBSCAN. Comput. Electron. Agric. 2020, 178, 105753. [Google Scholar] [CrossRef]
  33. Chen, L.; Zhu, G.; Liang, W.; Wang, Y. Multi-objective reinforcement learning approach for trip recommendation. Expert Syst. Appl. 2023, 226, 120145. [Google Scholar] [CrossRef]
  34. Nan, G.; Zhao, Y.; Fu, L.; Ye, Q. Object Detection by Channel and Spatial Exchange for Multimodal Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 8581–8593. [Google Scholar] [CrossRef]
  35. Yang, Z.; He, W.; Fan, X.; Tjahjadi, T. PlantNet: Transfer learning-based fine-grained network for high-throughput plants recognition. Soft Comput. 2022, 26, 10581–10590. [Google Scholar] [CrossRef]
  36. Mohammed, W.K.; Taha, M.A.; Jabar, H.D.; Alsaidi, S.A.A.A. Object Detection Techniques: A Review. Wasit J. Comput. Math. Sci. 2023, 2, 59–68. [Google Scholar] [CrossRef]
  37. Fu, Y. Recent Deep Learning Approaches for Object Detection. Highlights Sci. Eng. Technol. 2023, 31, 64–70. [Google Scholar] [CrossRef]
  38. Liu, H.; Ma, X.; Yu, Y.; Wang, L.; Hao, L. Application of Deep Learning-Based Object Detection Techniques in Fish Aquaculture: A Review. J. Mar. Sci. Eng. 2023, 11, 867. [Google Scholar] [CrossRef]
  39. Du, L.; Zhang, R.; Wang, X. Overview of two-stage object detection algorithms. J. Physics: Conf. Ser. 2020, 1544, 012033. [Google Scholar] [CrossRef]
  40. Zhang, S.; Wen, L.; Bian, X.; Lei, Z.; Li, S.Z. Single-Shot Refinement Neural Network for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  41. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  42. Nazir, A.; Wani, M.A. You Only Look Once—Object Detection Models: A Review. In Proceedings of the 2023 10th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 15–17 March 2023; pp. 1088–1095. [Google Scholar]
  43. Ge, X.; Chen, F.; Xu, S.; Tao, F.; Jose, J.M. Cross-modal Semantic Enhanced Interaction for Image-Sentence Retrieval. arXiv 2022, arXiv:2210.08908. [Google Scholar] [CrossRef]
  44. Xu, Y.; Zhang, M.; Fu, C.; Chen, P.; Yang, X.; Li, K.; Xu, C. Multi-modal Queried Object Detection in the Wild. arXiv 2023, arXiv:2305.18980. [Google Scholar] [CrossRef]
  45. Zhong, Y.; Yang, J.; Zhang, P.; Li, C.; Codella, N.; Li, L.H.; Zhou, L.; Dai, X.; Yuan, L.; Li, Y.; et al. RegionCLIP: Region-Based Language-Image Pretraining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16793–16803. [Google Scholar]
  46. Cheng, M.; Sun, Y.; Wang, L.; Zhu, X.; Yao, K.; Chen, J.; Song, G.; Han, J.; Liu, J.; Ding, E.; et al. ViSTA: Vision and Scene Text Aggregation for Cross-Modal Retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5184–5193. [Google Scholar]
  47. Li, L.H.; Zhang, P.; Zhang, H.; Yang, J.; Li, C.; Zhong, Y.; Wang, L.; Yuan, L.; Zhang, L.; Hwang, J.N.; et al. Grounded Language-Image Pre-training. arXiv 2021, arXiv:2112.03857. [Google Scholar] [CrossRef]
  48. Roy, A.M.; Bhaduri, J.; Kumar, T.; Raj, K. WilDect-YOLO: An efficient and robust computer vision-based accurate object localization model for automated endangered wildlife detection. Ecol. Inform. 2023, 75, 101919. [Google Scholar] [CrossRef]
  49. Verma, G.K.; Gupta, P. Wild Animal Detection Using Deep Convolutional Neural Network. In Proceedings of the 2nd International Conference on Computer Vision & Image Processing, Roorkee, India, 9–12 September 2017; Chaudhuri, B.B., Kankanhalli, M.S., Raman, B., Eds.; Springer: Singapore, 2018; pp. 327–338. [Google Scholar]
  50. Eikelboom, J.A.J.; Wind, J.; van de Ven, E.; Kenana, L.M.; Schroder, B.; de Knegt, H.J.; van Langevelde, F.; Prins, H.H.T. Improving the precision and accuracy of animal population estimates with aerial image object detection. Methods Ecol. Evol. 2019, 10, 1875–1887. [Google Scholar] [CrossRef]
  51. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv 2016, arXiv:1506.01497. [Google Scholar] [CrossRef]
  52. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar] [CrossRef]
  53. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. arXiv 2020, arXiv:2005.12872. [Google Scholar]
  54. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse R-CNN: End-to-End Object Detection with Learnable Proposals. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14449–14458. [Google Scholar] [CrossRef]
  55. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-aligned One-stage Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3490–3499. [Google Scholar] [CrossRef]
  56. Chen, S.; Sun, P.; Song, Y.; Luo, P. DiffusionDet: Diffusion Model for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 19773–19786. [Google Scholar]
  57. Zhou, X.; Girdhar, R.; Joulin, A.; Krähenbühl, P.; Misra, I. Detecting Twenty-thousand Classes using Image-level Supervision. arXiv 2022, arXiv:2201.02605. [Google Scholar] [CrossRef]
  58. Du, Y.; Wei, F.; Zhang, Z.; Shi, M.; Gao, Y.; Li, G. Learning to Prompt for Open-Vocabulary Object Detection with Vision-Language Model. arXiv 2022, arXiv:2203.14940. [Google Scholar] [CrossRef]
  59. Feng, C.; Zhong, Y.; Jie, Z.; Chu, X.; Ren, H.; Wei, X.; Xie, W.; Ma, L. PromptDet: Towards Open-Vocabulary Detection Using Uncurated Images. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  60. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
  61. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar] [CrossRef]
  62. Varghese, R.; M., S. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. Spatial distribution of key protected areas within Sichuan Wolong National Nature Reserve.
Figure 1. Spatial distribution of key protected areas within Sichuan Wolong National Nature Reserve.
Animals 14 02909 g001
Figure 2. The overall framework of CECS-CLIP. (a) Pre-training of the visual encoder. CLIP is utilized to match images with descriptions, employing contrastive learning to extract visual region representations from generated image–text pairs. (b) Text feature extraction process. After generating textual information, data cleaning is performed, and textual knowledge is extracted through a BiLSTM-CRF model to construct an expert knowledge base. (c) Species description processing. Species descriptions are input into the text encoder to extract key textual features. (d) Joint application of visual and textual encoders. The well-trained visual encoder is integrated with the textual encoder for the object detection task.
Figure 2. The overall framework of CECS-CLIP. (a) Pre-training of the visual encoder. CLIP is utilized to match images with descriptions, employing contrastive learning to extract visual region representations from generated image–text pairs. (b) Text feature extraction process. After generating textual information, data cleaning is performed, and textual knowledge is extracted through a BiLSTM-CRF model to construct an expert knowledge base. (c) Species description processing. Species descriptions are input into the text encoder to extract key textual features. (d) Joint application of visual and textual encoders. The well-trained visual encoder is integrated with the textual encoder for the object detection task.
Animals 14 02909 g002
Figure 3. Visual–linguistic pretraining: Transforming image–text pairs into region–text pairs using the pretrained CLIP language encoder and Region Proposal Network (RPN) for image region alignment.
Figure 3. Visual–linguistic pretraining: Transforming image–text pairs into region–text pairs using the pretrained CLIP language encoder and Region Proposal Network (RPN) for image region alignment.
Animals 14 02909 g003
Figure 4. Network structure of concept enhancement module.
Figure 4. Network structure of concept enhancement module.
Animals 14 02909 g004
Figure 5. Average precision results of CECS-CLIP and other models on the wild dataset by species.
Figure 5. Average precision results of CECS-CLIP and other models on the wild dataset by species.
Animals 14 02909 g005
Figure 6. Average precision results of CECS-CLIP and other models on the web dataset by species.
Figure 6. Average precision results of CECS-CLIP and other models on the web dataset by species.
Animals 14 02909 g006
Figure 7. Confusion matrix of CECS-CLIP and the other models on the wild dataset.
Figure 7. Confusion matrix of CECS-CLIP and the other models on the wild dataset.
Animals 14 02909 g007
Figure 8. Experimental results of CECS-CLIP and other models based on P, mAP50, and mAP50:90 metrics across the web and wild datasets.
Figure 8. Experimental results of CECS-CLIP and other models based on P, mAP50, and mAP50:90 metrics across the web and wild datasets.
Animals 14 02909 g008
Figure 9. Partial results of our method generalization test.
Figure 9. Partial results of our method generalization test.
Animals 14 02909 g009
Figure 10. Experimental results of our method compared with other models.
Figure 10. Experimental results of our method compared with other models.
Animals 14 02909 g010
Figure 11. Experimental results of our method compared with other multimodal models.
Figure 11. Experimental results of our method compared with other multimodal models.
Animals 14 02909 g011
Figure 12. Comparative experimental results of different weights.
Figure 12. Comparative experimental results of different weights.
Animals 14 02909 g012
Figure 13. Visualization results of the improved model.
Figure 13. Visualization results of the improved model.
Animals 14 02909 g013
Table 1. Comparison of CECS-CLIP model with other models, including single-stage object detection models (FCOS), two-stage object detection models (Fast RCNN), novel object detection models (DETR, Sparse R-CNN, TOOD, DiffusionDet), and multimodal object detection models (RegionCLIP, DeTIC, DetPro, and PromptDet)—experimental results.
Table 1. Comparison of CECS-CLIP model with other models, including single-stage object detection models (FCOS), two-stage object detection models (Fast RCNN), novel object detection models (DETR, Sparse R-CNN, TOOD, DiffusionDet), and multimodal object detection models (RegionCLIP, DeTIC, DetPro, and PromptDet)—experimental results.
ModelBackboneAP (web)AP (wild)AP50 (web)AP50 (wild)
Faster R-CNN [51]r500.5540.6980.8510.931
r1010.5880.7100.8630.933
FCOS [52]r500.4060.7010.6090.919
r1010.4220.7230.6230.933
DETR [53]r500.5610.3970.7160.61
Sparse R-CNN [54]r500.6500.7340.8330.945
r1010.6830.7440.8390.945
TOOD [55]r500.6850.7510.8510.938
r1010.6960.7580.8560.941
DiffusionDet [56]r500.7070.7610.8650.941
r1010.6980.7620.850.954
RegionCLIP [45]r500.6960.5590.8250.776
Detic [57]r500.7210.5440.8460.743
DetPro [58]r500.7490.7130.8670.824
PromptDet [59]r500.70.6210.8280.802
Oursr500.8890.9580.9050.976
Table 2. Performance metrics of YOLO and our method on the wild dataset.
Table 2. Performance metrics of YOLO and our method on the wild dataset.
ModelPDRFPRMR
YOLOv5s [24]0.9360.9220.0680.078
YOLOv6 [60]0.9250.9100.0810.090
YOLOv7s [61]0.9490.9250.0540.075
YOLOv8s [62]0.9470.8970.0560.103
Ours0.9520.9180.0500.082
Table 3. Performance metrics of YOLO and our method on the web dataset.
Table 3. Performance metrics of YOLO and our method on the web dataset.
ModelPDRFPRMR
YOLOv5s [24]0.8140.7630.2290.237
YOLOv6 [60]0.7760.7200.2890.280
YOLOv7s [61]0.8880.8200.1260.180
YOLOv8s [62]0.8370.7680.1950.232
Ours0.8950.8500.1170.150
Table 4. Fine-grained metric experimental results of CECS-CLIP compared with other models: APs, APm, and APl.
Table 4. Fine-grained metric experimental results of CECS-CLIP compared with other models: APs, APm, and APl.
ModelAPsAPmAPl
YOLOv60.040.2740.605
RegionCLIP0.2820.060.582
Ours0.2870.5270.613
Table 5. Comparison of ablation experiments.
Table 5. Comparison of ablation experiments.
ModelAPAP50APsAPmAPl
RegionCLIP0.5590.7760.2820.060.582
RegionCLIP+CE0.7960.8540.2850.5340.60
RegionCLIP+CS0.8560.8910.2810.4620.59
RegionCLIP+CE+CS0.9580.9760.2870.5270.613
Table 6. Comparative analysis of experimental results under different weight settings and comparative study with RegionCLIP fine-tuning method.
Table 6. Comparative analysis of experimental results under different weight settings and comparative study with RegionCLIP fine-tuning method.
Weight ParameterIterAPAP50
0.510,0000.2140.5
0.710,0000.3820.694
0.910,0000.5740.859
RegionCLIP fine-tuning10,0000.2960.596
RegionCLIP fine-tuning20,0000.4220.728
RegionCLIP fine-tuning30,0000.5590.776
Table 7. Comparative study of accuracy results and training time with different numbers of nodes.
Table 7. Comparative study of accuracy results and training time with different numbers of nodes.
IterAPAP50
50000.4620.79
10,0000.5730.859
15,0000.6940.892
20,0000.7370.929
30,0000.9580.976
Table 8. Statistics on the number of correct identifications of rare wildlife species and blank identifications in robustness testing.
Table 8. Statistics on the number of correct identifications of rare wildlife species and blank identifications in robustness testing.
SpeciesActualCorrect EstimateCorrect Rate
Giant panda35351.000
Red panda25230.920
Tibetan macaque28240.857
Golden snub-nosed monkey19170.895
Porcupine27250.926
Tibetan macaque20180.900
Sambar30160.533
Tufted deer18160.888
Chinese serow24220.917
Blue sheep23180.783
All2492140.859
Empty shots50460.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, F.; Hu, C.; Liang, A.; Wang, S.; Su, Y.; Xu, F. CECS-CLIP: Fusing Domain Knowledge for Rare Wildlife Detection Model. Animals 2024, 14, 2909. https://doi.org/10.3390/ani14192909

AMA Style

Yang F, Hu C, Liang A, Wang S, Su Y, Xu F. CECS-CLIP: Fusing Domain Knowledge for Rare Wildlife Detection Model. Animals. 2024; 14(19):2909. https://doi.org/10.3390/ani14192909

Chicago/Turabian Style

Yang, Feng, Chunying Hu, Aokang Liang, Sheng Wang, Yun Su, and Fu Xu. 2024. "CECS-CLIP: Fusing Domain Knowledge for Rare Wildlife Detection Model" Animals 14, no. 19: 2909. https://doi.org/10.3390/ani14192909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop