Next Article in Journal
AHEAD: A Novel Technique Combining Anti-Adversarial Hierarchical Ensemble Learning with Multi-Layer Multi-Anomaly Detection for Blockchain Systems
Previous Article in Journal
Contextual Intelligence: An AI Approach to Manufacturing Skills’ Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Federated Learning with Multi-Method Adaptive Aggregation for Enhanced Defect Detection in Power Systems

1
State Grid Sichuan Electric Power Research Institute, Power Internet of Things Key Laboratory of Sichuan Province, Chengdu 610094, China
2
School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu 610031, China
*
Authors to whom correspondence should be addressed.
Big Data Cogn. Comput. 2024, 8(9), 102; https://doi.org/10.3390/bdcc8090102
Submission received: 25 July 2024 / Revised: 24 August 2024 / Accepted: 27 August 2024 / Published: 2 September 2024

Abstract

:
The detection and identification of defects in transmission lines using computer vision techniques is essential for maintaining the safety and reliability of power supply systems. However, existing training methods for transmission line defect detection models predominantly rely on single-node training, potentially limiting the enhancement of detection accuracy. To tackle this issue, this paper proposes a server-side adaptive parameter aggregation algorithm based on multi-method fusion (SAPAA-MMF) and formulates the corresponding objective function. Within the federated learning framework proposed in this paper, each client executes distributed synchronous training in alignment with the fundamental process of federated learning. The hierarchical difference between the global model, aggregated using the improved joint mean algorithm, and the global model from the previous iteration is computed and utilized as the pseudo-gradient for the adaptive aggregation algorithm. This enables the adaptive aggregation to produce a new global model with improved performance. To evaluate the potential of SAPAA-MMF, comprehensive experiments were conducted on five datasets, involving comparisons with several algorithms. The experimental results are analyzed independently for both the server and client sides. The findings indicate that SAPAA-MMF outperforms existing federated learning algorithms on both the server and client sides.

1. Introduction

Ensuring a stable power supply through the power grid is a crucial factor in maintaining quality of life and the seamless functioning of society. Timely detection of faults and defects in transmission lines is essential for maintaining the stability of power system operations. Consequently, transmission line defect detection has emerged as a research hotspot, garnering significant attention from both academia and the power industry, and has been extensively studied by numerous scholars.
In recent years, the rapid advancement of deep learning has prompted researchers to employ it for object detection, leading to the development of numerous deep learning-based object detection algorithms. For insulator defect detection, Miao et al. [1] employed transfer learning strategies, utilizing the COCO (Common Objects in Context) pre-trained model for domain adaptation to construct a Single Shot MultiBox Detector (SSD) model that incorporates various insulator types using diverse backgrounds. Zhao et al. [2] proposed an improved insulator detection method based on Faster R-CNN, fine-tuning the model and introducing an enhanced anchor box generation technique to improve detection precision and significantly enhance performance in handling occlusion. Feng et al. [3] introduced an automatic insulator detection approach utilizing the YOLOv5x object detection model in combination with the K-means algorithm. This approach optimizes model performance, facilitating the rapid and accurate detection and localization of targets in images. For detecting foreign objects in bird nests, Ju et al. [4] proposed a real-time bird nest detection approach leveraging the spatial relationship between bird nests and tower poles. This approach employs SSD as the detector and utilizes deep neural networks to establish the spatial relationship between tower poles and bird nests. By leveraging tower pole detection results to refine bird nest detection, this approach enhances both the accuracy and efficiency of bird nest identification.
Although the aforementioned detection models have achieved considerable success in identifying defects in transmission lines, their training methodologies primarily rely on single-node systems. Training defect detection models on a single node restricts effective data sharing and connectivity, hindering comprehensive integration and analysis of data from diverse environments. This, in turn, affects the diversity and generalization capabilities of transmission line defect detection models. Furthermore, single-node training raises concerns regarding privacy security and the inability to handle large-scale data [5]. In 2016, Google Research introduced the concept of federated learning to enhance privacy protection and address the challenge of data silos [6]. The primary goal of federated learning is to enable model training without centrally storing or transmitting raw data. Due to its superior performance and enhanced security compared to traditional distributed machine learning, federated learning has rapidly become a prominent paradigm in the field [7]. Hong et al. [8] proposed a joint adversarial debiasing method that allows users to opt out of the debiasing session based on privacy or computational cost concerns without accessing sensitive group information. This method addresses the issue of model bias and unfairness arising from the heterogeneity of user data in federated learning. Karimireddy et al. [9] proposed the stochastic control averaging algorithm, which utilizes control variables to reduce drift between clients during local updates, thereby addressing the issues of unstable and slow model convergence. Yuan et al. [10] proposed the joint pairwise averaging algorithm, which overcomes the limitations of traditional averaging by introducing an innovative server-side pairwise averaging procedure. This approach effectively improves convergence speed and optimization in federated learning tasks involving non-smooth regularization terms. Traditional federated learning algorithms have achieved progress by optimizing parameter aggregation and improving communication efficiency, but the aggregated global model continues to suffer from poor performance and generalization. Most traditional federated learning algorithms optimize parameter aggregation from a single perspective, overlooking the complementary advantages of integrating different algorithms, which limits model performance.
To tackle these challenges, this paper introduces a Server-side Adaptive Parameter Aggregation Algorithm Based on Multi-method Fusion (SAPAA-MMF). This approach takes into account data silos, privacy security, and hardware performance issues inherent in defect detection tasks dependent on single-node training models. The study constructs a transmission line defect detection network utilizing the single-stage object detection algorithm SSD as the client model for local training within a federated learning framework. By leveraging the complementarity of algorithms, the adaptive aggregation algorithm FedAdam [11] serves as the base algorithm, which is combined with the improved federated averaging algorithm FedAvg [11] to enhance server-side parameter aggregation in federated learning. This approach yields a more effective global model.
  • The SAPAA-MMF method integrates various algorithms by utilizing the average model variance as the pseudo-gradient for adaptive aggregation. This approach addresses the limitations of previous single-parameter aggregation algorithms, which often struggled to ensure both the convergence and stability of the global model. By leveraging the average model variance, SAPAA-MMF enables a more robust and stable parameter aggregation process, resulting in enhanced performance and generalization of the global model in federated learning scenarios.
  • In this paper, we propose a weight balancing method that comprehensively considers factors such as client-side local data distribution and data quality in the aggregation process. This method ensures that different training nodes or devices can contribute effectively to the global model, thereby maximizing the advantages of various algorithms. By accounting for the variability in local data and the quality of client contributions, the weight balancing method improves the aggregation process, resulting in a more accurate and robust global model in federated learning.
  • The SAPAA-MMF method proposed in this paper significantly improves defect detection in transmission lines. Extensive experiments against several state-of-the-art methods demonstrate the effectiveness and superiority of SAPAA-MMF. The results indicate that SAPAA-MMF outperforms existing methods by providing more accurate and reliable defect detection, thereby confirming its potential and advantages in practical applications.

2. Related Work

2.1. Defect Detection

In the traditional field of image processing research, researchers mainly use artificially designed feature extraction algorithms to build models for visual characteristics such as color and morphological edges, aiming to effectively distinguish key objects from complex background environments, performing various defect detection tasks accordingly. Wu et al. [12] proposed a new active contour model, which uses a semi-local operator to extract texture features, defines a new function to handle texture inhomogeneity, and uses a fast dual form to make contour evolution more efficient. The model can extract uneven insulators from aerial images and overcome the recognition difficulties caused by texture inhomogeneity. Using only a single feature cannot significantly improve the detection effect. Therefore, Wang et al. [13] proposed a recognition method that integrates the shape, color, and texture information of insulators. First, parallel line features are perceived from different directions as candidate features of the insulators. Then, the candidate region is expanded and semantic analysis is performed using local binary patterns to identify the insulator region. Finally, the main color component analysis is used to compensate for the insulator region, the region is adaptively divided according to the average distance, the texture feature changes are analyzed, and the insulator shedding defect is detected. Considering that image noise can also affect the detection results, Yan et al. [14] proposed a maximum inter-class variance algorithm based on morphological methods, designed a new filtering method to remove tiny noise according to the characteristics of power lines, and successfully identified the redundant materials on the power lines by comparing the number of local maxima in the measured power lines and Hough transfer accumulators through the Hough transform feature. Chen et al. [15] addressed the limitations of traditional pavement defect detection methods by proposing MANet, a multi-scale mobile attention-based network that integrates multi-scale convolutions and hybrid attention mechanisms to enhance feature extraction and improve the accuracy of pavement defect detection. Yu et al. [16] addressed the limitations of using methods like graph convolutional networks (GCN) by constructing high-quality graphs with multi-source sensors and utilizing a two-stage framework for fault diagnosis, thereby enhancing the robustness of fault diagnosis.
Although the above algorithms have achieved some results, they have limitations in feature acquisition and real-time performance, which limits the accuracy and efficiency of detection. In recent years, with the rapid development of deep learning [17], researchers have adopted deep learning for target detection, and many deep learning-based target detection algorithms have emerged. The target detection algorithm based on deep learning has achieved good results in image feature extraction and complex and diverse image features. This type of algorithm can automatically identify the features in an image and overcomes the shortcomings of traditional algorithms. In insulator defect detection, Miao et al. [1] borrowed the transfer learning strategy, made domain adaptability adjustments through the COCO (Common Objects in Context) pre-trained model, and constructed an SSD (Single Shot MultiBox Detector) detection model that includes multiple insulator types using multivariate backgrounds. On this basis, the initial model is fine-tuned using training data from specific insulator samples and their complex application scenarios to accurately identify insulators. Zhao et al. [2] proposed an insulator detection method based on an improved Faster R-CNN. By fine-tuning the Faster R-CNN model and using an improved anchor box generation method, the accuracy of insulator detection is improved while significant improvements are made in handling occlusion. Feng et al. [3] proposed an automatic insulator detection method based on the YOLOv5x target detection model. This method combines the YOLOv5x model and the K-Means algorithm to achieve the purpose of automatically identifying insulator defects. The performance of the model is optimized by K-means clustering. The YOLOv5x model can quickly and accurately detect and locate targets in the image. In the defect detection of bird nest foreign bodies, Ju et al. [4] proposed a real-time bird nest detection method based on the relationship with the tower pole. This method uses SSD as a detector, uses a deep neural network to establish the relationship between the tower pole and the bird nest, and uses the tower pole detection results to correct the bird nest detection effect, thereby improving the accuracy and efficiency of bird nest detection. Li et al. [18] proposed an automatic detection method for bird nests in power transmission lines based on Faster R-CNN. By magnifying the bird nest image, it not only solves the problem of insufficient training samples and overfitting of the neural network classifier, but also has a good bird nest target detection effect in complex environments. Li et al. [19] addressed the limitations of traditional PV panel defect detection methods by incorporating Ghost convolution with BottleneckCSP and a tiny target prediction head into YOLOv5, significantly enhancing detection accuracy and speed for multiscale PV panel defects. Wang et al. [20] addressed the limitations of traditional road defect detection methods by enhancing YOLOv8 with BiFPN and LSK-attention mechanisms, improving both detection accuracy and model efficiency for real-time applications.
Although the aforementioned detection models have achieved good results in identifying defects in transmission lines, their training methods are primarily based on single nodes. However, in practical applications, transmission lines are widely distributed, and defect data is dispersed across equipment or systems in different regions. The single-node training method for defect detection models leads to a lack of effective sharing and connection between data, preventing comprehensive integration and analysis of data from different environments, thus affecting the data diversity and generalization ability of transmission line defect detection models. Additionally, there are issues related to privacy security and the inability to train large-scale data on a single node.

2.2. Federated Learning

The goal of federated learning is to build a global model by aggregating local updates from multiple clients to adapt to the overall data distribution. The client shares global model parameters with other clients during local training, and the server aggregates local updates to form a global model.
McMahan et al. [6] were the first to propose the concept of federated learning. In response to the needs of massive data training and privacy and security protection, they advocated distributing training data to mobile devices and using local calculated updates for shared model learning. They proposed a practical method for deep network federated learning based on the average iteration model, and the global model was obtained by calculating the mean of the parameters. Blanchard et al. [21] proposed the Krum aggregation algorithm, the core idea of which is to judge the abnormal behavior of nodes based on the similarity between weights, identify abnormal nodes by calculating the distance between weights, and exclude these abnormal nodes to obtain a reliable aggregation result. Yin et al. [22] proposed a robust distributed gradient descent algorithm based on median and truncated mean operations, which enables the algorithm to better solve the Byzantine problem [22] and has good robustness. Pan et al. [23] started by reducing the frequency of federated learning communication, and optimized the balance between local updates and global parameter aggregation by intelligently selecting the number of local updates from the client, thereby reducing the communication frequency between the client and the server. Li et al. [24] added a proximal term to the loss function of the client model to penalize any client model that deviates too much from the global model, so as to prevent the client model from deviating too much from the global model and ensure the convergence of the global model. Li et al. [25] proposed model contrast federated learning by using improved cross entropy loss to perform local optimization and solve the inconsistency problem between foreground and background and by using the similarity between model representations to correct the local training of each party. Zhu et al. [26] proposed a data-free knowledge distillation method to solve heterogeneous federated learning. This method integrates client information without using actual data by training a lightweight generator on the server side and then broadcasting it to the client, adjusting the client’s training process. Reddi et al. [11] proposed an adaptively optimized federated learning algorithm, combined with an adaptive optimization strategy, to achieve effective learning and optimization of the global model on the server side, thereby improving the performance of federated learning. Fallah et al. [27] proposed combining federated learning algorithms with a model-agnostic meta-learning framework. In this approach, models are initialized via meta-learning, and clients perform one or more gradient descent steps on their local datasets to adapt the models. This method retains the advantages of the federated learning architecture while providing more personalized models for each client. Li et al. [28] proposed a private model inheritance scheme that fuses a client’s historical personalized models to guide the subsequent personalized adaptation of the global model, effectively leveraging prior personalized results. Chen et al. [29] proposed a trunk patch-based federated learning architecture, where the shared trunk refines the common knowledge of all participants, and the private patch serves as a compact and efficient module to retain each participant’s domain-specific information, thereby improving model performance. Niu et al. [30] proposed an activation function-based regularizer to correct the gradient of category embeddings by accurately injecting cross-client gradient terms, starting from backpropagation for gradient correction. Collins et al. [31] proposed shared feature representations with local head learning algorithms, which utilize the computational power of distributed clients to perform multiple rounds of local low-dimensional parameter updates. This approach addresses the issues of learning inefficiency and limited model generalization caused by data heterogeneity in federated learning. Shen et al. [32] proposed a novel federated learning paradigm called joint mutual learning to address the issue of data heterogeneity. This paradigm allows each client to independently train personalized models and obtain local models for similar but distinct tasks through collaborative training. Additionally, it enables clients to design customized models for different scenarios and tasks. Shi et al. [33] investigated the challenge of balancing generalization and personalization in blended learning and proposed the personalized joint upper confidence bound algorithm. This algorithm optimally selects the exploration length to balance learning a local model with providing global information for the blended learning objective.
The federated learning algorithm has made some achievements in terms of parameter aggregation effect and communication efficiency, but the aggregated global model still has problems such as poor model performance and the generalization effect. Most traditional federated learning algorithms only optimize the parameter aggregation effect from a single perspective, ignoring the complementary advantages of different algorithm fusions, which limits the model performance. Starting from the perspective of multi-method fusion, this paper makes full use of the complementary advantages of different algorithms and designs a multi-method fusion adaptive aggregation parameter algorithm to improve the performance and generalization ability of the global model.

3. Proposed Method

This section is organized into subsections and offers a comprehensive description of the proposed SAPAA-MMF structure, the objective function, and the specific design elements. Additionally, it presents the pseudo-code along with an analysis of the model’s time and space complexity.

3.1. Overall Algorithm Architecture

The SAPAA-MMF algorithm is designed to ensure that the global model aggregated by the server demonstrates robust performance and generalization capabilities. This paper leverages the complementary strengths of algorithms by integrating the improved joint mean algorithm with the adaptive aggregation algorithm. By combining the broad applicability and high performance of the joint mean algorithm with the adaptability and stability of the adaptive aggregation algorithm, the SAPAA-MMF algorithm further improves the performance and generalization capabilities of the global model. The overall architecture of SAPAA-MMF is depicted in Figure 1.
The SAPAA-MMF algorithm operates within the federated learning framework developed in this study. As illustrated in Figure 1, each client utilizes its local dataset to perform distributed synchronous training, adhering to the standard federated learning process. After each training round, the aggregated global model is disseminated to the clients as updated model parameters, and iterative training continues until the termination condition is satisfied. During server-side aggregation, the hierarchical difference between the global model aggregated by the improved joint mean algorithm and the global model from the previous round is calculated. This difference is utilized as the pseudo-gradient for the adaptive aggregation algorithm, enabling it to generate a new global model with improved performance.

3.2. Objective Function

The local client device in federated learning is responsible solely for storing and processing data. Under this condition, it uploads the client model parameters to the server, which aggregates them to generate a single global model parameter w. In this distributed setting, the global optimization problem involves finding the set of parameters w that minimizes the average loss across all client devices participating in the training process. Thus, the global optimal objective function F ( w ) on the server side can be expressed as
min w F ( w ) = 1 K k = 1 K F k ( w ) .
Here, K represents the total number of client devices participating in the joint training and F k ( w ) denotes the local objective function of the k-th client, which is expressed as
F k ( w ) = 1 n k i d k f i ( w ) .
Here, d k represents the local dataset of the k-th client device, n k indicates the size of the k-th client’s data, and f i ( w ) refers to the loss value of the local client defect detection model SSD on sample i. The corresponding loss function f ( w ) is expressed as
f ( w ) = 1 N ( L c o n f ( w , x , c ) + λ L l o c ( w , x , l , g ) ) ,
where x is the matching indicator variable, c is the confidence score of each category predicted by the network, l is the offset of each default box predicted by the network relative to its original shape, and g represents the parameters of the real box, including the center coordinates ( c x , c y ) , width ( w ) , and height ( h ) . N is the number of default boxes corresponding to the matched real box, L c o n f is the classification loss, L l o c is the positioning loss, and λ is a hyperparameter that balances the weights between classification and positioning losses. L c o n f and L l o c are calculated using Equation (4) and Equation (6), respectively. In Equation (3),
L c o n f ( w , x , c ) = i P o s N x i j p log c ^ i p i N e g log c ^ i 0 ,
where x i j p = { 1 , 0 } indicates whether the i-th default box matches the real bounding box of category p. In Equation (4),
c ^ i p = exp c i p p exp c i p
indicates the probability that the i-th default box belongs to category p.
L l o c ( w , x , l , g ) = i P o s N m { c x , c y , w , h } x i j f smooth L 1 l i m g ^ j m ,
where x i j f = { 1 , 0 } indicates whether the i-th default box matches the j-th real bounding box. In Equation (6), smooth L 1 ( · ) is the smooth L1 loss function used to calculate the difference in positional parameters between the bounding box predicted by the network and the real bounding box.

3.3. Algorithm Principle

The SAPAA-MMF algorithm integrates two algorithms to fully leverage the complementary strengths of different approaches, thereby enhancing the performance of the global model. This process is divided into two stages of aggregation. In the first stage, an improved joint mean algorithm is used for aggregation, followed by the calculation of hierarchical differences between the aggregated global model and the previous round’s global model. The objective is to leverage the adaptability and strong performance of the improved joint mean algorithm, fully accounting for the differences in client data and local training states, to obtain a high-performance one-stage global model. In the second stage, the hierarchical model differences from the first stage serve as pseudo-gradients for adaptive aggregation. The aim is to exploit the strong convergence and stability of adaptive aggregation algorithms to further enhance the stability and performance of the global model. Due to the use of joint mean and adaptive aggregation algorithms in multi-method fusion, this section primarily provides a detailed introduction and analysis of the joint mean algorithm, adaptive aggregation algorithm, and server-side adaptive parameter aggregation algorithm used in multi-method fusion.

3.3.1. Joint Mean Algorithm

McMahan et al. [6] proposed the joint mean algorithm FedAvg, applicable to horizontal federated learning. This algorithm accounts for the influence of data samples on aggregation performance, adapts to various environments, and fully utilizes the computing resources of diverse devices. By aggregating multiple local models, this algorithm enhances the model’s generalization ability and improves its robustness.
In the FedAvg algorithm, the server aggregates the model parameters uploaded by the clients, while the clients train the model on their local datasets. The detailed training process is as follows: First, the server initiates the global model’s initialization process and distributes the initialized model to various clients. Then, the server and clients jointly enter the iterative training stage. In the t + 1 round of federated learning, the server selects K clients from all available clients for local model training and sends the aggregated global model w t from the t-th round to the selected clients. After receiving the global model, the clients train on their local datasets according to Equation (7). Finally, after local training, the local model parameters w t + 1 k are generated as follows:
w t + 1 k = w t k η g k .
Here, η represents the learning rate, and g k denotes the model gradient. After all client models complete training, the model parameters are uploaded to the server, which aggregates the parameters from each client model according to
w t + 1 = k = 1 K n k n · w t + 1 k .
Here, n represents the total number of client samples and n k denotes the number of samples for the k-th client. After server-side aggregation, a global model is generated and distributed to each client for subsequent rounds of training until the model converges or meets the termination condition.

3.3.2. Adaptive Aggregation Algorithm

The FedAdam algorithm, proposed by Reddi et al. [11], is a federated learning algorithm. It is an improved federated learning algorithm that utilizes the Adam optimizer [34] within the federated learning framework. It primarily addresses the issues of slow convergence, low accuracy, and poor stability encountered by traditional federated learning algorithms during model training.
The FedAdam algorithm is adapted from the Adam algorithm framework to accommodate the unique characteristics of federated learning. This algorithm retains the concept of adaptive learning rates from the Adam algorithm, dynamically adjusting the learning rate through first-order and second-order moment estimations to accelerate model convergence and improve stability. The formula for calculating the first-order moment in the FedAdam algorithm is:
m t = β 1 m t 1 + ( 1 β 1 ) Δ t
Here, t denotes the training round and β 1 represents the decay rate of first-order moment estimation. First-order moment estimation simulates the effect of momentum, helping to stabilize model parameter updates. Here, Δ t represents the model pseudo-gradient. In the FedAdam algorithm, the average model difference for each client is computed as:
Δ t k = w t k w t .
Here, Δ t k represents the difference between the k-th client and the global model w t in round t. Based on the above equation, the average model difference Δ t can be obtained as:
Δ t = 1 | K | i = 1 K Δ t k .
The calculation formula for second-order moment estimation in the FedAdam algorithm is:
v t = β 2 v t 1 + ( 1 β 2 ) ( Δ t ) 2 .
Here, β 2 represents the decay rate of second-order moment estimation. By incorporating second-order moment estimation, the algorithm can more flexibly respond to gradient changes in the parameter space. In some cases, gradients can undergo significant changes, and second-order moment estimation captures this trend, facilitating learning rate adjustments and enhancing the algorithm’s robustness.
Finally, the formula for updating the model parameters is
w t + 1 = w t + σ m t v t + τ .
Here, σ represents the learning rate, controlling the magnitude of parameter updates, while τ is a small constant added for numerical stability to prevent division by zero.

3.3.3. A Server-Side Adaptive Parameter Aggregation Algorithm for Multi-Method Fusion

The first two sections provide a comprehensive introduction to the core concepts of the FedAvg and FedAdam algorithms. The FedAvg algorithm primarily addresses data distribution differences, offering strong effectiveness and adaptability, while the FedAdam algorithm focuses on improving model convergence speed and stability. Therefore, this paper posits that integrating these two algorithms can harness their complementary strengths, further enhancing the effectiveness of federated learning.
This paper proposes a server-side adaptive parameter aggregation algorithm, SAPAA-MMF, based on multi-method fusion. In the adaptive aggregation algorithm FedAdam, the pseudo-gradient Δ t is calculated by averaging model differences, as shown in Equations (10) and (11). This paper posits that using average model differences as pseudo-gradients for adaptive aggregation overlooks the impact of client data samples on the global model. Since the number of data samples is a key factor in determining the contribution of each client, overlooking client data samples can result in weight allocation bias. For instance, clients with a large number of data samples may be undervalued, while clients with fewer data samples may be overvalued, leading to weight bias that can affect the accuracy of model training. Furthermore, overlooking client data samples can also diminish the model’s generalization ability. The diversity of data samples is crucial to the model’s generalization ability, and neglecting this factor may lead to poor model performance on new data. Therefore, during the model aggregation process, the influence of the number of client samples on the aggregation results must be carefully considered.
To address the above issues and integrate the advantages of different algorithms, this chapter adopts FedAdam as the base algorithm, incorporates the weighted client data samples from the FedAvg algorithm, and proposes a new model hierarchical difference as the pseudo-gradient Δ t in the FedAdam algorithm:
Δ t = [ w a v g , 1 w t , 1 , w a v g , 2 w t , 2 , , w a v g , i w t , i ] .
Here, i represents the number of model layers, w a v g denotes the global model aggregated using the joint mean algorithm in the first stage, and w t is the global model from the previous round. The detailed calculation of Δ t is presented in Algorithm 1.
Algorithm 1: Calculation of hierarchical differences in models
Bdcc 08 00102 i001
Using hierarchical model differences that account for data samples as pseudo-gradients can effectively resolve the weight bias and generalization decline caused by average model differences. However, there are still challenges in calculating w a v g . In the calculation of w a v g , only the number of client samples is considered, meaning that the more samples a client has, the greater the client’s influence on the global model aggregation. If a client’s model parameters are of poor quality but the client has a large number of samples, this will negatively impact the performance of the global model.
In real-world federated learning, various factors influence client samples, including not only sample size but also differences in data distribution and quality. Therefore, the aggregation method for w a v g can be further improved to fully account for the impact of factors such as client local data distribution and quality on the aggregation results. Since no direct indicator exists to measure data distribution and quality, the loss during client training is used as an indirect indicator to assess the local data distribution and quality factors of the client, further optimizing the calculation of
w a v g = k = 1 K 1 θ · n k n + θ · L k · w t + 1 k .
Here, L k represents the proportion of losses incurred by the k-th client relative to all clients. This improvement considers both the impact of local data samples and the loss incurred during training on the aggregation results. The parameter θ adjusts the ratio between sample size and loss. When θ equals 0, the calculation of w a v g is equivalent to the original FedAvg algorithm. L k is calculated using Equation (16). In Equation (15),
L k = e l k i = 1 K e l i ,
where l i represents the model loss for the k-th client trained on the local dataset. This formula converts a set of losses into a probability distribution with values ranging between 0 and 1. For the exponential term e l k , taking the negative value of the input l k ensures that smaller l k values approach 1 after the exponential operation, while larger l k values approach 0. Accordingly, for larger losses, if the model’s learning performance is poor, the proportion of losses should be small, whereas for smaller losses, if the model’s learning performance is good, the proportion of losses should be large.
Weight decay is a technique used to regularize neural networks, with its effects primarily manifesting in regularization, preventing overfitting, enhancing robustness, simplifying model structure, and improving convergence stability. By introducing weight decay, the model can learn simpler and smoother representations, thereby enhancing its generalization ability and robustness. Weight decay is introduced in the adaptive aggregation algorithm to prevent the global model parameters from becoming too large after server-side aggregation, thereby reducing the risk of overfitting and improving the stability and generalization performance of the global model. The weight decay penalty is introduced in the adaptive aggregation algorithm updates to control the complexity of the model. The optimized model update is given by:
w t + 1 = w t + σ m t v t + ϵ λ w t .
Here, λ represents the weight decay term, which is a non-negative real number typically set to 0.0001 during training.
Based on the aforementioned algorithm steps and fundamental principles, a server-side adaptive parameter aggregation algorithm for multi-method fusion has been implemented. The detailed steps of the algorithm are presented in Algorithm 2. Algorithm 3 provides the client-side update procedure, which is used to receive the global model and perform updates and local training.
Algorithm 2: A server-side adaptive parameter aggregation algorithm for multi-method fusion (SAPAA-MMF)
Bdcc 08 00102 i002
Algorithm 3: Client update algorithm
Bdcc 08 00102 i003

3.3.4. Algorithm Flow

Combining the SAPAA-MMF algorithm with the basic process of federated learning, the specific algorithm process is illustrated in Figure 2. Initially, the server selects k clients for joint training. Each client loads its local dataset and waits for the initialization of the model. Once the server initializes the model, it distributes the global model to each client. After receiving the initialized global model, clients load it and use it as their initial model parameters for training. After the initialization phase, the training proceeds according to the specified number of rounds. If training reaches the specified number of rounds, the joint training task ends. Otherwise, each client uses its local dataset for training and waits for all k clients to complete their training, after which the server aggregates the k models. In the SAPAA-MMF algorithm, the improved joint mean algorithm is first used to aggregate and obtain a one-stage global model. The hierarchical model difference is subsequently calculated based on the previous round’s global model. This hierarchical model difference is used as a pseudo-gradient to calculate the first-order moment estimation and the second-order moment estimation. A two-stage global model is generated through adaptive aggregation, and the global model parameters for the current round are stored. This process constitutes one round of SAPAA-MMF training. Iterative training continues until the specified number of rounds is reached or the model converges, at which point the training concludes.

3.3.5. Algorithm Complexity Analysis

SAPAA-MMF is a federated learning algorithm that operates within the federated learning framework, with its primary task being model aggregation. Therefore, this section will not consider the complexity of the federated learning framework but will focus solely on analyzing the time and space complexity of the SAPAA-MMF algorithm. In terms of time complexity, the primary factor affecting the complexity of the SAPAA-MMF algorithm is the number of model layers, denoted as m. The algorithm first iterates through the model layers for one-stage aggregation, resulting in a time complexity of O ( m ) . Next, the hierarchical model difference is calculated. Since the model layers need to be matched, the time complexity of this step is O ( m 2 ) . Finally, the adaptive algorithm aggregates the two-stage model, with a time complexity of O ( m ) . Therefore, the overall time complexity of the SAPAA-MMF algorithm is O ( m 2 + 2 m ) , which simplifies to O ( m 2 ) . In terms of space complexity, the primary factor influencing the complexity of the SAPAA-MMF algorithm is the number of model layers, denoted as m. The algorithm requires loading all the models into the memory during execution. Therefore, the space complexity of the algorithm is O ( m ) .

4. Experiments

4.1. Experiments Setup

Datasets. This paper utilizes transmission line defect data as the experimental dataset, provided by the State Grid Sichuan Electric Power Company Electric Power Research Institute. After thorough screening and collation, the defect data has been meticulously marked and reviewed by professionals, ensuring its high quality and credibility, thereby accurately reflecting various transmission line defects in actual operation. The entire dataset is categorized based on defect characteristics, comprising four main types: bird’s nest foreign bodies, cement rod damage, shockproof hammer slip, and insulator self-explosion. To enable efficient detection of multiple defects using a single model, the four types of defect data are combined into a mixed dataset containing various defect types. Table 1 illustrates the details of all the datasets. Figure 3 and Figure 4 present sample images of the four defect types and the mixed dataset, respectively.
As shown in Table 1, federated learning involves multiple clients and requires distributed training. Therefore, this paper divides the five datasets into four parts, each of which is used as local data for individual clients. Client_4 is designated for global evaluation and does not participate in model training.
Software and hardware environment. The experimental environment is divided into a hardware environment and a software environment. The host using the Ubuntu operating system on the hardware uses system version number 18.04 and has a memory size of 86 G, a CPU of AMD EPYC 7402, a graphics card of NVIDIA A40 (19.5 TFLOPs), and a video memory size of 48G. The programming language of the software is python, version 3.8. The depth learning framework used is python, version 1.9. The programming IDE is Pycharm 2023.2.
Experimental framework. On the experimental framework, Flower is used as the basic framework for federated learning and the SSD network is used as the transmission line defect detection model to build a multi-client transmission line defect detection federated learning synchronous training framework.
Comparison Algorithms. The SAPAA-MMF algorithm proposed in this paper is used for federated learning, and compared with FedAvg [11], FedAvgM [35], FedYogi [11], FedAdam [11], FedAdagrad [11], FedMedia [22], TrimmedAvg [22], and Krum [21] algorithms.
Parameter settings. The initial learning rate of the SSD network is set to 0.002, the Batchsize is set to 2, 20 epochs are trained in each round, a total of 20 rounds, and 1200 epochs are trained for three clients. The initialization parameter of SAPAA-MMF algorithm θ is set to 0.003, β 1 is set to 0.9, β 2 is set to 0.99, and σ is set to 0.01.
Evaluation metrics. The experimental results are evaluated and compared separately for the server and client. Server evaluation measures the performance of the global model, while client evaluation assesses the performance of the client model. The client evaluation result is obtained by averaging all client evaluations. For evaluation metrics, five indicators— A P , A P 50 , A P 75 , A R 100 , and A R 300 —which effectively represent the model’s performance, are selected. These five indicators are standard metrics from the COCO dataset [36]. Detailed definitions of these evaluation metrics are provided in the next subsection.

4.2. Evaluation Metrics Definition

In this paper, a target detection model is used to identify transmission line defects and a federated learning model for defect detection is trained using the federated learning algorithm. Therefore, the evaluation metrics used in this paper are commonly employed in target detection and the effectiveness of the proposed federated learning algorithm is assessed using these metrics. The commonly used evaluation metrics in target detection include AP and AR values. These evaluation metrics are described below.
Intersection over Union ( I o U ) is a metric commonly used to evaluate the degree of overlap between bounding boxes in target detection. I o U calculates the ratio of the overlapping area between two bounding boxes to their union. It is typically used to evaluate the match between the predicted bounding box and the actual target box. The formula is as follows:
I o U = a r e a ( B p B g t ) a r e a ( B p B g t ) .
Here, B p represents the predicted box, B g t is the actual target box, and the I o U value ranges from 0 to 1.
Generally, in target detection, when the Intersection over Union ( I o U ) between the predicted box and the ground truth box exceeds the threshold of 0.5, the predicted box is considered a positive sample. Conversely, if the I o U is lower than 0.5, the predicted box is considered a negative sample. Under this standard, the instances that the model correctly identifies as positive samples are defined as true positives ( T P ), while the instances correctly identified as negative samples are referred to as true negatives ( T N ). Conversely, when an actual positive sample is mistakenly classified as negative by the model, it is recorded as a false negative ( F N ). Similarly, when an actual negative sample is mistakenly classified as positive, it is recorded as a false positive ( F P ). Based on the above, Precision and Recall can be calculated using the following formulas:
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N .
The Precision–Recall ( P R ) curve illustrates the trade-off between these two metrics as the threshold changes, making it a common method for evaluating the effectiveness of classification models. Average precision ( A P ) is the area under the P R curve, obtained by integrating the area enclosed by the P R curve and the coordinate axes, and serves as a measure of the model’s overall performance. The larger the area value, the better the overall performance of the detection model. Its calculation formula is as follows:
A P = 0 1 P ( r ) d r .
In target detection tasks, A P is an important evaluation metric, typically calculated based on different I o U thresholds. These thresholds include A P 50 and A P 75 , which represent the average precision when the I o U threshold is 0.5 and 0.75, respectively. Additionally, an I o U step of 0.05 can be used to calculate A P values from A P 50 to A P 95 , with the average taken as A P 50 : 95 . The model’s performance can be evaluated using multiple metrics at different thresholds.
Average recall ( A R ) is the average of all R e c a l l values within the I o U range of [0.5, 1], used to evaluate the recall performance of the detection model. The calculation formula is as follows:
A R = 2 0.5 1 r e c a l l ( o ) d o
In target detection tasks, the maximum number of targets to be detected is often limited. Therefore, metrics such as A R 100 and A R 300 are typically used to represent the average recall when the number of detections is limited to 100, 300, and so on, to more comprehensively evaluate the performance of the detection model.

4.3. Comparison of Server Evaluation

This section will compare the performance of SAPAA-MMF algorithm and each comparison algorithm in the server evaluation using five datasets.
The server evaluation results based on the Bird’s Nest Foreign Bodies dataset are presented in Table 2. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in four of the server evaluation metrics. The A P , A P 50 , A R 100 , and A R 300 metrics outperform those of the other comparison algorithms. Although A P 75 is lower than the maximum value among the comparison algorithms, the overall A v e r a g e value remains higher than that of the other comparison algorithms.
The server evaluation results based on the Cement Rod Damage dataset are presented in Table 3. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in four of the server evaluation metrics. The A P , A P 50 , A R 100 , and A R 300 metrics outperform those of the other comparison algorithms. Although A P 75 is lower than the maximum value among the comparison algorithms, the overall A v e r a g e value remains higher than that of the other comparison algorithms.
The server evaluation results based on the Shockproof Hammer Slip dataset are presented in Table 4. The SAPAA-MMF algorithm ranks first in all five server evaluation metrics: A P , A P 50 , A P 75 , A R 100 , and A R 300 , outperforming the other comparison algorithms.
The server evaluation results based on the Insulator Self-Explosion dataset are presented in Table 5. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in four of the server evaluation metrics. The A P , A P 50 , A R 100 , and A R 300 metrics outperform those of the other comparison algorithms. Although A P 75 is lower than the maximum value among the comparison algorithms, the overall A v e r a g e value remains higher than that of the other comparison algorithms.
The server evaluation results based on the mixed dataset are presented in Table 6. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in three of the server evaluation metrics. The A P , A P 50 , and A P 75 metrics outperform those of the other comparison algorithms. Although A R 100 and A R 300 are lower than the maximum values among the comparison algorithms, the overall A v e r a g e value remains higher than that of the other comparison algorithms.
From Table 2, Table 3, Table 4, Table 5 and Table 6, it is evident that the SAPAA-MMF algorithm consistently achieves the best results across the five datasets in the server evaluation, excelling in key metrics such as A P , A P 50 , A P 75 , A R 100 , and A R 300 . Notably, the A P and A P 50 metrics consistently rank first across all datasets, and the overall average value of all SAPAA-MMF indicators remains higher than those of the comparison algorithms. This consistent performance highlights the superiority of the SAPAA-MMF algorithm, demonstrating its innovation and advancement over traditional methods. The SAPAA-MMF algorithm, as a multi-method fusion federated learning approach, leverages the strengths of different algorithms to achieve superior performance. In the one-stage aggregation process, it effectively balances the impact of data sample size and training loss, leading to strong initial results. The two-stage adaptive aggregation further enhances the stability of the training process, ensuring that the model not only performs well but also exhibits robust generalization capabilities. Consequently, the SAPAA-MMF algorithm produces a federated learning model that excels in defect detection, significantly improving both detection accuracy and generalization across diverse datasets.

4.4. Comparison of Client Evaluation

This section will compare the performance of the SAPAA-MMF algorithm with that of each comparison algorithm in the client evaluation using five datasets.
The client evaluation results based on the Bird’s Nest Foreign Bodies dataset are presented in Table 7. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in three of the client evaluation metrics. The A P , A P 50 , and A P 75 metrics outperform those of the other comparison algorithms. Although A R 100 and A R 300 are lower than the maximum values among the comparison algorithms, the overall A v e r a g e value remains higher than that of the other comparison algorithms.
The client evaluation results based on the Cement Rod Damage dataset are presented in Table 8. The SAPAA-MMF algorithm ranks first in all five client evaluation metrics, A P , A P 50 , A P 75 , A R 100 , and A R 300 , outperforming the other comparison algorithms.
The client evaluation results based on the Shockproof Hammer Slip dataset are presented in Table 9. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in four of the client evaluation metrics. The A P , A P 50 , A R 100 , and A R 300 metrics outperform those of the other comparison algorithms. The overall A v e r a g e value remains higher than that of the other comparison algorithms.
The client evaluation results based on the Insulator Self-Explosion dataset are presented in Table 10. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in four of the client evaluation metrics. The A P , A P 75 , A R 100 , and A R 300 metrics outperform those of the other comparison algorithms. The overall A v e r a g e value remains higher than that of the other comparison algorithms.
The client evaluation results based on the mixed dataset are presented in Table 11. From this dataset, it is evident that the SAPAA-MMF algorithm ranks first in four of the client evaluation metrics. The A P , A P 50 , A R 100 , and A R 300 metrics outperform those of the other comparison algorithms. Although A P 75 is lower than the maximum value among the comparison algorithms, the overall A v e r a g e value remains higher than that of the other comparison algorithms.
From Table 7, Table 8, Table 9, Table 10 and Table 11, it is evident that the SAPAA-MMF algorithm consistently achieves the best results across the five client evaluation metrics: A P , A P 50 , A P 75 , A R 100 , and A R 300 on the five datasets. The A P metric consistently ranks first across all five datasets, and the average value of all metrics for the SAPAA-MMF algorithm remains higher than that of the other comparison algorithms. SAPAA-MMF combines the strengths of multiple algorithms and enhances the server evaluation, resulting in a better global model. This indicates that the global model performs better across the entire dataset, demonstrating strong generalization capabilities and consistent performance across different data distributions. As a result, client evaluation has also improved, further confirming that the global model exhibits strong generalization capabilities on client data. This underscores the effectiveness and innovation of the SAPAA-MMF algorithm, which consistently delivers strong performance and generalization.

4.5. Hyper-Parameter Experimental Verification

In the SAPAA-MMF algorithm, factors such as data distribution, data quality, and the number of samples also influence the effectiveness of parameter aggregation. Since there is no direct indicator to measure these factors, the client training loss is used as an indirect measure of local data distribution and quality. In the SAPAA-MMF algorithm, the hyperparameter θ is used to control the balance between sample size and loss during parameter aggregation. The optimal value of the hyperparameter θ is determined through experiments, validating the rationale for using client loss as a condition for parameter aggregation. Table 12 and Table 13 in the mixed dataset present the optimal values of the hyperparameter θ for server and client evaluations across different metrics, including A P , A P 50 , A P 75 , A R 100 , and A R 300 .
From Table 12 and Table 13, it is evident that different values of the hyperparameter θ have varying effects on the server and client evaluation metrics. This fully demonstrates the effectiveness and rationale of considering both the number of client samples and training loss in the SAPAA-MMF algorithm. To intuitively determine the optimal value of the hyperparameter θ , the average values of various metrics and the corresponding values of θ are illustrated in a bar chart, as shown in Figure 5.
From Figure 5, it is evident that different values of the parameter θ in the SAPAA-MMF algorithm result in varying effects on both client-side and server-side evaluations. In the server evaluation, when the parameter θ is set to 0.003, the average metric value is 0.416 higher compared to the case where client loss is not considered. In the client evaluation, when the parameter θ is set to 0.003, 0.005, 0.007, 0.009, or 0.01, the average metric value is higher than the baseline value of 0.349, where client loss is not considered. Therefore, when the parameter θ is set to 0.003, the average evaluation metrics for both server and client are higher than those that do not consider client loss, achieving the best performance. This fully verifies that incorporating client loss during model parameter aggregation yields better results. Based on this analysis, the optimal value of θ used throughout this study is 0.003.

4.6. Validation of Multi-Method Fusion

This paper proposes the SAPAA-MMF algorithm. By leveraging the concept of multi-method fusion, different algorithms are combined, fully utilizing their complementary advantages to enhance the performance and generalization ability of the model while ensuring stability and convergence. To comprehensively verify the effectiveness of multi-method fusion, the FedAdam algorithm and the SAPAA-MMF algorithm without fusion are tested on five datasets. The detailed server and client evaluation metrics are presented in Table 14 and Table 15.
From Table 14 and Table 15, it is evident that after multi-method fusion, server-side evaluations showed improvements across multiple datasets. On the Bird’s Nest Foreign Bodies dataset, A P , A P 75 , A R 100 , and A R 300 increased by 1.6%, 2.1%, 0.8%, and 0.8%, respectively. On the Cement Pole Damage dataset, A P , A P 50 , A R 100 , and A R 300 increased by 1.2%, 1.6%, 3.8%, and 3.8%. On the Shockproof Hammer Slip dataset, A P , A P 50 , A P 75 , A R 100 , and A R 300 increased by 0.4%, 0.5%, 3.8%, 2.8%, and 1.5%, respectively. On the Insulator Self-Explosion dataset, A P , A P 50 , A R 100 , and A R 300 increased by 1.1%, 0.8%, 2%, and 1.5%, respectively. On the mixed dataset, A P , A R 100 , and A R 300 increased by 0.3%, 2.2%, and 2.2%.
In the client evaluation on the Bird’s Nest Foreign Bodies dataset, A P , A P 50 , A P 75 , A R 100 , and A R 300 increased by 1.8%, 0.9%, 2%, 0.4%, and 0.3%, respectively. On the Cement Pole Damage dataset, A P , A P 75 , A R 100 , and A R 300 increased by 3.9%, 3.9%, 2.4%, and 2.5%, respectively. On the Shockproof Hammer Slip dataset, A P , A P 50 , A R 100 , and A R 300 increased by 3.1%, 5.6%, 2.5%, and 2.5%, respectively. On the Insulator Self-Explosion dataset, A P , A P 50 , A P 75 , A R 100 , and A R 300 increased by 0.9%, 1.5%, 4.3%, 1%, and 0.7%, respectively. On the mixed dataset, A P , A P 50 , and A P 75 increased by 1.9%, 0.2%, and 3.3%, respectively.
From both server-side and client-side evaluations, fully considering the stability of client data samples and utilizing adaptive aggregation during server-side parameter aggregation leads to better aggregation outcomes. This demonstrates that multi-method fusion effectively leverages the complementary strengths of different algorithms, significantly improving model performance and effectiveness.

4.7. Training Effectiveness Experiment

To verify the effectiveness of applying federated learning to the training of transmission line defect detection models and to assess the effectiveness of the constructed federated learning training framework, nine federated learning algorithms, including the SAPAA-MMF algorithm proposed in this chapter, were selected for evaluation in this section. Figure 6 illustrates the trends of various evaluation metrics ( A P , A P 50 , A P 75 , A R 100 , and A R 300 ) for SAPAA-MMF and other algorithms on the mixed dataset as training progresses. As shown in the figure, SAPAA-MMF and the other algorithms exhibit an upward trend in all five evaluation metrics as training progresses, eventually stabilizing after a certain point, demonstrating the good convergence ability of SAPAA-MMF. Furthermore, except for the A P metric, where SAPAA-MMF is sub-optimal, SAPAA-MMF achieves the best results in the other four evaluation metrics after 20 training rounds, further demonstrating its effectiveness.

4.8. Statistical Hypothesis Testing

The Friedman statistical test is a non-parametric method for comparing multiple related samples. It evaluates whether there are significant differences across conditions by ranking observed values and comparing the differences between these rankings. In this section, the Friedman test is used to conduct hypothesis testing to determine whether there are significant differences in the performance of each algorithm. The first step involves sorting all algorithms according to performance indicators for each dataset and assigning ranks from 1 to k. Next, the average ranking of each algorithm across all datasets is calculated. Finally, the statistics required for the test are calculated using Equations (23) and (24) as follows:
τ χ 2 = 12 N k ( k + 1 ) i = 1 k r i 2 k ( k + 1 ) 2 4 ,
τ F = ( N 1 ) τ χ 2 N ( k 1 ) τ χ 2 .
where k is the number of algorithms participating in the statistical test, N is the number of datasets, and r i represents the average ranking value of the i algorithm in all datasets. τ χ 2 is used to measure the ranking difference among algorithms and τ F is the conversion of τ χ 2 to make it obey the approximate normal distribution, which is used to further analyze that the ranking difference τ F obeys the F distribution with degrees of freedom of ( k 1 ) and ( k 1 ) ( N 1 ) .
The significance level α is typically set at 0.05 in hypothesis testing. The null hypothesis posits that “there is no significant difference in the performance of the algorithms between the server and the client”. If the value of τ F is less than the critical value of the F distribution at the significance level α , the null hypothesis is accepted, indicating no significant difference in the performance of the algorithms between the server and the client. Otherwise, the null hypothesis is rejected, indicating a significant difference in performance.
To examine the performance differences between the server and the client across the nine algorithms, the Friedman statistical test was applied to analyze the experimental results. All statistical tests were conducted on the A P metric. The experiment involved five datasets and nine algorithms. According to Equations (23) and (24), the value of τ F for the server evaluation is 19.38, and for the client evaluation it is 16.17. The critical value of the F distribution at a significance level of 0.05 is F 0.05 ( 8 , 32 ) = 2.244 . Since 19.38 > 2.244 and 16.17 > 2.244 , the null hypothesis is rejected, indicating a significant difference in the performance of the algorithms between the server and the client. Based on these experimental results, it can be concluded that SAPAA-MMF outperforms the other comparison algorithms.

5. Conclusions

This study proposes a server-side Adaptive Parameter Aggregation Algorithm based on Multi-method Fusion (SAPAA-MMF) for transmission line defect detection. A federated learning framework for transmission line defect detection was constructed to enable distributed model training without the need to centrally store or transmit original data, addressing the issues of data silos, privacy concerns, and hardware performance limitations associated with traditional transmission line defect detection model training. To address the limitation that existing federated learning algorithms only use a single algorithm for server-side parameter aggregation, making it challenging to fully leverage the advantages of different algorithms, this study proposes a server-side adaptive parameter aggregation algorithm based on multi-method fusion to enhance the aggregation effect, model performance, and generalization ability. The performance of the SAPAA-MMF algorithm was compared with multiple baseline algorithms across five datasets, and the effectiveness of SAPAA-MMF was verified. This comprehensive performance across multiple datasets underscores the effectiveness of SAPAA-MMF in addressing challenges associated with federated learning in defect detection tasks. By integrating multiple methods and carefully balancing key factors, SAPAA-MMF demonstrates its capacity to provide a more reliable and adaptable solution for real-world transmission lines defect detection applications.
However, existing federated learning algorithms often overlook personalized client feature information during the client update process. Additionally, the generalization ability of these models for performing defect detection tasks in other domains has not been thoroughly demonstrated. Future research should explore additional strategies and optimize client parameter updates to further enhance model performance. Furthermore, applying SAPAA-MMF to defect detection tasks in other real-world scenarios could further highlight the effectiveness of our research approach.

Author Contributions

Conceptualization, L.Z. and B.B.; methodology, L.Z.; software, B.B.; validation, B.B. and L.L.; formal analysis, H.W.; investigation, H.W.; resources, H.W.; data curation, L.Z.; writing—original draft preparation, L.L.; writing—review and editing, S.L.; visualization, S.L.; supervision, S.L.; project administration, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Project of the State Grid Sichuan Electric Power Company No. 52199722000Y.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data in this study are available upon request from the corresponding author, as this dataset was obtained from State Grid Sichuan Electric Power Company Electric Power Research Institute, which is a private dataset and therefore no public link is provided. There are no copyright/ethics issues with these datasets.

Conflicts of Interest

Author Linhao Zhang was employed by the company State Grid Sichuan Electric Power Research Institute, Power Internet of Things Key Laboratory of Sichuan Province. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Miao, X.; Liu, X.; Chen, J.; Zhuang, S.; Fan, J.; Jiang, H. Insulator detection in aerial images for transmission line inspection using single shot multibox detector. IEEE Access 2019, 7, 9945–9956. [Google Scholar] [CrossRef]
  2. Zhao, Z.; Zhen, Z.; Zhang, L.; Qi, Y.; Kong, Y.; Zhang, K. Insulator detection method in inspection image based on improved faster r-cnn. Energies 2019, 12, 1204. [Google Scholar] [CrossRef]
  3. Feng, Z.; Guo, L.; Huang, D.; Li, R. Electrical insulator defects detection method based on yolov5. In Proceedings of the 2021 IEEE 10th Data Driven Control and Learning Systems Conference, Suzhou, China, 14–16 May 2021; pp. 979–984. [Google Scholar]
  4. Ju, M.; Yoo, C.D. Detection of bird’s nest in real time based on relation with electric pole using deep neural network. In Proceedings of the 2019 34th International Technical Conference on Circuits/Systems, Computers and Communications, Jeju, Republic of Korea, 23–26 June 2019; pp. 1–4. [Google Scholar]
  5. Zhang, C.; Xie, Y.; Bai, H.; Yu, B.; Li, W.; Gao, Y. A survey on federated learning. Knowl.-Based Syst. 2021, 216, 106775. [Google Scholar] [CrossRef]
  6. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A.Y. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  7. Li, L.; Fan, Y.; Tse, M.; Lin, K.-Y. A review of applications in federated Learning. Comput. Ind. Eng. 2020, 149, 106854. [Google Scholar] [CrossRef]
  8. Hong, J.; Zhu, Z.; Yu, S.; Wang, Z.; Dodge, H.H.; Zhou, J. Federated adversarial debiasing for fair and transferable representations. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Virtual Event, 14–18 August 2021; pp. 617–627. [Google Scholar]
  9. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic controlled averaging for federated learning. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 5132–5143. [Google Scholar]
  10. Yuan, H.; Zaheer, M.; Reddi, S. Federated composite optimization. In Proceedings of the 38th International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 12253–12266. [Google Scholar]
  11. Reddi, S.; Charles, Z.; Zaheer, M.; Garrett, Z.; Rush, K.; Konečnỳ, J.; Kumar, S.; McMahan, H.B. Adaptive federated optimization. arXiv 2020, arXiv:2003.00295. [Google Scholar]
  12. Wu, Q.; An, J. An active contour model based on texture distribution for extracting inhomogeneous insulators from aerial images. IEEE Trans. Geosci. Remote Sens. 2013, 52, 3613–3626. [Google Scholar] [CrossRef]
  13. Wang, W.; Wang, Y.; Han, J.; Liu, Y. Recognition and drop-off detection of insulator based on aerial image. In Proceedings of the 2016 9th International Symposium on Computational Intelligence and Design, Hangzhou, China, 10–11 December 2016; pp. 162–167. [Google Scholar]
  14. Yan, S.; Jin, L.; Duan, S.; Zhao, L.; Yao, C.; Zhang, W. Power line image segmentation and extra matter recognition based on improved otsu algorithm. In Proceedings of the 2013 2nd International Conference on Electric Power Equipment-Switching Technology, Matsue, Japan, 20–23 October 2013; pp. 1–4. [Google Scholar]
  15. Chen, J.; Wen, Y.; Nanehkaran, Y.A.; Zhang, D.; Zeb, A. Multiscale attention networks for pavement defect detection. IEEE Trans. Instrum. Meas. 2023, 72, 1–12. [Google Scholar] [CrossRef]
  16. Yu, Y.; He, Y.; Karimi, H.R.; Gelman, L.; Cetin, A.E. A two-stage importance-aware subgraph convolutional network based on multi-source sensors for cross-domain fault diagnosis. Neural Netw. 2024, 179, 106518. [Google Scholar] [CrossRef] [PubMed]
  17. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  18. Li, F.; Xin, J.; Chen, T.; Xin, L.; Wei, Z.; Li, Y.; Zhang, Y.; Jin, H.; Tu, Y.; Zhou, X.; et al. An automatic detection method of bird’s nest on transmission line tower based on faster_rcnn. IEEE Access 2020, 8, 164214–164221. [Google Scholar] [CrossRef]
  19. Li, L.; Wang, Z.; Zhang, T. Gbh-yolov5: Ghost convolution with bottleneckcsp and tiny target prediction head incorporating yolov5 for pv panel defect detection. Electronics 2023, 12, 561. [Google Scholar] [CrossRef]
  20. Wang, X.; Gao, H.; Jia, Z.; Li, Z. Bl-yolov8: An improved road defect detection model based on yolov8. Sensors 2023, 23, 8361. [Google Scholar] [CrossRef] [PubMed]
  21. Blanchard, P.; El Mhamdi, E.M.; Guerraoui, R.; Stainer, J. Machine learning with adversaries: Byzantine tolerant gradient descent. Adv. Neural Inf. Process. Syst. 2017, 30, 119–129. [Google Scholar]
  22. Yin, D.; Chen, Y.; Kannan, R.; Bartlett, P. Byzantine-robust distributed learning: Towards optimal statistical rates. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 5650–5659. [Google Scholar]
  23. Pan, Z.; Geng, H.; Wei, L.; Zhao, W. Adaptive client model update with reinforcement learning in synchronous federated learning. In Proceedings of the 2022 32nd International Telecommunication Networks and Applications Conference, Wellington, New Zealand, 30 November–2 December 2022; pp. 1–3. [Google Scholar]
  24. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2020, 2, 429–450. [Google Scholar]
  25. Li, Q.; He, B.; Song, D. Model-contrastive federated learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10713–10722. [Google Scholar]
  26. Zhu, Z.; Hong, J.; Zhou, J. Data-free knowledge distillation for heterogeneous federated learning. In Proceedings of the 38th International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 12878–12889. [Google Scholar]
  27. Fallah, A.; Mokhtari, A.; Ozdaglar, A. Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach. Adv. Neural Inf. Process. Syst. 2020, 33, 3557–3568. [Google Scholar]
  28. Li, X.-C.; Zhan, D.-C.; Shao, Y.; Li, B.; Song, S. Fedphp: Federated personalization with inherited private models. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Cham, Switzerland, 2021; pp. 587–602. [Google Scholar]
  29. Chen, J.; Zhang, R.; Guo, J.; Fan, Y.; Cheng, X. Fedmatch: Federated learning over heterogeneous question answering data. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual Event, 1–5 November 2021; pp. 181–190. [Google Scholar]
  30. Niu, Y.; Deng, W. Federated learning for face recognition with gradient correction. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 28 February–1 March 2022; pp. 1999–2007. [Google Scholar]
  31. Collins, L.; Hassani, H.; Mokhtari, A.; Shakkottai, S. Exploiting shared representations for personalized federated learning. In Proceedings of the 38th International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 2089–2099. [Google Scholar]
  32. Shen, T.; Zhang, J.; Jia, X.; Zhang, F.; Lv, Z.; Kuang, K.; Wu, C.; Wu, F. Federated mutual learning: A collaborative machine learning method for heterogeneous data, models, and objectives. Front. Inf. Technol. Electron. Eng. 2023, 24, 1390–1402. [Google Scholar] [CrossRef]
  33. Shi, C.; Shen, C.; Yang, J. Federated multi-armed bandits with personalization. In Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, Virtual Event, 13–15 April 2021; pp. 2917–2925. [Google Scholar]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Hsu, T.-M.H.; Qi, H.; Brown, M. Federated visual classification with real-world data distribution. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part X 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 76–92. [Google Scholar]
  36. Lin, T.-Y.; Maire, M.; Belongie, S.J.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
Figure 1. Architecture diagram of server-side adaptive parameter aggregation algorithm based on multi-method fusion.
Figure 1. Architecture diagram of server-side adaptive parameter aggregation algorithm based on multi-method fusion.
Bdcc 08 00102 g001
Figure 2. SAPAA-MMF algorithm flow chart.
Figure 2. SAPAA-MMF algorithm flow chart.
Bdcc 08 00102 g002
Figure 3. Examples of datasets. The image dataset was obtained from the State Grid Sichuan Electric Power Research Institute, Power Internet of Things Key Laboratory of Sichuan Province.
Figure 3. Examples of datasets. The image dataset was obtained from the State Grid Sichuan Electric Power Research Institute, Power Internet of Things Key Laboratory of Sichuan Province.
Bdcc 08 00102 g003
Figure 4. Mixed data example. The image dataset was obtained from the State Grid Sichuan Electric Power Research Institute, Power Internet of Things Key Laboratory of Sichuan Province.
Figure 4. Mixed data example. The image dataset was obtained from the State Grid Sichuan Electric Power Research Institute, Power Internet of Things Key Laboratory of Sichuan Province.
Bdcc 08 00102 g004
Figure 5. Mean value of client evaluation indicators using different parameter values (The dashed line represents the mean of indicators that do not take into account client losses.).
Figure 5. Mean value of client evaluation indicators using different parameter values (The dashed line represents the mean of indicators that do not take into account client losses.).
Bdcc 08 00102 g005
Figure 6. Evaluation metrics change curves for SAPAA-MMF and comparison algorithms run for 20 rounds on a mixed dataset.
Figure 6. Evaluation metrics change curves for SAPAA-MMF and comparison algorithms run for 20 rounds on a mixed dataset.
Bdcc 08 00102 g006
Table 1. Defect dataset information.
Table 1. Defect dataset information.
Datasetclient_1client_2client_3client_4Total
Bird’s nest foreign bodies11511511536381
Cement rod damage9010011023323
Shockproof hammer slip8510511529334
Insulator self-explosion1149310626339
Mixed dataset17517517554579
Table 2. The server evaluation results of each algorithm, based on the Bird’s Nest Foreign Bodies dataset.
Table 2. The server evaluation results of each algorithm, based on the Bird’s Nest Foreign Bodies dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.1660.4970.0700.2600.2650.252
FedAvgM0.1850.4550.1140.2650.2650.257
FedYogi0.1820.4830.0790.2400.2400.245
FedAdam0.1740.5060.0630.2470.2470.247
FedAdagrad0.1960.5030.0740.2550.2550.257
FedMedia0.1710.4960.0940.2430.2430.249
TrimmedAvg0.1970.4690.1560.2650.2650.270
Krum0.1380.4430.0470.2170.2320.215
SAPAA-MMF0.2020.5090.1230.2670.2670.274
Table 3. The server evaluation results of each algorithm, based on the Cement Rod Damage dataset.
Table 3. The server evaluation results of each algorithm, based on the Cement Rod Damage dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.6010.7830.7520.6330.6500.684
FedAvgM0.5430.7760.7030.5880.5880.640
FedYogi0.5970.7930.7540.6420.6420.686
FedAdam0.6020.7760.7520.6040.6040.668
FedAdagrad0.5340.7710.5670.5670.5670.601
FedMedia0.5770.7630.7520.6130.6130.664
TrimmedAvg0.5400.7860.6210.5960.5960.628
Krum0.5220.7270.6550.5670.5670.608
SAPAA-MMF0.6140.7950.7520.6540.6540.694
Table 4. The server evaluation results of each algorithm, based on the Shockproof Hammer Slip dataset.
Table 4. The server evaluation results of each algorithm, based on the Shockproof Hammer Slip dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.4430.6920.4990.5210.5210.535
FedAvgM0.4400.6910.4780.5140.5140.527
FedYogi0.4450.7000.4980.5450.5450.547
FedAdam0.4380.7030.5210.5210.5340.543
FedAdagrad0.2640.6290.1160.3930.3930.359
FedMedia0.3980.7140.4140.4970.4970.504
TrimmedAvg0.4390.6990.5540.5070.5070.541
Krum0.3100.6730.2880.3930.4000.412
SAPAA-MMF0.4460.7160.5860.5520.5520.570
Table 5. The server evaluation results of each algorithm, based on the Insulator Self-Explosion dataset.
Table 5. The server evaluation results of each algorithm, based on the Insulator Self-Explosion dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.2610.4630.2730.3550.3700.344
FedAvgM0.2930.5230.3490.3850.3910.388
FedYogi0.3460.5280.3810.3910.3910.407
FedAdam0.3300.5170.4070.3700.3760.400
FedAdagrad0.2560.5310.2600.3000.3000.329
FedMedia0.2640.4450.3330.3700.3700.356
TrimmedAvg0.3140.5260.3710.3970.3970.401
Krum0.2740.5260.3040.3730.3760.371
SAPAA-MMF0.3490.5330.3940.3980.3980.414
Table 6. The server evaluation results of each algorithm, based on the mixed dataset.
Table 6. The server evaluation results of each algorithm, based on the mixed dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.3200.5980.3440.3940.3940.410
FedAvgM0.3110.5700.3010.3940.3940.394
FedYogi0.2900.5810.2620.3560.3560.369
FedAdam0.3320.6280.3650.3680.3680.412
FedAdagrad0.3420.6510.2990.3680.3680.406
FedMedia0.2900.5620.3210.3970.3970.393
TrimmedAvg0.2840.5800.3010.3920.3970.391
Krum0.2630.5640.2520.3630.3650.361
SAPAA-MMF0.3460.6530.3700.3960.3960.432
Table 7. The client evaluation results of each algorithm, based on the Bird’s Nest Foreign Bodies dataset.
Table 7. The client evaluation results of each algorithm, based on the Bird’s Nest Foreign Bodies dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.2140.5710.1210.2870.2890.296
FedAvgM0.2240.5850.1750.3180.3180.324
FedYogi0.2030.5510.1470.3030.3030.302
FedAdam0.2140.5640.1400.2810.2820.296
FedAdagrad0.2190.5890.1550.2810.2810.305
FedMedia0.2170.5870.1360.3160.3160.314
TrimmedAvg0.2240.5620.1600.3160.3190.316
Krum0.2030.5180.1350.2950.3050.291
SAPAA-MMF0.2440.5970.1780.3170.3170.331
Table 8. The client evaluation results of each algorithm, based on the Cement Rod Damage dataset.
Table 8. The client evaluation results of each algorithm, based on the Cement Rod Damage dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.6110.8370.7200.6590.6590.697
FedAvgM0.6090.8110.7490.6530.6550.695
FedYogi0.6050.8470.7180.6670.6670.701
FedAdam0.5640.8530.7030.6310.6310.676
FedAdagrad0.5160.8220.5460.5660.5660.603
FedMedia0.6040.8510.6830.6710.6740.697
TrimmedAvg0.5970.8530.6920.6500.6500.688
Krum0.5690.7800.7080.6180.6190.659
SAPAA-MMF0.6210.8670.7580.6770.6780.720
Table 9. The client evaluation results of each algorithm, based on the Shockproof Hammer Slip dataset.
Table 9. The client evaluation results of each algorithm, based on the Shockproof Hammer Slip dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.4730.7510.5000.5440.5440.562
FedAvgM0.4650.7460.5540.5410.5410.569
FedYogi0.4700.7710.5670.5320.5320.574
FedAdam0.4360.7070.5060.5230.5230.539
FedAdagrad0.2760.6410.1780.3700.3760.368
FedMedia0.4660.7670.4990.5340.5340.560
TrimmedAvg0.4410.7320.4710.5050.5050.531
Krum0.4030.7300.4250.5040.5040.513
SAPAA-MMF0.4730.7720.5590.5530.5530.582
Table 10. The client evaluation results of each algorithm, based on the Insulator Self-Explosion dataset.
Table 10. The client evaluation results of each algorithm, based on the Insulator Self-Explosion dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.2630.4850.2620.3570.3610.346
FedAvgM0.2750.5030.2840.3720.3780.362
FedYogi0.3160.5540.3200.3770.3780.389
FedAdam0.3180.5390.3440.3660.3700.387
FedAdagrad0.2550.5290.2560.3340.3350.342
FedMedia0.2100.4340.1660.3190.3250.291
TrimmedAvg0.2750.5220.3000.3530.3600.362
Krum0.2330.4610.2360.3400.3480.324
SAPAA-MMF0.3280.5390.3480.3810.3890.397
Table 11. The client evaluation results of each algorithm, based on the mixed dataset.
Table 11. The client evaluation results of each algorithm, based on the mixed dataset.
AlgorithmIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
FedAvg0.2630.5340.2080.3790.3810.353
FedAvgM0.2770.5180.2150.3820.3820.355
FedYogi0.2700.5160.2620.3410.3410.346
FedAdam0.2650.5370.2380.3340.3340.342
FedAdagrad0.2780.5370.2160.3460.3460.345
FedMedia0.2800.5270.2390.3640.3640.355
TrimmedAvg0.2640.5140.2300.3690.3720.350
Krum0.2270.4980.1810.3330.3330.314
SAPAA-MMF0.2850.5400.2400.3830.3840.366
Table 12. Optimal indicators for server evaluation with different parameter values based on mixed datasets.
Table 12. Optimal indicators for server evaluation with different parameter values based on mixed datasets.
ParameterIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
θ = 0 0.3350.6240.3390.3900.3900.416
θ = 0.001 0.3040.6490.2750.4000.4000.406
θ = 0.003 0.3460.6530.3700.3960.3960.432
θ = 0.005 0.3290.5500.3520.3970.3970.405
θ = 0.007 0.3200.6290.2520.4110.4110.405
θ = 0.009 0.3160.6260.3030.4030.4030.410
θ = 0.01 0.3190.5700.3140.3990.3990.400
θ = 0.03 0.2440.5150.1960.3730.3780.341
θ = 0.05 0.1980.4330.1310.3430.3460.290
θ = 0.07 0.1310.3390.1120.2730.2730.226
θ = 0.09 0.1070.2690.0880.2550.2570.195
Table 13. Optimal indicators for client evaluation with different parameter values based on mixed datasets.
Table 13. Optimal indicators for client evaluation with different parameter values based on mixed datasets.
ParameterIndicator
AP AP 50 AP 75 AR 100 AR 300 Average
θ = 0 0.2840.5390.2710.3260.3260.349
θ = 0.001 0.2780.5330.2710.3310.3310.349
θ = 0.003 0.2850.5400.2400.3830.3840.366
θ = 0.005 0.2870.5200.2590.3520.3520.354
θ = 0.007 0.2770.5460.2650.3510.3510.358
θ = 0.009 0.2750.5340.2810.3470.3470.357
θ = 0.01 0.2780.5310.2480.3510.3510.352
θ = 0.03 0.2510.5460.1900.3580.3580.341
θ = 0.05 0.1860.4590.1150.3040.3050.274
θ = 0.07 0.1350.3580.0700.2910.2910.229
θ = 0.09 0.0910.2800.0490.2060.2060.166
Table 14. Comparison between non-multi-method fusion and multi-method fusion server evaluation based on different datasets.
Table 14. Comparison between non-multi-method fusion and multi-method fusion server evaluation based on different datasets.
DatasetNo FusionFusion ( θ = 0 )
AP AP 50 AP 75 AR 100 AR 300 AP AP 50 AP 75 AR 100 AR 300
Bird’s nest foreign bodies0.1740.5060.0630.2470.2470.1900.4910.0840.2550.255
Cement rod damage0.6020.7760.7520.6040.6040.6140.7920.7520.6420.642
Shockproof hammer slip0.4380.7030.5210.5210.5340.4420.7080.5590.5490.549
Insulator self-explosion0.3300.5170.4070.3700.3760.3410.5250.4040.3900.391
Mixed0.3320.6280.3650.3680.3680.3350.6240.3390.3900.390
A v e r a g e 0.3750.6260.4220.4220.4260.3840.6280.4280.4450.445
Table 15. Comparison between non-multi-method fusion and multi-method fusion client evaluation based on different datasets.
Table 15. Comparison between non-multi-method fusion and multi-method fusion client evaluation based on different datasets.
DatasetNo FusionFusion ( θ = 0 )
AP AP 50 AP 75 AR 100 AR 300 AP AP 50 AP 75 AR 100 AR 300
Bird’s nest foreign bodies0.2140.5640.1400.2810.2820.2320.5730.1600.2850.285
Cement rod damage0.5640.8530.7030.6310.6310.6030.8500.7420.6550.656
Shockproof hammer slip0.4360.7070.5060.5230.5230.4670.7630.4980.5480.548
Insulator self-explosion0.3180.5390.3440.3660.3700.3270.5540.3870.3760.377
Mixed0.2650.5370.2380.3340.3340.2840.5390.2710.3260.326
A v e r a g e 0.3590.6400.3870.4270.4280.3830.6560.4120.4380.438
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Bian, B.; Luo, L.; Li, S.; Wang, H. Federated Learning with Multi-Method Adaptive Aggregation for Enhanced Defect Detection in Power Systems. Big Data Cogn. Comput. 2024, 8, 102. https://doi.org/10.3390/bdcc8090102

AMA Style

Zhang L, Bian B, Luo L, Li S, Wang H. Federated Learning with Multi-Method Adaptive Aggregation for Enhanced Defect Detection in Power Systems. Big Data and Cognitive Computing. 2024; 8(9):102. https://doi.org/10.3390/bdcc8090102

Chicago/Turabian Style

Zhang, Linghao, Bing Bian, Linyu Luo, Siyang Li, and Hongjun Wang. 2024. "Federated Learning with Multi-Method Adaptive Aggregation for Enhanced Defect Detection in Power Systems" Big Data and Cognitive Computing 8, no. 9: 102. https://doi.org/10.3390/bdcc8090102

Article Metrics

Back to TopTop