Next Article in Journal
Extended Kalman Filter-Based Vehicle Tracking Using Uniform Planar Array for Vehicle Platoon Systems
Previous Article in Journal
Accuracy, Repeatability, and Reproducibility of a Hand-Held Structured-Light 3D Scanner across Multi-Site Settings in Lower Limb Prosthetics
Previous Article in Special Issue
Enhancing Urban Mobility with Self-Tuning Fuzzy Logic Controllers for Power-Assisted Bicycles in Smart Cities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vehicle Position Detection Based on Machine Learning Algorithms in Dynamic Wireless Charging

Electrical and Computer Engineering Department, Florida International University, Miami, FL 33174, USA
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(7), 2346; https://doi.org/10.3390/s24072346
Submission received: 27 February 2024 / Revised: 19 March 2024 / Accepted: 5 April 2024 / Published: 7 April 2024
(This article belongs to the Special Issue Integrated Control and Sensing Technology for Electric Vehicles)

Abstract

:
Dynamic wireless charging (DWC) has emerged as a viable approach to mitigate range anxiety by ensuring continuous and uninterrupted charging for electric vehicles in motion. DWC systems rely on the length of the transmitter, which can be categorized into long-track transmitters and segmented coil arrays. The segmented coil array, favored for its heightened efficiency and reduced electromagnetic interference, stands out as the preferred option. However, in such DWC systems, the need arises to detect the vehicle’s position, specifically to activate the transmitter coils aligned with the receiver pad and de-energize uncoupled transmitter coils. This paper introduces various machine learning algorithms for precise vehicle position determination, accommodating diverse ground clearances of electric vehicles and various speeds. Through testing eight different machine learning algorithms and comparing the results, the random forest algorithm emerged as superior, displaying the lowest error in predicting the actual position.

1. Introduction

Electric vehicles are progressively gaining popularity as viable alternatives to traditional fossil fuel-powered counterparts, sparking a strategic transformation within the automotive industry toward embracing electric propulsion. This shift is primarily motivated by growing concerns regarding the depletion of fossil resources and the environmental repercussions associated with conventional fuel-powered vehicles. Despite notable advancements in the design of electric vehicle charging infrastructure, a persistent challenge looms large—the prevalence of range anxiety. This enduring issue acts as a formidable barrier, impeding the widespread acceptance of electric vehicles among the general public. In response to the aforementioned challenges, dynamic wireless charging (DWC) has emerged as a promising technological solution. DWC not only addresses the concerns related to range anxiety but also holds the potential to alleviate the necessity for larger battery capacities in mobile electric vehicles. This groundbreaking technology ensures a reliable and uninterrupted charging experience for electric vehicles in motion. As electric vehicles continue to gain traction as a sustainable mode of transportation, the significance of DWC becomes increasingly evident in mitigating challenges and fostering the widespread acceptance of these innovative vehicles [1].
When exploring the diverse realm of dynamic wireless charging, a pivotal distinction arises based on the length of the transmitter’s coil, leading to the categorization of two primary types: the long-track transmitter and the segmented coil array [2]. The long-track transmitter, as the name suggests, implements an extended transmitter track, enabling the simultaneous charging of multiple vehicles over considerable distances. This design, while promising in facilitating long-distance charging scenarios, comes with inherent drawbacks. Notably, it grapples with challenges such as reduced efficiency and heightened electromagnetic interference. The extended track introduces complexities that impact the overall performance of this charging method. In contrast, the segmented coil array takes a more intricate approach to overcome the limitations associated with the long-track transmitter. This alternative strategy involves the use of multiple transmitter coils, each precisely sized to match the receiver. As the receiver advances along the transmitter track, these coils are activated selectively. This thoughtful activation ensures a targeted and efficient energy transfer, effectively mitigating the efficiency and electromagnetic interference issues encountered with the long-track transmitter design [3].
As the receiver pad travels along the transmitter track, it selectively picks up energy only from the transmitters that are coupled with it. Nonetheless, the high-frequency current in each uncoupled transmitter coil not only induces increased losses and decreased efficiency but also generates problems with electromagnetic interference (EMI). Consequently, it is essential to turn off the uncoupled transmitters, as these inactive transmitter coils are not transferring power to the receiver pad. This control mechanism is referred to as segmentation control [4]. Segmentation control necessitates knowledge of the vehicle’s position to appropriately turn on and off each transmitter coil. Accurate positioning detection is essential to ensure efficient power transfer and precise manipulation of the electromagnetic field [5]. Scholars have extensively investigated receiver position identification, primarily categorized into three approaches: utilizing additional sensors, deploying extra coils for detection, and leveraging circuit parameters of transmitter coils.
Conventional methods for locating and detecting the receiver involve the use of optical, ultrasonic, and magnetic sensors [6]. While coil positioning can be achieved using radio-frequency detection and optic/ultrasound sensors, their effectiveness may be impacted by the magnetic field of the DWC system, necessitating the use of extra hardware [7]. However, sensors are significantly affected by environmental factors, such as rain, shadows, or dirt, leading to inaccurate judgments, especially in rapid switching processes among multiple charging units [8]. Moreover, the inclusion of additional magnetic sensors in these methods results in an elevated cost for the DWC system, and the installation of these sensors adds complexity to the overall DWC system.
Another strategy for detecting the vehicle’s position is the deployment of auxiliary coils. Auxiliary detection coils can be deployed on the primary side [9] or the secondary side [10] of the DWC system to detect the position by identifying changes in magnetic flux. Several additional coils were deployed in [11] to determine the position of electric vehicles. Moreover, to enhance the accuracy of receiver position detection, additional detection coils were strategically placed on both the primary and secondary sides of the DWC system [12]. The typical configuration of this system includes an extra coil for sensing and an identification circuit. The detection circuit transforms the induced voltage from the sensor coil into a signal that the main controller can interpret. The authors of [13] investigated four diverse layouts of detection coils, featuring configurations classified into single-coil, double-coil, and three-coil structures. In [10], an innovative receiver position identification system was employed, which utilized a ferrite core counter system with a Ferrite Position Identification (FPID) group messaging unit, to determine the receiver’s position through the counting of strategically placed ferrite cores within the DWC system. The authors of [14] proposed a dual-side magnetic integration of the inductor of the LCC-LCC compensation network to detect the vehicle position by utilizing a solenoid-integrated coil at the transmitter and a bipolar-integrated coil at the receiver, thereby eliminating the need for extra sensors. However, incorporating additional detection coils in the aforementioned approaches presents a challenge, as it causes an increase in the volume and cost of the DWC system. Furthermore, some of the aforementioned methods are constrained to detecting the receiver’s position in a single direction rather than across a full plane.
The third approach for identifying the vehicle’s position involves leveraging the circuit parameters of the transmitter coils. This method is preferred for detecting the position of electric vehicles as it eliminates the need for additional sensors or coils for detection. The authors of [15] identified the vehicle’s position by measuring the input impedance observed from the first resonator of the array. In [16], an approach was developed for position detection based on measuring the angle of the primary current, which varies depending on the vehicle’s position. Li et al. developed a vehicle position detection method in [17] based on monitoring the primary current, as the magnitude and phase of the current in the transmitter vary according to the receiver’s position, enabling position detection through the calculation of these current characteristics. However, these methods entail the need for multiple current samplings, and determining the RX position relies on the comparison of current values, resulting in a relatively slow response speed.
Machine learning stands out as a highly effective methodology for the prediction and detection of vehicle positions, primarily owing to its inherent self-learning capabilities, adaptability to diverse environments, and rapid response mechanisms. In the work presented by Shen et al. [18], machine learning algorithms were strategically utilized to enhance the adaptability and speed of response in estimating the position of the receiver. Nevertheless, it is imperative to assess the efficiency of machine learning algorithms within the study. This evaluation entails a comprehensive examination of system efficiency, which, in turn, necessitates the meticulous measurement or calculation of parasitic losses occurring in both inductors and capacitors. It is important to note that such measurements may be susceptible to inaccuracies, introducing a layer of complexity to the overall analysis.
The current literature lacks studies considering the influence of speed on predicting the vehicle’s position. Furthermore, there is a noticeable gap in research employing a diverse set of machine learning algorithms for vehicle detection. This absence of comprehensive studies serves as the driving force behind the authors’ motivation to present this paper. In this research, the authors employ various machine learning algorithms for vehicle position detection, providing a novel contribution to the existing body of knowledge. This study meticulously compares results in terms of accuracy and minimum error, shedding light on the performance disparities among the utilized algorithms. To ensure real-world relevance, this investigation incorporates different ground clearances of electric vehicles (air gap) and explores various vehicle speeds, contributing to a more representative analysis. Key input features considered in this study include transmitter current, Δ C (change in current), pad number, air gap, and speeds. This holistic approach aims to capture the multifaceted nature of real-world scenarios and provide insights into the algorithms’ adaptability across diverse conditions. The machine learning algorithms employed encompass gradient boosting, decision tree, support vector regression (SVR), random forest, neural network, K-nearest neighbor (KNN), and Bayesian ridge, each offering distinct advantages and nuances in addressing the complexities of vehicle position detection. The distinctive contributions of this paper, setting it apart from other studies, are as follows:
  • The methodology employed in this research encompasses the strategic utilization of an extensive array consisting of eight distinct machine learning algorithms. The primary objective is to predict the precise position of the vehicle, and this involves conducting a meticulous and comprehensive comparison of the results obtained using each algorithm. The overarching goal is to discern and isolate the most effective algorithm among the diverse set, thereby enhancing the accuracy and reliability of the predictive model.
  • An additional layer of depth is infused into this study through the deliberate inclusion of critical parameters that wield considerable influence in real-world scenarios. Noteworthy considerations encompass the dynamic variability in ground clearances exhibited by different vehicles and the range of speeds at which these vehicles operate. By incorporating these factors, this study endeavors to create a simulation environment that mirrors the complexities inherent in practical situations, thus fortifying the relevance and applicability of the findings.
  • The results for each algorithm are meticulously presented, encompassing detailed regression plots, a thorough analysis of errors specific to each position, and a robust evaluation of the mean squared error. This comprehensive approach is intentionally crafted to foster a nuanced and thorough understanding of the performance intricacies exhibited by each algorithm. The goal is to move beyond mere numerical outputs, offering a comprehensive insight that helps clarify the strengths and limitations of each algorithm in the specific context of predicting vehicle positions.

2. Structure of the DWC System

Figure 1 illustrates the structure of the segmented coil array employed in the case study. The transmitter track features five bipolar coils, each the same size as the receiver pad. Table 1 displays the dimensions of the receiver pad and transmitter coils.
When the vehicle aligns with the transmitter coil, the activation of the corresponding transmitter coil becomes essential. Upon the vehicle’s departure, the same transmitter coil should be promptly deactivated, ensuring optimal energy consumption. This strategic activation and deactivation process, synchronized with the vehicle’s presence, not only conserves energy but also minimizes electromagnetic interference (EMI). By selectively activating transmitter coils only within the coverage range of the receiver coil, the dynamic wireless charging (DWC) system achieves significant energy savings. Simultaneously, keeping non-relevant coils inactive contributes to a reduction in EMI, enhancing the overall efficiency and safety of the DWC system. The intricate dance of activation and deactivation is orchestrated by a sophisticated switching control system, reliant on precise information about the vehicle’s position. In essence, this mechanism ensures a harmonious interplay between the vehicle and the charging infrastructure, optimizing both energy utilization and electromagnetic compatibility.
The diagram in Figure 2 elucidates the mechanism for activating transmitter coils through the integration of advanced machine learning (ML) algorithms. In this setup, each transmitter coil is intricately linked to a dedicated compensation network via individual switches. The orchestration of these switches falls under the purview of a sophisticated switching control unit. This control unit operates the switches in alignment with the vehicle’s position, a determination facilitated by a robust machine learning (ML) algorithm. This study employs a diverse set of machine learning algorithms to enhance the precision of vehicle position identification. The algorithms employed include random forest, decision tree, gradient boosting, support vector regression (SVR), neural network, K-nearest neighbor (KNN), and Bayesian ridge. Each algorithm brings its unique capabilities to the forefront, contributing to the comprehensive evaluation of their effectiveness in optimizing transmitter coil activation based on vehicle position.
After accurately determining the vehicle’s real-time position and mapping the locations of individual transmitter pads, we can optimize coil activation using a lookup table. This table serves to activate specific transmitter coils precisely as the vehicle approaches, ensuring an efficient and responsive dynamic wireless charging (DWC) system.

3. Estimation of Vehicle’s Position Using Different Machine Learning Algorithms

In this section, we present the various machine learning algorithms used for estimating the vehicle’s position, considering different air gaps and speeds, and sampling the transmitter current. We can formulate the relationship between the input and output of the machine learning algorithms as follows:
y = f ( x ) = f ( d , V , i p )
where x and y represent the input and output, respectively, with y denoting the position. The input x comprises d, V, and i p , representing the air gap, speed, and primary current, respectively. The input x can be written as follows:
x = x 1 x 2 x t = d 1 V 1 i p 1 d 2 V 2 i p 2 d n V n i p n
Table 2 presents the database for the parameters, such as the air gap, speed, and position.
We conducted a simulation in ANSYS Maxwell (ANSYS Electronics Desktop 2023) and obtained mutual inductance data for various air-gap values, as listed in Table 2, corresponding to positions ranging from 0 to 1800mm. Subsequently, MATLAB Simulink was employed to derive primary current data based on the mutual inductance data at different vehicle speed values. Ultimately, a total of 22,200 samples were collected. The sample database was formed as follows:
Data = x 1 , y 1 , x 2 , y 2 x i , y i , i = 1 , 2 , 3 n
To significantly enhance the accuracy of obtaining vehicle positions, our emphasis is on creating a variety of machine learning (ML) algorithms for position estimation models. Subsequently, we conduct a thorough comparative analysis to evaluate the precision achieved by these ML algorithms in position estimation.

3.1. Random Forest

Random forest operates as an ensemble method and is renowned for its effectiveness in addressing classification challenges. Unlike single decision trees, this algorithm trains multiple trees within an ensemble. The collective decision of these trees, based on a majority consensus, determines the final class [19]. Random forest boasts several advantages, including speed, scalability, resilience to noise, and resistance to overfitting. It is user-friendly, eliminating the need for intricate parameter management. Additionally, random forest provides ease of interpretation and visualization. Figure 3 visually represents the structure of the random forest algorithm, showcasing its ensemble-based approach. This methodology ensures versatility and reliability, making the random forest algorithm a robust solution across various classification scenarios.
In the domain of random forests, each tree within the ensemble relies on a set of randomly chosen variables. To formalize this, let x = ( x 1 , , x p ) T denote a p-dimensional random vector representing the real-valued input or predictor variables, and let y be a random variable representing the real-valued response. We assume an unknown joint distribution P ( x , y ) . The main objective is to discover a prediction function f ( x ) capable of predicting y. Ensemble methods construct this prediction function f using a set of base learners h 1 ( x ) , , h J ( x ) , and these base learners are combined to create the ensemble predictor f ( x ) . In regression tasks, the base learners are typically averaged.
h ^ ( x ) = y ¯ = 1 n i = 1 n y k i ; f ^ ( x ) = 1 j j = 1 J h ^ j ( x )
where h ^ j ( x ) is the prediction of the response variable at x using the jth tree.
Key hyperparameters include the number of trees in the random forest, controlling the maximum depth of each tree to limit the depth, the minimum samples required to split an internal node, and the minimum samples required to be at a leaf node. In the model used, 100 trees were employed with no specified maximum depth, a minimum of two samples were required to split an internal node, and one sample was required to be at a leaf node.

3.2. Decision Tree

A decision tree stands out as a supervised machine learning technique specifically designed to tackle both classification and regression problems through a systematic process of data division based on distinct parameters [20]. The leaves of the tree represent the decision outcomes, whereas the nodes facilitate the segmentation of the data. In the context of a classification tree, the decision variable is categorical, leading to binary outcomes like yes/no. Conversely, in a regression tree, the decision variable is continuous, accommodating numerical predictions. The advantages of decision trees are manifold. They exhibit versatility in handling diverse scenarios, whether regression or classification, and provide interpretability. Their effectiveness extends to managing both categorical and quantitative values, along with the ability to handle missing attribute values through imputation. The tree traversal algorithm ensures high performance. However, decision trees face challenges, particularly the risk of overfitting. To overcome this limitation, the random forest technique offers a solution by adopting an ensemble modeling approach [21]. The structure of the decision tree algorithm is visually depicted in Figure 4, emphasizing its hierarchical and branching nature.
The concept of entropy, quantifying the amount of information required for an accurate description of data, is expressed as follows:
E n t r o p y ( S ) = i = 1 C p i log 2 ( p i )
where S is the training dataset, C is the number of classes, and p is the proportion of S classified as i. The aim of a split in a tree is to decrease the impurity (uncertainty) in the dataset with respect to the class in the next stage. This objective is achieved by calculating the information gain as follows:
G a i n ( S , a ) = E n t r o p y ( S ) v Values | S v | | S |
where S v = s S : a ( s ) = V , with v being the value of the attribute.
Key hyperparameters include the maximum depth of the tree, regulating the maximum depth, the minimum samples required to split an internal node, specifying the minimum number of samples required to split an internal node, and the minimum samples needed to be at a leaf node. In the model used, there was no specified maximum depth, two samples were required to split an internal node, and a minimum of one sample per leaf node was set.

3.3. Gradient Boosting

Gradient boosting, a robust algorithm, employs an iterative approach to enhance model performance. It achieves this by fitting a weak learner to the residual errors in each iteration, progressively refining predictions. The strength of gradient boosting lies in its ability to unravel intricate data structures, capture nonlinearity, and detect high-order interactions within the data. This technique proves particularly effective in scenarios featuring a vast number of potential predictors, ranging from hundreds to tens of thousands. As the algorithm iterates, it autonomously refines its understanding of the data, leading to continuous improvements in overall model accuracy [22]. The structural overview of the gradient boosting algorithm is depicted in Figure 5.
In the context of a training dataset D = { ( x i , y i ) } i = 1 N , the primary objective of gradient boosting is to derive an approximation, denoted as F ^ ( x ) , for the underlying function F * ( x ) . This function maps input instances x to their corresponding output values y. The optimization process involves minimizing the expected value of a predefined loss function L ( y , F ( x ) ) . Gradient boosting achieves this by constructing an additive approximation of F * ( x ) through a weighted sum of functions. Initially, an initial constant approximation of F * ( x ) is acquired as follows:
F ^ 0 ( x ) = arg min γ i = 1 N L ( y i , γ )
Yet, rather than directly addressing the optimization problem, each h m can be interpreted as a greedy step within a gradient descent optimization for F * . In this context, every model h m undergoes training on a distinct dataset D = { ( x i , r m i ) } i = 1 N , where r m i represents the pseudo-residuals, derived as follows:
r m i = L ( y i , F ^ m 1 ( x i ) ) F ^ m 1 ( x i ) F ^ ( x i ) = F ^ m 1 ( x i )
A step size β is chosen as the learning rate, and the model can be updated as follows:
F ^ m ( x ) = F ^ m 1 ( x ) + β h m ( x )
where the final model can be derived as follows:
F ^ ( x ) = F ^ M ( x )
Key hyperparameters include the number of boosting stages, the learning rate, controlling the step-size shrinkage, and the maximum depth of the individual trees, which limits the depth. In the model used, 100 boosting stages were applied with a learning rate of 0.1, and the tree depth was limited to three levels.

3.4. K-Nearest Neighbor

Operating as a non-parametric classification algorithm, K-nearest neighbor assigns an unlabeled sample point the class of its nearest neighbor from a set of previously labeled points [21]. This rule operates independently of the joint distribution of sample points and their classifications. Particularly effective for multi-modal classes and scenarios where objects can have multiple labels, it employs a straightforward lazy learning approach, albeit with reduced efficiency. Notably, its performance hinges on the prudent selection of the ‘k’ parameter, with no principled method available except through computationally expensive techniques such as cross-validation. The algorithm is susceptible to the adverse effects of noise and demonstrates sensitivity to irrelevant features. Furthermore, its performance dynamics vary with dataset size, as it necessitates revisiting all data points [23]. Figure 6 shows the structure of the KNN algorithm.
The KNN algorithm employs Euclidean distance metrics for locating the nearest neighbor. The Euclidean distance between x q u e r y and each x i in the training set ( D = { ( x i , y i ) } i = 1 N ) is calculated as follows:
d i = j = 1 M ( x query , j x i , j ) 2
where M is the number of features. The distances d i are sorted in ascending order, maintaining the corresponding indices. Then, the first K indices are selected from the sorted list. These indices correspond to the K-nearest neighbors. Finally, the predicted value for the query point ( y ^ q u e r y ) is calculated as follows:
y ^ query = i = 1 K 1 d i · y i i = 1 K 1 d i
The primary hyperparameter is the number of neighbors to consider during prediction. In the model used, KNN was configured with five neighbors, influencing the local smoothing of predictions.

3.5. Support Vector Regression

Support vector regression (SVR) stands out for its utilization of kernels, sparse solutions, and VC (Vapnik–Chervonenkis) control over the margin and the number of support vectors. While not as widely embraced as support vector machine (SVM), SVR has proven effective in the estimation of real-valued functions. Operating as a supervised learning technique, SVR undergoes training with a symmetrical loss function that uniformly penalizes both high and low estimates. A noteworthy advantage of SVR lies in its computational complexity, which remains unaffected by the dimensionality of the input space. Additionally, SVR demonstrates exceptional generalization capabilities, resulting in high prediction accuracy [24]. Figure 7 illustrates the structure of the support vector regression algorithm.
Figure 6. KNN algorithm structure.
Figure 6. KNN algorithm structure.
Sensors 24 02346 g006
Figure 7. Support vector regression structure.
Figure 7. Support vector regression structure.
Sensors 24 02346 g007
Training dataset D = { ( x i , y i ) } i = 1 N , where x i is the input feature vector and y i is the corresponding output. The objective is to find a regression function by solving the following optimization problem:
min w , b , ξ , ξ * 1 2 w 2 + C i = 1 N ( ξ i + ξ i * )
subject to the following constraints:
y i ( w · x i + b ) ε + ξ i , i = 1 , , N ( w · x i + b ) y i ε + ξ i * , i = 1 , , N ξ i , ξ i * 0 , i = 1 , , N
Given the nonlinear kernel function, the nonlinear SVR model is expressed as follows:
f ( x i ) = i = 1 N ( α i α i * ) x i x i + C i = 1 N ( α i α i * ) , α i , α i * [ 0 , C ]
Crucial hyperparameters include the regularization parameter, which determines the regularization strength, and the kernel function, which selects the type of kernel function to be used. In the model used, the regularization parameter was set to 1.0, and the radial basis function kernel was employed.

3.6. Neural Networks

Neural networks serve as computational models inspired by the intricacies of the human brain’s structure, processing, and learning mechanisms, albeit on a smaller scale. They excel in handling scenarios where relationships are nonlinear or dynamically evolving. Unlike traditional methods with rigid assumptions, such as normality and linearity, neural networks offer a flexible alternative. Their ability to capture a wide range of relationships enables users to model phenomena that might be challenging or impossible to explain using conventional approaches [25]. Figure 8 shows the structure of a three-layer backpropagation neural network.
The output of the hidden layer can be derived as follows:
h j = i = 1 n W i j I i j = 1 , 2 , , l
The output of the output layer can be calculated as follows:
O k = j = 1 l h k + 1 ( i ) W j k k = 1 , 2 , , m
Hyperparameters in the sequential model of Keras include the number of units in each dense layer, the number of training epochs, the batch size for optimization, the optimizer algorithm, and the loss function. In the model used, the neural network comprised two hidden layers, each with 10 units and ReLU activation. It was trained for 50 epochs with a batch size of 32, using the ’Adam’ optimizer and optimizing for mean squared error.

3.7. Bayesian Ridge

Bayesian ridge regression embraces a probabilistic methodology, leveraging the Gaussian probability distribution. The optimization of posterior predictions in Bayesian regression incorporates the use of l2 regularization. This sets Bayesian ridge regression apart, particularly in the derivation of the weighted coefficient ’w,’ which is deduced from a spherical Gaussian [26].
While Bayesian ridge regression demands computational time, it demonstrates notable adaptability concerning small data parameters. Its user-friendly nature is evident in effectively handling regularization challenges and facilitating the tuning of hyperparameters. Despite the computational demands, Bayesian ridge regression proves to be a valuable tool, especially in scenarios with limited data parameters, offering a practical and efficient approach to regularization problem-solving and hyperparameter fine-tuning.
Let observations y = ( y 1 , , y n ) T R n . Define features X = [ 1 n , x 1 , , x m ] R n × ( m + 1 ) , where x i represents the column vectors in R n , i = 1 , , m , and 1 n = ( 1 , , 1 ) T R n .
We make the assumption that each y i has a likelihood, given by:
p ( y i | ω , α ) = N y i | j = 1 m ω i X j , α
where ω = ( ω 0 , ω 1 , , ω m ) T R m represents the weights and α R represents the variance (indicating the noise).
Bayesian ridge regression focuses on determining the “posterior” distribution of the model parameters instead of directly finding these parameters. Consequently, Bayesian ridge regression requires a substantial volume of training data to enhance the accuracy of the model. Key hyperparameters include the values controlling the shape and precision of the distribution of the weights. In the model used, very small values were configured for these hyperparameters, implying weak regularization.

3.8. Linear Regression

Linear regression is recognized for its straightforward model structure, representing the regression function as a linear combination of predictors. Its popularity in various applications can be attributed to several factors. The linear form of the model allows for easily interpretable parameters. Furthermore, linear model theories boast well-established mathematical elegance. Additionally, linear regression serves as a foundational element for numerous contemporary modeling tools. Particularly in scenarios where the sample size is limited or the signal strength is modest, linear regression frequently delivers a satisfactory approximation of the underlying regression function [27].
Contemplate the dataset D , denoted as { ( y i , x i ) : i = 1 , , n } , where y i represents the ith response measured on a continuous scale, x i is the corresponding predictor vector, and n p denotes the sample size. The linear model is formally defined as follows:
y i = β 0 + β 1 x i 1 + + β p x i p + ϵ i
In the form of a matrix, we can write:
y = X β + ϵ
where y = [ y i ] n × 1 represents the n-dimensional response vector; X = ( x i j ) n × ( p + 1 ) , where x i 0 = 1 is commonly referred to as the design matrix; and ε = [ ε i ] n × 1 . Models (1) or (2) involve four primary statistical assumptions, which are as follows [27]:
  • Linearity: μ [ E ( y i | x i ) ] n × 1 = X β ;
  • Independence: ε i ’s are independent of each other;
  • Homoscedasticity: ε i ’s exhibit equal variance σ 2 ;
  • Normality: ε i ’s follow a normal distribution.
It is essential to note that many properties of linear models hold true even when all four assumptions are not met. Linear Regression does not have many hyperparameters to tune. It relies on the linear relationship between the features and the target variable.

4. Results and Discussion

To acquire the necessary data for predicting mutual inductance, the transmitter and receiver coils were simulated in ANSYS Maxwell. This simulation was conducted to obtain mutual inductance data at various positions for different air gaps. Subsequently, the obtained mutual inductance data were utilized in MATLAB R2022b Simulink to capture current data corresponding to different positions at different vehicle speeds. Figure 9 shows the mutual inductance between the transmitter coils and receiver with different air gaps at various positions. As the receiver moved across the transmitter coils, the mutual inductance reached its maximum when it aligned precisely with each transmitter coil and decreased when the receiver entered the space between adjacent transmitter coils, reaching a minimum at the midpoint between them. This cyclic variation occurred as the receiver crossed each transmitter coil, with the number of cycles corresponding to the number of transmitter coils. In our case, with five transmitter coils, we observed five cycles, resulting in five maxima and minima, as shown in the figure. Importantly, due to the similar structure of the transmitter coils, the maximum and minimum values of mutual inductance remained consistent within each cycle.
Table 3 presents the descriptions and values of the circuit parameters, which were utilized for obtaining the current data in MATLAB Simulink.
Figure 10 depicts regression plots illustrating the prediction of the vehicle’s position using eight different machine learning algorithms, namely random forest, decision tree, gradient boosting, KNN, support vector regression, neural network, and Bayesian ridge. Notably, the analysis revealed that random forest exhibited superior performance and accuracy in predicting the actual position, followed closely by the decision tree algorithm, which demonstrated better accuracy compared to the other algorithms in predicting the vehicle’s position. Furthermore, it can be seen that for each algorithm, the largest errors occurred when estimating the position near the center of each coil. This may be due to the small Δ C in these positions, which can otherwise be a strong identifier of the position depending on whether it is a large positive (increasing mutual inductance) or negative (decreasing mutual inductance) value.
Figure 11 shows a graph of the neural network training per epoch. This graph shows that the neural network effectively fitted the data within a few epochs (50 epochs in total). The validation of the training also showed that no overfitting in the neural network occurred. Additionally, although the mean squared error (MSE) settled at around 0.0052, it may be possible to achieve even better generalization (lower MSE) through appropriate tuning and optimization of the hyperparameters and architecture using various methods, as seen in [28].
In Figure 12, the MSE is illustrated for the different machine learning algorithms employed for predicting the vehicle’s position. The random forest algorithm exhibited the lowest MSE, followed closely by the decision tree algorithm, which exhibited a lower MSE compared to the other machine learning algorithms. The random forest algorithm’s superior performance may be due to the fact that it acts as a preventative measure against overfitting and yields more accurate predictions. Moreover, the random forest algorithm introduces extra randomness during tree growth. Instead of solely focusing on the most crucial feature during node splitting, it explores the best feature within a random subset of features. This approach fosters broad diversity, typically leading to superior model performance.
Figure 13 illustrates the errors of various machine learning algorithms at different positions. These findings highlight the superior performance of the random forest algorithm, which consistently exhibited the lowest error across different positions, followed by the decision tree algorithm, which demonstrated excellent performance and consistently exhibited very low errors in predicting the vehicle’s position.
We chose three machine learning algorithms (random forest, neural network, and SVR) for comparison. Their errors per position are displayed in a single graph, shown in Figure 14. The graph illustrates the superiority of the random forest algorithm in predicting the vehicle’s position with the lowest error.

5. Conclusions

Accurate detection of a vehicle’s position is crucial for dynamic wireless charging (DWC) of electric vehicles, ensuring efficient and safe operation. Traditional methods relying on sampling transmitters’ currents become ineffective due to the high speed of vehicles in DWC systems. This paper adopts machine learning algorithms for predicting the vehicle’s position, leveraging their self-learning capability, adaptability to the environment, and swift response. Eight distinct machine learning algorithms are employed for vehicle position detection. Mean squared error bar graphs for each of the eight machine learning algorithms and regression plots for each algorithm demonstrate their accuracy in predicting the actual position. Moreover, this paper illustrates the actual errors of each algorithm for each receiver position, with coil numbers labeled consistently with those in the mutual inductance figure. This aids readers in understanding the ability of each algorithm to predict the position corresponding to each coil or the distance between coils.
Based on the comparison, the ‘tree’-based methods, (decision tree and random forest) demonstrate better generalization with the provided dataset in this application. Finally, the results underscore the superiority of the random forest algorithm in accurately predicting the actual position. Future work will involve incorporating additional real-world parameters for predicting vehicle positions. Furthermore, future work may involve obtaining vehicle position detection data by integrating electric vehicle charging path planning into the existing framework. It should be noted that the results of this research can assist researchers in determining the most accurate machine learning algorithm for detecting vehicle positions in DWC systems. Additionally, it can help engineers in the field adopt the most suitable machine learning algorithm for implementing vehicle position detection systems.

Author Contributions

Conceptualization, M.B.; Formal analysis, M.B., A.S. (Alexander Stevenson), M.T. and A.S. (Arif Sarwat); Funding acquisition, A.S. (Arif Sarwat); Investigation, M.B., A.S. (Alexander Stevenson) and M.T.; Methodology, M.B., A.S. (Alexander Stevenson), M.T. and A.S. (Arif Sarwat); Project administration, A.S. (Arif Sarwat); Supervision, A.S. (Arif Sarwat); Validation, M.B., A.S. (Alexander Stevenson), M.T. and A.S. (Arif Sarwat); Writing—original draft, M.B.; Writing—review and editing, A.S. (Alexander Stevenson), M.T. and A.S. (Arif Sarwat). All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the U.S. Department of Energy (DOE) under grant number DE-NA0004109.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Z.; Pang, H.; Georgiadis, A.; Cecati, C. Wireless power transfer—An overview. IEEE Trans. Ind. Electron. 2018, 66, 1044–1058. [Google Scholar] [CrossRef]
  2. Behnamfar, M.; Tariq, M.; Sarwat, A.I. Novel Control Method With Five-Phase Interleaved Boost Converter to Reduce Power Pulsation in Dynamic Charging of Electric Vehicle. In Proceedings of the 2023 IEEE Wireless Power Technology Conference and Expo (WPTCE), San Diego, CA, USA, 4–8 June 2023; pp. 1–6. [Google Scholar]
  3. Deng, Z.; Hu, H.; Su, Y.; Chen, F.; Xiao, J.; Tang, C.; Lin, T. Design of a 60kW EV Dynamic Wireless Power Transfer System with Dual Transmitters and Dual Receivers. IEEE J. Emerg. Sel. Top. Power Electron. 2023, 12, 316–327. [Google Scholar] [CrossRef]
  4. Li, Y.; Hu, J.; Lin, T.; Li, X.; Chen, F.; He, Z.; Mai, R. A new coil structure and its optimization design with constant output voltage and constant output current for electric vehicle dynamic wireless charging. IEEE Trans. Ind. Inform. 2019, 15, 5244–5256. [Google Scholar] [CrossRef]
  5. Luo, Y.; Song, Y.; Wang, Z.; Mai, R.; Yang, B.; He, Z. Auto-segment control system with high spatially average power for dynamic inductive power transfer. IEEE Trans. Transp. Electrif. 2022, 9, 3060–3071. [Google Scholar] [CrossRef]
  6. Nagendra, G.R.; Chen, L.; Covic, G.A.; Boys, J.T. Detection of EVs on IPT highways. IEEE J. Emerg. Sel. Top. Power Electron. 2014, 2, 584–597. [Google Scholar] [CrossRef]
  7. Liu, H.; Tan, L.; Huang, X.; Zhang, M.; Zhang, Z.; Li, J. Power stabilization based on switching control of segmented transmitting coils for multi loads in static-dynamic hybrid wireless charging system at traffic lights. Energies 2019, 12, 607. [Google Scholar] [CrossRef]
  8. Song, C.; Kim, H.; Kim, Y.; Kim, D.; Jeong, S.; Cho, Y.; Lee, S.; Ahn, S.; Kim, J. EMI reduction methods in wireless power transfer system for drone electrical charger using tightly coupled three-phase resonant magnetic field. IEEE Trans. Ind. Electron. 2018, 65, 6839–6849. [Google Scholar] [CrossRef]
  9. Hasan, N.; Wang, H.; Saha, T.; Pantic, Z. A novel position sensorless power transfer control of lumped coil-based in-motion wireless power transfer systems. In Proceedings of the 2015 IEEE Energy Conversion Congress and Exposition (ECCE), Montreal, QC, Canada, 20–24 September 2015; pp. 586–593. [Google Scholar]
  10. Shin, Y.; Hwang, K.; Park, J.; Kim, D.; Ahn, S. Precise vehicle location detection method using a wireless power transfer (WPT) system. IEEE Trans. Veh. Technol. 2018, 68, 1167–1177. [Google Scholar] [CrossRef]
  11. Al Mahmud, S.A.; Panhwar, I.; Jayathurathnage, P. Large-area free-positioning wireless power transfer to movable receivers. IEEE Trans. Ind. Electron. 2022, 69, 12807–12816. [Google Scholar] [CrossRef]
  12. Patil, D.; Miller, J.M.; Fahimi, B.; Balsara, P.T.; Galigekere, V. A coil detection system for dynamic wireless charging of electric vehicle. IEEE Trans. Transp. Electrif. 2019, 5, 988–1003. [Google Scholar] [CrossRef]
  13. Azad, A.N.; Echols, A.; Kulyukin, V.A.; Zane, R.; Pantic, Z. Analysis, optimization, and demonstration of a vehicular detection system intended for dynamic wireless charging applications. IEEE Trans. Transp. Electrif. 2018, 5, 147–161. [Google Scholar] [CrossRef]
  14. Wang, J.; Wan, L.; Cai, C.; Xue, M.; Du, Y.; Zhang, J. A Dual-Side Magnetic Integration Based Receiver Detection Method of Long-Track DWPT System. IEEE Trans. Power Electron. 2024. [Google Scholar] [CrossRef]
  15. Simonazzi, M.; Sandrolini, L.; Mariscotti, A. Receiver–Coil Location Detection in a Dynamic Wireless Power Transfer System for Electric Vehicle Charging. Sensors 2022, 22, 2317. [Google Scholar] [CrossRef]
  16. Ruffo, R.; Cirimele, V.; Diana, M.; Khalilian, M.; La Ganga, A.; Guglielmi, P. Sensorless control of the charging process of a dynamic inductive power transfer system with an interleaved nine-phase boost converter. IEEE Trans. Ind. Electron. 2018, 65, 7630–7639. [Google Scholar] [CrossRef]
  17. Li, X.; Hu, J.; Wang, H.; Dai, X.; Sun, Y. A new coupling structure and position detection method for segmented control dynamic wireless power transfer systems. IEEE Trans. Power Electron. 2020, 35, 6741–6745. [Google Scholar] [CrossRef]
  18. Shen, H.; Tan, P.; Song, B.; Gao, X.; Zhang, B. Receiver position estimation method for multitransmitter WPT system based on machine learning. IEEE Trans. Ind. Appl. 2021, 58, 1231–1241. [Google Scholar] [CrossRef]
  19. Lorena, A.C.; Jacintho, L.F.; Siqueira, M.F.; De Giovanni, R.; Lohmann, L.G.; De Carvalho, A.C.; Yamamoto, M. Comparing machine learning classifiers in potential distribution modelling. Expert Syst. Appl. 2011, 38, 5268–5275. [Google Scholar] [CrossRef]
  20. Roy, S.; Tufail, S.; Tariq, M.; Sarwat, A. Photovoltaic Inverter Failure Mechanism Estimation Using Unsupervised Machine Learning and Reliability Assessment. IEEE Trans. Reliab. 2024, 1–15. [Google Scholar] [CrossRef]
  21. Ray, S. A quick review of machine learning algorithms. In Proceedings of the 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon), Faridabad, India, 14–16 February 2019; pp. 35–39. [Google Scholar]
  22. Zhang, Z.; Zhao, Y.; Canes, A.; Steinberg, D.; Lyashevska, O. Predictive analytics with gradient boosting in clinical medicine. Ann. Transl. Med. 2019, 7, 152. [Google Scholar] [CrossRef]
  23. Pernkopf, F. Bayesian network classifiers versus selective k-NN classifier. Pattern Recognit. 2005, 38, 1–10. [Google Scholar] [CrossRef]
  24. Awad, M.; Khanna, R.; Awad, M.; Khanna, R. Support vector regression. In Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers; Springer: Berlin/Heidelberg, Germany, 2015; pp. 67–80. [Google Scholar]
  25. Singh, A.; Thakur, N.; Sharma, A. A review of supervised machine learning algorithms. In Proceedings of the 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 16–18 March 2016; pp. 1310–1315. [Google Scholar]
  26. Massaoudi, M.; Refaat, S.S.; Abu-Rub, H.; Chihi, I.; Wesleti, F.S. A hybrid Bayesian ridge regression-CWT-catboost model for PV power forecasting. In Proceedings of the 2020 IEEE Kansas Power and Energy Conference (KPEC), Manhattan, KS, USA, 13–14 July 2020; pp. 1–5. [Google Scholar]
  27. Su, X.; Yan, X.; Tsai, C.L. Linear regression. Wiley Interdiscip. Rev. Comput. Stat. 2012, 4, 275–294. [Google Scholar] [CrossRef]
  28. Abdolrasol, M.G.M.; Hussain, S.M.S.; Ustun, T.S.; Sarker, M.R.; Hannan, M.A.; Mohamed, R.; Ali, J.A.; Mekhilef, S.; Milad, A. Artificial Neural Networks Based Optimization Techniques: A Review. Electronics 2021, 10, 2689. [Google Scholar] [CrossRef]
Figure 1. Structure of the coil.
Figure 1. Structure of the coil.
Sensors 24 02346 g001
Figure 2. The structure of activating transmitter coils based on machine learning (ML) algorithms.
Figure 2. The structure of activating transmitter coils based on machine learning (ML) algorithms.
Sensors 24 02346 g002
Figure 3. Random forest algorithm structure.
Figure 3. Random forest algorithm structure.
Sensors 24 02346 g003
Figure 4. Decision tree algorithm structure.
Figure 4. Decision tree algorithm structure.
Sensors 24 02346 g004
Figure 5. Gradient boosting algorithm structure.
Figure 5. Gradient boosting algorithm structure.
Sensors 24 02346 g005
Figure 8. Three-layer backpropagation neural network structure.
Figure 8. Three-layer backpropagation neural network structure.
Sensors 24 02346 g008
Figure 9. Mutual inductance between the transmitter coils and receiver with different air gaps at various positions.
Figure 9. Mutual inductance between the transmitter coils and receiver with different air gaps at various positions.
Sensors 24 02346 g009
Figure 10. Regression plots of eight different machine learning algorithms in predicting the vehicle’s position.
Figure 10. Regression plots of eight different machine learning algorithms in predicting the vehicle’s position.
Sensors 24 02346 g010
Figure 11. Neural network training per epoch.
Figure 11. Neural network training per epoch.
Sensors 24 02346 g011
Figure 12. Mean squared error (MSE) for different machine learning algorithms employed for predicting the vehicle’s position.
Figure 12. Mean squared error (MSE) for different machine learning algorithms employed for predicting the vehicle’s position.
Sensors 24 02346 g012
Figure 13. The errors of eight different machine learning algorithms employed for predicting the vehicle’s position at different positions.
Figure 13. The errors of eight different machine learning algorithms employed for predicting the vehicle’s position at different positions.
Sensors 24 02346 g013
Figure 14. The errors per position of three machine learning algorithms displayed in a single graph.
Figure 14. The errors per position of three machine learning algorithms displayed in a single graph.
Sensors 24 02346 g014
Table 1. Transmitter and receiver coupler dimensions.
Table 1. Transmitter and receiver coupler dimensions.
ParameterDescriptionValue
l A L Aluminum plate length410 mm
w A L Aluminum plate width410 mm
l f e Ferrite plate length370 mm
w f e Ferrite plate width370 mm
w c Coil width25 mm
l i n , c o i l Inner length of coil230 mm
l o v e r l a p Overlapping length of coils185 mm
d c o i l Distance between adjacent transmitter coils50 mm
dAir gap150 mm
Table 2. Database for machine learning parameters.
Table 2. Database for machine learning parameters.
ParameterValues
Air gap (d)[150 mm, 160 mm, 170 mm, 180 mm, 190 mm, 200 mm]
Speed (V)[40 km/h, 50 km/h, 60 km/h, 70 km/h, 80 km/h]
Position (y)[0, 100 mm, 200 mm, 300 mm, 400 mm, …, 1800 mm]
Table 3. Circuit parameters.
Table 3. Circuit parameters.
ParameterDescriptionValue
V i n Input voltage200 V
L p 1 , L p 2 , L p 3 Self-inductance of transmitter coils50 μ H
C p 1 , C p 2 , C p 3 Primary series resonant capacitors70 nF
L 2 Self-inductance of receiver coil50 μ H
C s Secondary resonant capacitor70 nH
C b 1 Input capacitor of buck converter100 μ F
C b 2 Output capacitor of buck converter75 μ F
L d Inductor of buck converter9 mH
f s Operating frequency of inverter85 kHz
fFrequency of buck converter20 kHz
R L Load resistor22 Ω
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Behnamfar, M.; Stevenson, A.; Tariq, M.; Sarwat, A. Vehicle Position Detection Based on Machine Learning Algorithms in Dynamic Wireless Charging. Sensors 2024, 24, 2346. https://doi.org/10.3390/s24072346

AMA Style

Behnamfar M, Stevenson A, Tariq M, Sarwat A. Vehicle Position Detection Based on Machine Learning Algorithms in Dynamic Wireless Charging. Sensors. 2024; 24(7):2346. https://doi.org/10.3390/s24072346

Chicago/Turabian Style

Behnamfar, Milad, Alexander Stevenson, Mohd Tariq, and Arif Sarwat. 2024. "Vehicle Position Detection Based on Machine Learning Algorithms in Dynamic Wireless Charging" Sensors 24, no. 7: 2346. https://doi.org/10.3390/s24072346

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop