Next Article in Journal
New Insights into Stroke from Continuous Passively Collected Temperature and Sleep Data Using Wrist-Worn Wearables
Next Article in Special Issue
Experimental Study of Deep Submersible Structure Defect Monitoring Based on Flexible Interdigital Transducer Surface Acoustic Wave Technology
Previous Article in Journal
An Efficient and Conditional Privacy-Preserving Heterogeneous Signcryption Scheme for the Internet of Drones
Previous Article in Special Issue
Dynamic Maritime Traffic Pattern Recognition with Online Cleaning, Compression, Partition, and Clustering of AIS Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Attribution Analysis to Quantify the Impact of Oceanographic and Maneuverability Factors on Vessel Shaft Power Using Explainable Tree-Based Model

1
Korea Marine Equipment Research Institute, Busan 49111, Republic of Korea
2
Department of Industrial and Data Engineering, Major in Industrial Data Science and Engineering, Pukyong National University, Busan 48513, Republic of Korea
3
Lab021 Shipping Analytics, Busan 48508, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1072; https://doi.org/10.3390/s23031072
Submission received: 28 November 2022 / Revised: 12 January 2023 / Accepted: 13 January 2023 / Published: 17 January 2023
(This article belongs to the Special Issue Advanced Sensing Technologies for Marine Intelligent Systems)

Abstract

:
A vessel sails above the ocean against sea resistance, such as waves, wind, and currents on the ocean surface. Concerning the energy efficiency issue in the marine ecosystem, assigning the right magnitude of shaft power to the propeller system that is needed to move the ship during its operations can be a contributive study. To provide both desired maneuverability and economic factors related to the vessel’s functionality, this research studied the shaft power utilization of a factual vessel operational data of a general cargo ship recorded during 16 months of voyage. A machine learning-based prediction model that is developed using Random Forest Regressor achieved a 0.95 coefficient of determination considering the oceanographic factors and additional maneuver settings from the noon report data as the model’s predictors. To better understand the learning process of the prediction model, this study specifically implemented the SHapley Additive exPlanations (SHAP) method to disclose the contribution of each predictor to the prediction results. The individualized attributions of each important feature affecting the prediction results are presented.

1. Introduction

Marine transport is believed to be the primary medium of transportation for worldwide trade due to its efficiency of load. A major part of a whole cluster in global economic activities was carried on the ocean [1], but this comes with the contribution to the total greenhouse emission. International Maritime Organization (IMO) published its Fourth IMO Greenhouse Gas (GhG) Study in 2020 and stated that global shipping generated 1056 million tons of CO2 in 2018, which was equivalent to about 2.89% of all anthropogenic CO2 emissions globally. Additionally, using a voyage-based allocation method, international shipping contributed 740 million tons of CO2 to global shipping in 2018 [2].
Because global climate change is driven by the excessive production of pollution emissions, the shipping industry, which is the world’s largest transport sector, must place a high priority on lowering its emissions [3]. For the purpose of promoting energy efficiency, there has been a significant increase in research over the past few years on how to estimate the ship’s propulsion power.
Marine propulsion power can be examined by measuring the shaft power generated by the engine and it provides insight into how well the system is functioning [4]. Integrated systems that record the performance data of every element of a vessel engine have been widely implemented in the field of maritime engineering and it has made it set in motion of multiple research initiatives to conduct various studies by analyzing these data [5]. The research on vessel shaft power prediction took a major proportion of the many studies utilizing such data, and it has shown consistent improvement in recent years.
From 1977 until 1984, four research conducted by Holtrop [6,7] and Mennen [8,9], were carried out to develop the numerical description of a ship’s propulsion power properties and the related measurement. The regression-based method was implemented for each of the research, and it is believed to be one of the major initiatives for developing a prediction model of vessel shaft power. The use of various regression-based analyses in forecasting ship propulsion power has grown significantly during the last several decades. Along with the development of machine learning methodology, numerous research that aimed to predict vessel shaft power also made use of the progression of this advancement by utilizing each of the recent algorithms when the research was conducted.
Traced back to the past decade, there was some research deploying machine learning models to predict the shaft power or vessel propulsion power. In 2011 [10], two statistical investigations of ship energy efficiency analyzing the data collected from a domestic ferry were conducted by comparing the Gaussian process (GP) and Artificial Neural Networks. ANN resulted in slightly better performance than GP. Hence the continuation of the research in [11] that applied ANN in both real-time and predictive propulsion settings. There were some findings that ANN was indeed widely used with the aim of predicting shaft power. In 2014, data collected from the towboat were analyzed using ANN that was compared with an ensemble neural network (ENN) [12]. However, later, different machine learning models were showing some advancement. Thus, in 2017, a prediction of shaft power along with the prediction of shaft torque and fuel consumption were conducted by deploying a collection of machine learning techniques, including multiple Linear Regression (LR), LASSO regression, and Random Forest (RF) [13], which was categorized as the black-box model. Other than that, shaft power or vessel propulsion power-related research has been conducted using support vector-based machine learning models. Among those are Support Vector Machine (SVM) [14] and Support Vector Regression (SVR) [15], both in 2021. Both showed satisfactory results. Then, in 2021, another research also attempted to create a benchmark study by comparing multiple machine learning models, including eXtreme Gradient Boosting (XGBoost), Support Vector Machine, Linear Regression (MLR), Polynomial Regression, Generative Additive Model (GAM), Neural Networks, SVM, and a decision tree-based algorithm, XGBoost, which was the best-performing model in terms of error rate, as well as prediction accuracy (R-squared) [16]. A comparative study of the machine learning method to predict vessel shaft power was also conducted by [17], evaluating five models, which are MLR, AdaBoost, K-Nearest Neighbors (KNN), ANN, and RF, which resulted in RF as the best model.
Subsequently, before the recent period, numerous research has proved the optimum usage of machine learning algorithms to predict shaft power. However, those studies only experimented with the development of a prediction model without explainability. Thus, there has been no further in-depth investigation of how the machine learning system arrives at its prediction findings or how the many predictors that were considered throughout the process of developing the model influence the outcome of the prediction.
Numerous studies utilizing machine learning-based methodologies have been expanded with the aided Explainable Artificial Intelligence (XAI) approach to assist in better decision-making processes. This approach is making its way into a wide variety of domains, including education [18]; lithology [19] and geology [20]; social science [21]; construction engineering [22,23]; transportation [24] and smart cities [25]; healthcare [26] and medical [27]; mass media and entertainment [28]; tourism, travel, and hospitality [29]; supply chain management and manufacturing [30]; law enforcement [31] and legal [32]; information technology [33]; and financial services [34,35]. Overall, the research utilizing XAI to explain their machine learning model stated that it provides transparency of how the machine learning model produces its decision. To the best of our knowledge, however, in the maritime domain, we might be among the first to adopt the XAI methodology. One previous research pioneered the implementation of XAI in the maritime domain conducted anomaly detection analysis [36]. Thus, we are motivated to propose an extended framework for shaft power prediction using the XAI approach.
The objective of this research is to provide an interpretable explanation of the vessel shaft power prediction model using explainable AI techniques. This research not only contributes to the development of an explainable machine learning model for shaft power prediction but also goes deeper into the individualized attribution of each factor affecting the magnitude of shaft power using Shapley values. Factual oceanographic factors including tidal current, wind, wave, sea salinity, and temperature, as well as operating factors, including vessel speed, the rudder angle, the ship heading, and the drafts, were among the predictors for this analysis. In addition to that, the analysis was not only done over entire the dataset but also considering each distinct voyage number to understand the dynamic changes of feature importance with respect to different vessel trips.
To build an interpretable prediction model, we have carefully chosen input variables so that only oceanographic factors and operating factors are included in the prediction model. Oceanographic factors include wave, current, and wind that are externally given but may have a significant effect on the propulsion power. Operating factors, including speed, draft, and cargo weight, are directly controlled by the operator. In addition to this, most of the previous research has included intermediate sensor values that can be observed from the sensors at the vessel machinery sensor (such as RPM, piston, or cooling system) as their predictors. Although including these intermediate variables may increase the model’s accuracy, it is not good for explaining to the model because these intermediate variables are also the results of the vessel’s operation and not the direct cause of the shaft power. Thus, in this study, all the intermediate variables are excluded from the input variable. The experimental results, however, confirm that the model trained without intermediate variables could achieve a quite good prediction accuracy.
After then, we tried to provide a clearer explanation of the model using an explainable AI method. Focused on the oceanographic factors and the operating factors as the predictors of the prediction model, this research aims to specifically reveal which among those factors take hold of the generation of shaft power by the vessel propeller. More importantly, it is widely known that recent developments in machine learning and deep learning algorithms have made it possible and uncomplicated to construct any model with high-performance metrics. These developments have been widely publicized in numerous mediums. However, this will not be sufficient if the results cannot be confirmed or validated in the open; to put it another way, the “black box” that represents the model itself needs to be opened to discover how it reached the predicted results. By then, even the specific attribution of a single feature can be explained in terms of how it influences the outcome of the prediction.
Thus, the goal of this research can be presented by answering these three questions:
  • The vessel sails against sea resistance by adjusting its engine operation which generates the shaft power to drive the propeller. Is it possible to predict the vessel shaft power considering the influence of uncontrollable variables such as the oceanographic factors and also the controllable variables such as the maneuverability factors?
  • Among the uncontrollable and controllable variables affecting the generation of vessel shaft power, what factors deploy the significant influence and how?
  • Does a different vessel voyage (trip) change the important factors affecting the vessel shaft power?

2. Data Description

This research utilized large datasets collected from vessel engine sensor data and the factual noon report of a 16-months voyage (January 2020 until June 2021) by a general cargo vessel with detailed specifications in Table 1. This vessel sailed for 24 different voyages crossing some Asian countries such as South Korea, China, Taiwan, Vietnam, and Thailand. The routes of the whole voyage can be seen in Figure 1.
In general, the sensor data consists of automated system records of all sensors installed in the vessel engine, and this system is used to control the performance of each of the engine parts. On the other hand, the day-to-day operating reports that were prepared by the vessel crew are recorded in the noon report. Thus, these two datasets were combined and for the specific purpose of this study, some features are chosen through the feature engineering step. The list of the features can be seen in Table 2.
The size of the original data set from the cargo vessel is 89,297 rows with a total of 151 columns, but only the 17 features in Table 2 were arbitrarily chosen for this analysis. These features are oceanographic and vessel maneuverability variables collected from the noon report. Oceanographic factors refer to the factors related to meteorology and (physical) oceanography, such as wind, wave, current, and sea surface measurements. Whereas maneuverability factors refer to the variables such as draft, rudder, ship heading, and speed. The target variable which is the shaft power was collected from the engine sensors data.

3. Research Methodologies

3.1. Machine Learning Prediction

In the process of developing a regression-based machine learning model to predict the vessel shaft power, several different approaches and ideas are put into practice to demonstrate that the attempt at model development was effective. In general, this research utilized a tree-based algorithm for several performance metrics. To begin the comprehensive procedure of choosing the regression model, we conducted a comparative analysis examining four different tree-based algorithms.

3.1.1. Tree-Based Regressor Comparative Study

The tree-based technique we referred to in this study is an ensemble method that combines many decision trees to acquire higher prediction performance than a single decision tree. There are at least two methods that are widely used to develop tree ensembles, which are bagging and boosting [37]. There are numerous developed algorithms based on decision trees and ensemble formation. Among all, Random Forest [38] has proven to outperform Tree Bagging and other random tree ensemble methods [39]. Thus, Random Forest which is also one advanced tree-based method is utilized in this research. Random Forest combines the principle of bagging bootstrap sampling with extra randomization of the input attributes used as candidates to divide an inner node of the tree. Instead of attempting to find the best split among all features, the method randomly chooses a subset of features at each node and then calculates the best test over these features to effectively divide the node. This strategy is extremely effective and has found several successful applications in a variety of industries.
Aside from the bagging method, the decision tree-based algorithm also has its boosting categorization. Up to this day, there have been some gradient boosting algorithms that have been extended into some different methods, mentioned in order of the most recently developed, which are CatBoost (Category Boosting) [40], LightGBM (Light Gradient Boosted Machine) [41], and XGBoost (eXtreme Gradient Boosting) [42]. Focusing on processing speed and accuracy, all these gradients boosting-based techniques are deployed in the combination of weak learners into strong learners.
Additionally, it is proven by [43] that to get the best model parameters for a tree-based machine learning algorithm, among all the parameters, the learning rate and the depth of the trees can be optimized by performing hyperparameter tuning. Thus, this research utilized Grid Search Cross Validation (GridSearchCV) to perform the hyperparameter tuning and evaluate the models.

3.1.2. Performance Evaluation

The model developed to predict the vessel shaft power was evaluated using metrics for regression-based machine learning models. These metrics provide the representation of the model’s error rate and accuracy. Among the existing metrics, this research used:
  • Mean Absolute Error (MAE)
Mean Absolute Error (MAE) is a model evaluation metric for regression models concerning its test set [44]. MAE measures the quality of fit in terms of the prediction error or the difference between the prediction results to the actual training data. It is calculated as:
M A E = 1 m i = 1 m | X i Y i | ,
2.
Root Mean Squared Error (RMSE)
RMSE emphasizes more on the larger absolute error of the model performance metrics.
R M S E = 1 m i = 1 m ( X i Y i ) 2 ,
The function has been widely adopted to standardize the unit measure of MSE [45].
M S E = 1 m i = 1 m ( X i Y i ) 2 ,
MSE and RMSE are connected to one another in a way that is always calculated the same way through the square root. Both the MSE ordering and the RMSE ordering of the regression models will produce the exact same results.
3.
Mean Absolute Percentage Error (MAPE)
Based on [46], MAPE has a consistent empirical risk minimization which represents a good basis to understand the limits of the machine learning algorithm. The calculated MAPE value shows the average deviation between the predicted value and the actual one.
M A P E = 1 m i = 1 m | Y i X i Y i | ,
4.
R2 or R-squared (Coefficient of Determination)
R-squared provides a more informative calculation of prediction performance metrics compared to MSE, RMSE, MAE, MAPE, and SMAPE [47]. According to [48], the coefficient of determination of R-squared is determined by calculating the proportion of the variance in the dependent variable that is predictable from the independent variable.
R 2 = 1 i = 1 m ( X i Y i ) 2 i = 1 m ( Y ¯ Y i ) 2 ,

3.2. Explainable Artificial Intelligence

As stated in the previous chapter, this research demonstrated a concept to interpret the black box of machine learning using Explainable Artificial Intelligence (XAI) methodology. There are several methodologies related to the implementation of XAI. One is categorized as heuristic explanations that are computed by approaches such as LIME [49], SHAP [50], or Anchor [51]. Among the three, this research implemented SHAP as it has the TreeExplainer method to specifically explain the output of the tree-based machine learning model.
The basic concept of SHAP is shown in Figure 2. Firstly, a black-box model is trained from the data. Based on this trained model, SHAP has the role to produce a model explanation given the testing data. The explanation was presented in a form of Shapley values that calculated the feature contribution to the model output.
SHAP (Shapley Additive exPlanations) values are calculated using a method called the “Shapley value” introduced by Lloyd S. Shapley [52] which measures the fair allocation results of the cooperative game. Shapley values provide a way to fairly distribute a value among a group of individuals, where each individual’s contribution is based on their unique characteristics and the characteristics of the others in the group.
The basic idea behind SHAP values is to assign each feature an importance value for a given prediction. The method assigns a value of importance for each feature by considering all possible combinations of features and the marginal contribution of each feature to the prediction. Suppose that there is a set of input X = { x 1 , x 2 , I , x n } and a machine learning model v for every subset of the inputs, and S is the subset of X with the size of k ( S ) , so that v ( S ) is the value of the subset. Then, the Shapley value for specific feature is estimated as the following:
φ x ( v ) = 1 n s [ v ( S { x } ) v ( S ) ] ( n 1 k ( S ) ) ,
where [ v ( S { x } ) v ( S ) ] is the marginal contribution of x for a given subset S . This calculation is repeated for all observations in the data set, resulting in a set of feature importance values for each observation. Once all the feature importance values are calculated, they can be used to interpret the importance of each feature for each observation in the data set. However, exact calculation of Shapely Value is computationally expensive because the size of feature permutation S increases exponentially with the number of features SHAP is a kind of an approximation for exact Shapely Value. Several types of SHAP model are proposed such as kernel SHAP, tree SHAP, or deep SHAP. To simplifies the calculation, each method assumes feature independence, or tries to exploit the structure of the black box model.
SHAP values provide a unified measure of feature importance that can be used for any model, regardless of whether it is a tree-based model, a linear model, or a neural network. SHAP values also have several attractive properties, such as being consistent with locally accurate feature importance measures and being able to consistently identify the feature importance of interaction effects. Correlation, on the other hand, only measures linear association between two variables, it does not take into account possible non-linear association, for example, SHAP values can detect feature importance when the correlation between input and output is zero. Also, correlation does not provide information of feature importance for a specific model or prediction.
In essence, local explanation typically relates to the process of explaining a single prediction result by breaking down each feature’s contribution represents by the Shapley value of one feature summed with other features’ contribution.
Global explanation, on the other hand, explains how features entirely contribute to the prediction result over entire data. Specifically for tree-based machine learning, Lundberg [53,54] developed what is called TreeExplainer to improve the interpretability of tree-based models like random forests, decision trees, and gradient-boosted trees. Originally [50], SHAP calculated the local explanation of one prediction as follows:
y i ^ = s h a p 0 + s h a p ( X 1 i ) + s h a p ( X 2 i ) + + s h a p ( X p i )  
The sum of all SHAP values is equal to the difference between the actual prediction value for observation i and the average prediction of overall data [55]. The model predicts the y i ^ by adding the s h a p 0 , the mean prediction across all data, and the s h a p ( X j i ) , which is the SHAP value for the jth feature for observation i that represents the marginal contribution of the feature to the model’s prediction, where in Equation (6) is the [ v ( S { x } ) v ( S ) ] .
Presented in Figure 3 is an illustrative example of how SHAP presented a local explanation of an individual prediction with feature attribution.
As shown, the individual prediction result of f ( x ) = 10 can be decomposed with added contribution value (summation of Shapley values) of all features, which in Figure 3 is equal to 1.6 + 0.7 2.9 0.9 = 1.5 , to the model’s fixed base value, 11.5 . In regression case, base value refers to the mean of the target variable over entire data points. Thus, the model output after the disclosure of contribution values would be the prediction base value added with the summation of features’ Shapley values. By this, we can quantify which feature mostly affects the prediction in that particular individual prediction. Then, the aggregation of all features attribution would provide the global explanation of the model [56].
Moreover, to better understand the relationship between an individual feature’s value and the model’s prediction, Partial Dependence Plots (PDP) [57] can disclose the feature’s marginal contribution to the prediction result. The PDP functions can be used to interpret the results of any “black box” learning method [58]. In Figure 4, X-axis indicates the actual value of the feature complete with the histogram distribution of the data. The Y-axis is the expected prediction result given only a subset of ‘Feature 8′ is considered in the prediction.
As this research utilized tree-based machine learning, SHAP TreeExplainer [53] is implemented to explain the prediction model. TreeExplainer provides a fast and exact feature attribution method by exploiting an ensemble-based decision tree structure.

3.3. Experimental Framework

The general framework of this research is presented in Figure 5. The experiment began with the data preparation procedure that will be detailed in Section 3.1. Generally, this step consists of data preprocessing and feature engineering. As the final dataset is created, regression-based analysis was done to build a prediction model for shaft power by considering the oceanographic and maneuverability factors. Section 3.2. will explain the machine learning algorithm used in this step, that is Random Forest Regressor.
Continued to the next step, to explain the learning mechanism of the machine learning model that is built in the previous step, utilizing Shapley Additive exPlanations (SHAP) methodology, the black box of the machine learning is disclosed. The calculation of the Shapley value in this procedure is able to mathematically explain each feature’s contribution to the machine learning prediction result. The concept of the explainable artificial intelligence method utilized in this step has been presented in Section 3.2. The resulting Shapley values then being presented in some plots to better interpret the results.

4. Experimental Results and Discussion

This section includes a precise summary of the experimental results, interpretation, and research findings, which are presented in the sequence of data preprocessing, machine learning prediction, model explanation, and research discussion.

4.1. Data Preprocessing

It is necessary to process the dataset for it to be used in the experimental procedures. Consequently, in the stage of the study known as “data preprocessing,” and in this research, there are several primary stages that are carried out before the data is fed to the machine learning model. This research empirically constructed this sequence of procedures of data preprocessing specifically for the purpose of feeding a dataset to a shaft power prediction model with general machine learning algorithms. These procedures are as follows:
  • Features selection
In the direction to predict vessel shaft power in this research, from the 151 features included in the original dataset, only features related to the oceanographic and maneuverability of the vessel during its operational status are chosen as seen in the Table 3. There are two ways in deciding which features to choose. One is by expert judgment, which means deciding the chosen features based on maritime engineering practice. In the end, some features that are chosen include oceanographic information such as wave, wind, current, water depth, and sea surface profile during the vessel voyage, and maneuverability features such as vessel speed over ground, both of the drafts, and the rudder angle, as well as the target variable, the main engine shaft power.
2.
Data filtering
Before the data is fed to the regression model that we have chosen based on the comparative study that we have conducted, we filtered the data based on domain expert percipience. First of all, when there is an error in the sensor systems, the system will record the data as −9999, and instances containing this value for each feature was removed. Seen in Table 4 below is the number of instances per feature that contain such error data and the ratio over the entire data.
Then, some specific feature values were filtered on their rational value range which is confirmed by the expert in maritime engineering. One is the total wave height which is no more than 6 m. Also, the speed over ground is set to be above 10 knots for it is indicated that the vessel is in cruise mode, as in it is working on the operational speed, sailing above the ocean.
From the chosen features, the statistical summary presented in Table 5 showed the distribution of the data numerically. Some indicators such as mean, maximum, and minimum value, as well as the standard deviation and variance of each feature, are presented.
3.
Features transformation
The first transformation of features was between the draft fore and draft aft values. Both are values of vessel trim when the respective front and rear part of the hull from sea level, or in maritime engineering jargon, they are
A v g D r a f t = D r a f t   F o r e + D r a f t   A f t 2
Vertical distance between the waterline and the bottom of the hull measured at the perpendicular of the bow. Vertical distance between the waterline and the bottom of the hull measured at the perpendicular of the stern.
The dataset also has features indicating angular measurements, such as ship heading, rudder angle, wind and relative wind direction, current direction, and total wave direction. These features need to be transformed as they have different nature from the other magnitudes from features with scalar quantities. Their value generally ranged between 0° to 360° or −360° to 360°. As the vessel moves in the direction of the ship heading, all directional values except the ship heading are measured by the difference between the ship heading angle. This calculation intended to transform those directions into a scalar by calculating the angular difference between all directional features with the main direction of the vessel which is the ship heading angle and make some new scalar quantities. The transformation procedure is seen in Figure 6.
The final dataset after preprocessing then consists of 47,444 rows with 15 columns as seen in Table 6.

4.2. Prediction Results

Shaft power prediction was carried out by considering the oceanographic and maneuverability factors of the vessel as predictors. Utilizing a regression-based machine learning algorithm, this process is considered successful as the performance evaluation metrics showed a good result. To better choose the machine learning algorithm for our prediction model, we compared at least four advanced tree-based regression models, and the results were evaluated based on the performance metrics such as R-squared for the prediction performance measure, and error rate evaluation with RMSE, MAE, and MAPE.
The comparative analysis was also done by conducting hyperparameter tuning using Grid Search Cross-validation (GridSearchCV) with k = 5 or 5-folds cross-validation. Configuration of the parameters for the four algorithms that we compared that were tuned using GridSearchCV is shown in Table 7.
Furthermore, with the tuned parameters as the result of hyperparameter tuning using 5-fold GridSearchCV, the four models were predicting the shaft power with the entire dataset and the result is presented in Table 8. Comparing the four tree-based regressors was meant to find the best-performing model and as seen in Table 8, Random Forest Regressor has the best evaluation score. Thus, we chose Random Forest Regressor as the regression model.
Additionally, the analysis for the shaft power prediction was done in two separate queries. The first one was done using all data regardless of voyage number. So, all 47,444 instances were fed to the machine learning model, and four metrics to evaluate the prediction quality including R-squared score, RMSE, MAE, and MAPE were calculated. Then, with the same procedure, we conducted a separate analysis per distinct voyage number with the initial indication that different voyages may have each unique results because of the different oceanographic and maneuverability factors as the vessel sailed above the different oceans on different voyage numbers.
Therefore, to better examine this indication, performance evaluation with error calculation using RMSE, MAE, and MAPE, was also conducted over different voyage numbers. The complete results of the prediction analysis per voyage trip can be seen in Table 9. There is some prediction that has lower or higher performance measurement compared to the analysis over the entire data regardless of the voyage number. This fact alone can imply that on the different oceanographic and maneuverability factors, there must be changes in prediction performance.
Up to this, the result of the analysis has answered the initial research question No.1, that it is possible to predict the vessel shaft power considering the influence of uncontrollable variables such as the oceanographic factors and also the controllable variables such as the maneuverability factors. However, to pinpoint what factors are affecting the shaft power the most, the black box of the prediction model has to be disclosed. The Explainable Artificial Intelligence (XAI) approach is presented by utilizing the SHAP methodology to answer the next research question of which among the predictors affects the shaft power the most.

4.3. Explainable Machine Learning with SHAP

Previously, it is presented that with Random Forest Regressor, a prediction model was developed and predicted the shaft power of the overall dataset, as well as the dataset from each distinct voyage number. Then, the model and test data were fed to the explainable AI tools, SHAP, that were utilized in this research.
In SHAP methodology, the calculated Shapley values are presented in several types of visualization. In this research, a beeswarm type of summary plot is presented to show the overall feature importance of all features by the calculated Shapley values, and the averaged SHAP value over the entire data for every feature is presented using the mean absolute SHAP value bar plot. Then, deeper into the mathematical representation, the dependence plot and partial dependence plot will show how each feature (or at least the top 3 features) contributes to the model output.
Presented in Figure 7, a summary plot visualized the Shapley values of each feature with respect to its impact on the model output along with the representation of each feature’s values. The color bar showed the real feature values where red indicates a higher value and blue otherwise. Each data point formed the horizontal plotting forming a violin shape. On the left side, the features are listed from top to the bottom in order of feature importance. Thus, SPEED_VG turned out to be the most important feature affecting the prediction of shaft power followed by REL_WIND_SPEED and AvgDraft.
Additionally, Figure 8 showed the mean absolute SHAP value for each feature that disclosed the average Shapley value of one predictor that contribute to the average prediction result. For example, for SPEED_VG, it is shown that for the average prediction result, or the base value, that is equal to 5540.4589 based on the SHAP calculation, the SPEED_VG averagely contributes to the 983.74 points of it.
These results drew the conclusion that, from the analysis of the prediction model towards the overall data of the 24 voyages, the speed over the ground of the vessel highly affects the prediction result. Simply, based on the SPEED_VG data fed into the model as the predictor, when the vessel moves at a faster speed over the ground, there will likely be bigger shaft power generated to move the propeller.
However, it is hard to tell the exact magnitude of SPEED_VG (or the other features) that affects the prediction results. To better understand this SHAP interpretation up to this, individualized attribution of the three most important features based on its Shapley Values are added respectively in Figure 9, Figure 10, Figure 11 and Figure 12, using a visualization called Dependence Plot (a) and Partial Dependence Plot (b).
Generally, the dependence plot showed the distribution of each feature’s Shapley value for each data instance and the red color represents high real values of the feature while blue is the opposite. Then, how the feature factual value impacts the prediction result can numerically be examined using the Partial Dependence Plot that shows the expected output of the prediction model.
First of all, as shown in Figure 9a, the dependence plot of feature SPEED_VG showed a constant tendency which is to have higher Shapley values given the higher feature value. Also, when the value falls below 16 knots, the SPEED_VG will be affecting the prediction result negatively or contributes a negative impact to the shaft power calculation as represented by the SHAP value line below 0.
To examine the attribution in a mathematical manner, Figure 9b showed the gray horizontal line in the plot that represents the expected value of the model when applied to the whole dataset. The vertical grey line represents the average value of the median feature value. So, for SPEED_VG, when the value is more than 16 knots, it will result in the expected value calculated for the target feature on the value above the horizontal line.
SPEED_VG is indeed the feature that is highly correlated with the target variable. It is a controllable variable as the value was controlled by the vessel operator along with other features like “AvgDraft” and “RudD” related to rudder angle. Thus, these controllable variables may have obvious feature attribution that is supposedly presented in the forms of the SHAP dependence plot and partial dependence plot.
If you see in Figure 10, the first and second figures showed the dependence plot of SPEED_VG and AvgDraft with interaction value. We can see how two features that are highly correlated contribute to the prediction result.
For REL_WIND_SPEED, is seen in Figure 11a that the positive impact of REL_WIND_SPEED on the prediction result happened when its value is relatively above 10.5 m/s.
The same manner can be practiced for the other feature as well, AvgDraft. The average draft the vessel maneuvered during its voyage affected the prediction results positively most likely when the draft of the ship (either draft aft or draft forward), showed the average draft above 8 will be more likely to increase the prediction results.
Additionally, this research also provided the analysis of vessel shaft power prediction over each distinct voyage number. Data were separated based on the recorded voyage number and a machine-learning model was run over each of the subsets.
Figure 13 showed the feature importance ranks based on the mean of Shapley values for each feature calculated by SHAP for analysis of all data and each distinct voyage number. Feature lists on the left side of the table showed sorted rank of feature importance overall data. Then, for the rest of the table, it presented the changes of rank for each respectable feature in the first column over each voyage number.
The top three features based on their average Shapley values representing their impact on the prediction result fluctuated in the terms of their feature importance ranks when the model analyze different voyage numbers.

5. Conclusions

Aligned with the initiative to reduce gas emissions overall transportation mode, as the major medium of transportation, the maritime industry has a big responsibility to the progression of the initiative. As one of the main measurements of a ship’s performance besides the speed, shaft power generated by the propulsion system motored by the vessel engine is analyzed in this research to find the best prediction model of this magnitude. This study began with three initial research questions.
  • Is it possible to predict the vessel shaft power considering the influence of uncontrollable variables such as the oceanographic factors and also the controllable variables such as the maneuverability factors?
  • Among the uncontrollable and controllable variables affecting the generation of vessel shaft power, what factors deploy the significant influence and how?
  • Does a different vessel voyage (trip) change the important factors affecting the vessel shaft power?
First of all, by considering oceanographic and maneuverability factors recorded during a real vessel voyage, utilizing Random Forest Regressor, this research reached a satisfactory result of prediction represented by the error rate and prediction accuracy. R-squared and MAPE of the prediction overall data showed 95% accuracy and 5% of error, while the average prediction accuracy of the analysis per distinct voyage number showed a higher score of 99% R-squared, while it resulted in a 5% of error on average. Thus, research question number 1 is attended.
Then, by implementing the SHAP approach, the model predicting the shaft power is explained by calculating each of the features of Shapley values. SHAP results feature the importance of model predictors, and the rank is presented in Figure 9, the SHAP Summary plot. Shown that the speed of the vessel (SPEED_VG) is the most affecting feature with respect to the overall prediction result, which is obvious because the generation of shaft power is linearly correlated with the generation of the speed of the ship. Comes as the second most affecting feature is one of the oceanographic factors, which is relative wind speed. Then, the third one is the average draft of the vessel. This answered question number 2.
For question number 3, it is proven that different voyage numbers or different vessel trips can change the feature importance of factors affecting the vessel shaft power. Certain features are seen to have stable ranks with respect to the Shapley value feature importance rank on different voyage numbers. However, the less important features are changing in terms of the impact of their value on the prediction results as represented by the fluctuated rank of the feature across different voyages.
Further work as the continuation of this research can consider a deeper analysis of what makes different vessel voyages affecting the result of SHAP feature importance. The indication is that different vessel voyages sailed above different sea profiles.

Author Contributions

Conceptualization, D.K.; methodology, J.L.; formal analysis, M.P.H.; data curation, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Ministry of Trade, Industry and Energy of the Republic of Korea and Korea Institute for Advancement of Technology (KIAT) through the National Innovation Cluster R&D program (Grant Number: P0015306).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cullinane, K.; Cullinane, S. Atmospheric Emissions from Shipping: The Need for Regulation and Approaches to Compliance. Transp. Rev. 2013, 33, 377–401. [Google Scholar] [CrossRef]
  2. International Maritime Organization. Fourth IMO Greenhouse Gas Study 2020; International Maritime Organization: London, UK, 2020. [Google Scholar]
  3. Zheng, W.; Makhloufi, A.; Chen, Y.; Tang, J. Decarbonizing the International Shipping Industry: Solutions and Policy Recommendations. Mar. Pollut. Bull. 2018, 126, 428–435. [Google Scholar]
  4. Soultatis, C. Systems Modeling for Electric Ship Design; Massachusetts Institute of Technology: Cambridge, MA, USA, 2004. [Google Scholar]
  5. Zhang, C.; Bin, J.; Wang, W.; Peng, X.; Wang, R.; Halldearn, R.; Liu, Z. AIS Data Driven General Vessel Destination Prediction: A Random Forest Based Approach. Transp. Res. Part C Emerg. Technol. 2020, 118, 102729. [Google Scholar] [CrossRef]
  6. Holtrop, J. A Statistical Analysis of Performance Test Results. Int. Shipbuild. Prog. 1977, 24, 23–28. [Google Scholar] [CrossRef]
  7. Holtrop, J. A Statistical Re-analysis of Resistance and Propulsion Data. Int. Shipbuild. Prog. 1984, 31, 272–276. [Google Scholar]
  8. Holtrop, J.; Mennen, G. A Statistical Power Prediction Method. Int. Shipbuild. Prog. 1978, 25, 253–256. [Google Scholar] [CrossRef]
  9. Holtrop, J.; Mennen, G. An Approximate Power Prediction Method. Int. Shipbuild. Prog. 1982, 29, 166–170. [Google Scholar] [CrossRef]
  10. Petersen, J.P.; Jacobsen, D.J.; Winther, O. Statistical Modeling for Ship Propulsion Efficiency. J. Mar. Sci. Technol. 2012, 17, 20–39. [Google Scholar] [CrossRef]
  11. Petersen, J.P.; Winther, O.; Jacobsen, D.J. A Machine-Learning Approach to Predict Main Energy Consumption under Realistic Operational Conditions. Ship Technol. Res. 2012, 59, 64–72. [Google Scholar] [CrossRef]
  12. Radonjic, A.; Vukadinovic, K. Application of Ensemble Neural Networks to Pediction of Towboat Shaft Power. J. Mar. Sci. Technol. 2014, 20, 64–80. [Google Scholar] [CrossRef]
  13. Coraddu, A.; Oneto, L.; Baldi, F.; Anguita, D. Vessel Fuel Consumption Forecast and Trim Optimisation: A Data Analytics Perspective. Ocean Eng. 2017, 130, 351–370. [Google Scholar] [CrossRef]
  14. Kim, D.-H.; Han, S.-J.; Jung, B.-K.; Han, S.-H.; Lee, S.-B. A Machine Learning-Based Method to Predict Engine Power. J. Korean Soc. Mar. Environ. Saf. 2019, 25, 851–857. [Google Scholar] [CrossRef]
  15. Kim, D.-H.; Lee, S.-B.; Lee, J.-H. Data-Driven Prediction of Vessel Propulsion Power Using Support Vector Regression with Onboard Measurement and Ocean Data. Sensors 2020, 20, 1588. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Lang, X.; Wu, D.; Mao, W. Benchmark Study of Supervised Machine Learning Methods for a Ship Speed-Power Prediction at Sea. In Proceedings of the ASME 40th International Conference on Ocean, Offshore, and Arctic Engineering, Virtual, 21–30 June 2021. [Google Scholar]
  17. Laurie, A.; Anderlini, E.; Dietz, J.; Thomas, G. Machine learning for shaft power prediction and analysis of fouling related performance deterioration. Ocean Eng. 2021, 234, 108886. [Google Scholar] [CrossRef]
  18. Khosravi, H.; Shum, S.B.; Chen, G.; Conati, C.; Tsai, Y.-S.; Kay, J.; Knight, S.; Martinez-Maldonado, R.; Sadiq, S.; Gašević, D. Explainable Artificial Intelligence in education. Comput. Educ. Artif. Intell. 2022, 3, 100074. [Google Scholar] [CrossRef]
  19. Jacinto, M.; Silva, M.; Medeiros, G.; Oliveira, L.; Montalvão, L.; de Almeida, R.V.; Ninci, B. Explainable Artificial Intelligence for O&G Machine Learning Solutions: An Application to Lithology Prediction. In Proceedings of the 83rd EAGE Annual Conference & Exhibition, Madrid, Spain, 6–9 June 2022. [Google Scholar]
  20. Ali, A.; Aliyuda, K.; Elmitwally, N.; Bello, A.M. Towards more accurate and explainable supervised learning-based prediction of deliverability for underground natural gas storage. Appl. Energy 2022, 327, 120098. [Google Scholar] [CrossRef]
  21. Cohausz, L. Towards Real Interpretability of Student Success Prediction Combining Methods of XAI and Social Science. In Proceedings of the International Conference on Educational Data Mining (EDM), Durham, UK, 24–27 July 2022. [Google Scholar]
  22. Abioye, S.O.; Oyedele, L.O.; Akanbi, L.; Ajayi, A.; Delgado, J.M.D.; Bilal, M.; Akinade, O.O.; Ahmed, A. Artificial intelligence in the construction industry: A review of present status, opportunities and future challenges. J. Build. Eng. 2021, 44, 103299. [Google Scholar] [CrossRef]
  23. Love, P.E.; Fang, W.; Matthews, J.; Porter, S.; Luo, H.; Ding, L. Explainable Artificial Intelligence (XAI): Precepts, Methods, and Opportunities for Research in Construction. arXiv 2022, arXiv:2211.06579. [Google Scholar]
  24. Gaur, L.; Sahoo, B.M. Introduction to Explainable AI and Intelligent Transportation. In Explainable Artificial Intelligence for Intelligent Transportation Systems; Springer: Cham, Switzerland, 2022. [Google Scholar]
  25. Procopiou, A.; Chen, T.M. Explainable AI in Machine/Deep Learning for Intrusion Detection in Intelligent Transportation Systems for Smart Cities. In Explainable Artificial Intelligence for Smart Cities; CRC Press: Boca Raton, FL, USA, 2021; pp. 297–321. [Google Scholar]
  26. Chaddad, A.; Peng, J.; Xu, J.; Bouridane, A. Survey of Explainable AI Techniques in Healthcare. Sensors 2023, 23, 634. [Google Scholar] [CrossRef]
  27. Sheu, R.-K.; Pardeshi, M.S. A Survey on Medical Explainable AI (XAI): Recent Progress, Explainability Approach, Human Interaction and Scoring System. Sensors 2022, 22, 8068. [Google Scholar] [CrossRef]
  28. Shin, D. The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. Int. J. Hum. Comput. Stud. 2021, 146, 102551. [Google Scholar] [CrossRef]
  29. Limna, P. Artificial Intelligence (AI) in the Hospitality Industry: A Review Article. Int. J. Comput. Sci. Res. 2022, 6, 1–12. [Google Scholar]
  30. Hrnjica, A.; Softic, S. Explainable AI in Manufacturing: A Predictive Maintenance Case Study. In Proceedings of the IFIP International Conference on Advances in Production Management Systems, Novi Sad, Serbia, 30 August–3 September 2020. [Google Scholar]
  31. Hore, S.; Bhattacharya, T. A Smart System for the Assessment of Genuineness or Trustworthiness of the Tip-Off Using Audio Signals: An Explainable AI Approach. In Explainable AI: Foundations, Methodologies and Applications. Intelligent Systems Reference Library; Springer: Cham, Switzerland, 2022; pp. 185–209. [Google Scholar]
  32. Atkinson, K.; Bench-Capon, T.; Bollegala, D. Explanation in AI and law: Past, present, and future. Artif. Intell. 2020, 289, 103387. [Google Scholar] [CrossRef]
  33. Giudici, P.; Raffinetti, E. Explainable AI methods in cyber risk management. Qual. Reliab. Eng. Int. 2022, 38, 1318–1326. [Google Scholar] [CrossRef]
  34. Sachan, S.; Yang, J.-B.; Xu, D.-l.; Benavides, D.E.; Li, Y. An explainable AI decision-support-system to automate loan underwriting. Expert Syst. Appl. 2020, 144, 113100. [Google Scholar] [CrossRef]
  35. Misheva, H.; Osterrieder, J.; Hirsa, A.; Kulkarni, O.; Lin, S.F. Explainable AI in Credit Risk Management. arXiv 2021, arXiv:2103.00949. [Google Scholar]
  36. Kim, D.; Antariksa, G.; Handayani, M.P.; Lee, S.; Lee, J. Explainable Anomaly Detection Framework for Maritime Main Engine Sensor Data. Sensors 2021, 21, 5200. [Google Scholar] [CrossRef] [PubMed]
  37. Opitz, A.; Maclin, R. Popular ensemble methods: An empirical study. J. Artif. Intell. Res. 1999, 11, 169–198. [Google Scholar] [CrossRef]
  38. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  39. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  40. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  41. Ke, A.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  42. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  43. Bentéjac, C.; Csörgő, A.; Martínez-Muñoz, G. A Comparative Analysis of Gradient Boosting Algorithms. Artif. Intell. Rev. 2021, 54, 1937–1967. [Google Scholar] [CrossRef]
  44. Sammut, A.; Webb, G.I.W. Encyclopedia of Machine Learning; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  45. Hancock, R.; Freeman, M.J. Power and Sample Size for the Root Mean Square Error of Approximation Test of not Close Fit in Structural Equation Modeling. Educ. Psychol. Meas. 2001, 61, 741–758. [Google Scholar] [CrossRef]
  46. de Myttenaere, A.; Golden, B.; Grand, B.L.; Rossi, F. Mean Absolute Percentage Error for regression models. Neurocomputing 2016, 192, 38–48. [Google Scholar] [CrossRef] [Green Version]
  47. Chicco, D.; Warrens, M.J.; Jurman, G. The coefficient of determination R-squared is more informative than SMAPE, MAE, MAPE, MSE and RMSE in regression analysis evaluation. PeerJ Comput. Sci. 2021, 7, e623. [Google Scholar] [CrossRef] [PubMed]
  48. Wright, S. Correlation and causation. J. Agric. Res. USDA 1921, 557–585. [Google Scholar]
  49. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the KDD ‘16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  50. Lundberg, S.M.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  51. Ribeiro, M.T.; Singh, S.; Guestrin, C. Anchors: High-Precision Model-Agnostic Explanations. In Proceedings of the AAAI Conference of Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  52. Shapley, L. A Value for n-person Games. In Contribution to the Theory of Games II; Kuhn, H., Tucker, A., Eds.; Princeton University Press: Princeton, NJ, USA, 1953; pp. 307–317. [Google Scholar]
  53. Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.-I. From Local Explanations to Global Understanding with Explainable AI for Trees. Nat. Mach. Intell. 2020, 2, 56–67. [Google Scholar] [CrossRef]
  54. Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.-I. Explainable AI for Trees: From Local Explanations to Global Understanding. arXiv 2019, arXiv:1905.04610. [Google Scholar] [CrossRef]
  55. Li, Z. Extracting spatial effects from machine learning model using local interpretation method: An example of SHAP and XGBoost. Comput. Environ. Urban Syst. 2022, 96, 101845. [Google Scholar] [CrossRef]
  56. Lundberg, S. SHAP. 2018. Available online: https://shap.readthedocs.io/ (accessed on 28 November 2022).
  57. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  58. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: New York, NY, USA, 2009. [Google Scholar]
Figure 1. Vessel voyage routes.
Figure 1. Vessel voyage routes.
Sensors 23 01072 g001
Figure 2. SHAP concept.
Figure 2. SHAP concept.
Sensors 23 01072 g002
Figure 3. TreeSHAP concept.
Figure 3. TreeSHAP concept.
Sensors 23 01072 g003
Figure 4. SHAP partial dependence plot illustration.
Figure 4. SHAP partial dependence plot illustration.
Sensors 23 01072 g004
Figure 5. Experimental framework.
Figure 5. Experimental framework.
Sensors 23 01072 g005
Figure 6. Directional feature transformation.
Figure 6. Directional feature transformation.
Sensors 23 01072 g006
Figure 7. SHAP summary plot.
Figure 7. SHAP summary plot.
Sensors 23 01072 g007
Figure 8. Mean absolute SHAP value bar plot.
Figure 8. Mean absolute SHAP value bar plot.
Sensors 23 01072 g008
Figure 9. SHAP dependence plot (a) and partial dependence plot (b) of SPEED_VG.
Figure 9. SHAP dependence plot (a) and partial dependence plot (b) of SPEED_VG.
Sensors 23 01072 g009
Figure 10. SHAP dependence plot of SPEED_VG with AvgDraft as interaction feature (a) and AvgDraft with SPEED_VG as interaction feature (b).
Figure 10. SHAP dependence plot of SPEED_VG with AvgDraft as interaction feature (a) and AvgDraft with SPEED_VG as interaction feature (b).
Sensors 23 01072 g010
Figure 11. SHAP dependence plot (a) and partial dependence plot (b) of REL_WIND_SPEED.
Figure 11. SHAP dependence plot (a) and partial dependence plot (b) of REL_WIND_SPEED.
Sensors 23 01072 g011
Figure 12. SHAP dependence plot (a) and partial dependence plot (b) of AvgDraft.
Figure 12. SHAP dependence plot (a) and partial dependence plot (b) of AvgDraft.
Sensors 23 01072 g012
Figure 13. SHAP feature importance ranking heatmap.
Figure 13. SHAP feature importance ranking heatmap.
Sensors 23 01072 g013
Table 1. Vessel specifications.
Table 1. Vessel specifications.
SpecificationMeasurement
Length (m)172.07
Beam (m)27.4
Gross Tonnage18,085
Deadweight (t)22,317
TEU1809
Year Build2020
Table 2. Original feature lists.
Table 2. Original feature lists.
Feature NameDescription
ME1_SHAFT_POWERPropeller shaft power of the vessel
SPEED_VGSpeed over ground
DRAFT_FOREVertical distance between the waterline and the bottom of the hull measured at the perpendicular of the bow
DRAFT_AFTVertical distance between the waterline and the bottom of the hull measured at the perpendicular of the stern
SHIP_HEADINGDirection of which the vessel is sailing
TOTAL_WAVE_HEIGHTVertical distance between the crest (peak) and the trough of a wave
TOTAL_WAVE_DIRECTIONDirection of which the wave is moving
CURRENT_SPEEDDirectional movement of seawater driven by gravity, wind (Coriolis Effect), and water density
CURRENT_DIRECTIONDirection of seawater
WIND_SPEEDSpeed of the geographic or ground wind, assuming no tidal flow
WIND_DIRECTIONDirection of the geographic or ground wind
REL_WIND_SPEEDSpeed of the apparent wind
REL_WIND_DIRDirection of the apparent wind
RUDDER_ANGLEThe angle of the vessel rudder
WATER_DEPTHThe depth of water where the ship sails
SEA_SURFACE_SALINITYAmount of salt on the body of the seawater
SEA_SURFACE_TEMPTemperature on the seawater
Table 3. Features the summary statistics.
Table 3. Features the summary statistics.
Feature NameMinMeanMaxStandard Deviation
ME1_SHAFT_POWER04072.269681.592920.40
SPEED_VG−9999.0011.482122.447.9097
REL_WIND_SPEED−9999.007.449229.524111.215
REL_WIND_DIR−9999.00184.9379360181.6887
DRAFT_FORE2.6337.579210.2971.4717
DRAFT_AFT5.7078.114710.0710.8196
SHIP_HEADING−9999.00144.976359.998135.9825
TOTAL_WAVE_HEIGHT−9999.00−141.7388889.151188.97
TOTAL_WAVE_DIRECTION−9999.005.85283601208.92
CURRENT_SPEED−9999.00−142.89921.891187.64
CURRENT_DIRECTION−9999.0030.86153601212.59
WIND_SPEED−9999.00−136.918827.251188.37
WIND_DIRECTION−9999.005.45833601209.35
RUDDER_ANGLE−9999.00−24.872836.8512.351
WATER_DEPTH−9999.00−1341.97834.63469.26
SEA_SURFACE_SALINITY−9999.00−111.022434.681191.48
SEA_SURFACE_TEMP−9999.00−120.984532.931190.31
Table 4. Error data counts.
Table 4. Error data counts.
FeatureError Instances CountError Ratio (%)
ME1_SHAFT_POWER00.00
SPEED_VG20.00
DRAFT_FORE00.00
DRAFT_AFT00.00
SHIP_HEADING70.01
TOTAL_WAVE_HEIGHT12781.43
TOTAL_WAVE_DIRECTION12781.43
CURRENT_SPEED12781.43
CURRENT_DIRECTION12781.43
WIND_SPEED12781.43
WIND_DIRECTION12781.43
RUDDER_ANGLE2350.26
SEA_SURFACE_SALINITY12781.43
SEA_SURFACE_TEMP12781.43
REL_WIND_SPEED110.01
REL_WIND_DIR110.01
WATER_DEPTH12,35213.83
Table 5. Features the summary statistics.
Table 5. Features the summary statistics.
Feature NameMinMeanMaxStandard Deviation
ME1_SHAFT_POWER03810.35919681.5942939.5802
SPEED_VG011.099222.47.2974
REL_WIND_SPEED08.560629.5245.7281
REL_WIND_DIR0187.1479360141.6656
DRAFT_FORE2.6337.448810.2971.4857
DRAFT_AFT5.7078.063110.0710.8459
SHIP_HEADING0148.0478359.998102.9816
TOTAL_WAVE_HEIGHT0.010.96485.970.8011
TOTAL_WAVE_DIRECTION0.01158.993236091.5081
CURRENT_SPEED00.1841.890.1725
CURRENT_DIRECTION0176.654736099.3439
WIND_SPEED0.036.186827.253.6007
WIND_DIRECTION0156.57360.0097.79
RUDDER_ANGLE−34.71.415336.81.8461
WATER_DEPTH0.748.3948834.675.1042
SEA_SURFACE_SALINITY17.5732.356334.682.3481
SEA_SURFACE_TEMP−1.7221.627632.628.2854
Table 6. Feature list after preprocessing.
Table 6. Feature list after preprocessing.
Feature NameDescription
ME1_SHAFT_POWERPropeller shaft power of the vessel
SPEED_VGSpeed over ground
TOTAL_WAVE_HEIGHTVertical distance between the crest (peak) and the trough of a wave
CURRENT_SPEEDDirectional movement of seawater driven by gravity, wind (Coriolis Effect), and water density
WIND_SPEEDSpeed of the geographic or ground wind, assuming no tidal flow
REL_WIND_SPEEDSpeed of the apparent wind
WATER_DEPTHThe depth of water where the ship sails
SEA_SURFACE_SALINITYAmount of salt on the body of the seawater
SEA_SURFACE_TEMPTemperature on the seawater
AvgDraft 1Averaged value of DRAFT_FORE and DRAFT_AFT
CurrD 1Angular difference between SHIP_HEADING and
CURRENT_DIRECTION
WaveD 1Angular difference between SHIP_HEADING and
TOTAL_WAVE_DIRECTION
WindD 1Angular difference between SHIP_HEADING and
WIND_DIRECTION
RelWD 1Angular difference between SHIP_HEADING and
REL_WIND_DIR
RudD 1Angular difference between SHIP_HEADING and
RUDDER_ANGLE
1 Transformed features.
Table 7. Hyperparameter tuning attribution.
Table 7. Hyperparameter tuning attribution.
ModelParameters
Grid ParametersGrid ValuesTuned Values
Random Forest Regressorn_estimators
max_depth
min_samples_split
min_samples_leaf
: [10,25,50,75,100]
: [10,25,50,75,100]
: [2,4,6,8,10]
: [1,2,3,4,5]
{n_estimators: 100},
{max_depth: 25},
{min_samples_split: 2},
{min_samples_leaf: 1}
CatBoost Regressordepth
learning_rate
iterations
: [10,25,50]
: [0.1,0.5,1]
: [50,100,250]
{depth: 10},
{learning_rate: 0.1},
{iterations: 250}
Extreme Gradient Boosting (XGB) Regressorn_estimators
max_depth
learning_rate
: [50,100,150,200,250]
: [5,10,25,50]
: [0,0.5,1]
{n_estimators: 250},
{max_depth: 7},
{learning_rate: 0.01}
Light Gradient Boosting Machine (LightGBM) Regressor)n_estimators
max_depth
learning_rate
: [50,75,100]
: [10,50,100]
: [0.05,0.1,0.5,1]
{n_estimators: 100},
{max_depth: 10},
{learning_rate: 0.5}
Table 8. Comparative performance measurements.
Table 8. Comparative performance measurements.
ModelR-Squared ScoreRMSEMAEMAPE
TrainTestTrainTestTrainTestTrainTest
Random Forest Regressor0.990.95156.95414.8882.76221.000.020.05
CatBoost Regressor0.980.95237.69393.73164.66248.590.040.06
Extreme Gradient Boosting (XGB) Regressor1.000.942.84446.561.88263.600.000.06
Light Gradient Boosting Machine (LightGBM) Regressor)0.980.94221.95449.48159.27289.120.030.07
Average Scoring0.990.95154.86426.16102.14255.580.020.06
Table 9. Performance metrics per voyage number.
Table 9. Performance metrics per voyage number.
VOY.NO#InstancesR-Squared ScoreRMSEMAEMAPE
TrainTestTrainTestTrainTestTrainTest
15220.980.93105.85210.4437.7077.920.020.03
218520.990.95115.49302.6658.23146.600.020.04
323910.990.92157.32502.2677.55228.380.020.06
423670.990.92158.76432.5975.43201.870.020.05
521920.990.93163.04384.8667.86166.490.020.04
622580.990.92131.93351.8058.13155.410.020.04
722600.980.90164.63313.5165.74138.470.020.05
824580.970.79156.69391.2262.43160.650.020.05
917960.980.85147.04420.9758.71164.660.020.05
1021120.990.94131.45339.4658.26148.700.020.04
1121830.990.93199.58484.9690.11219.330.030.06
1220640.990.94134.09419.9168.52209.990.020.06
1316410.990.95179.03353.5873.79169.670.020.04
1422530.990.93150.80376.1161.88179.480.020.05
1524180.990.95177.57492.0088.89238.770.020.05
1620650.990.93183.39517.2299.79285.360.020.07
1720631.000.96160.98421.9475.34213.150.020.06
1818770.990.96169.40448.2579.10214.770.020.05
195600.990.93147.37328.3674.77194.510.010.03
2020410.990.96160.51459.9178.96213.190.020.05
2119671.000.96169.01456.1568.84195.630.020.06
2219690.990.94162.54426.9676.08198.780.020.05
2327730.990.95164.23402.2074.28189.410.020.05
2413620.990.96149.60391.09107.18216.510.030.07
Average per Voyage19770.990.93155.85401.1872.40188.650.020.05
All Data47,4440.990.95156.95414.8882.76221.000.020.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, D.; Handayani, M.P.; Lee, S.; Lee, J. Feature Attribution Analysis to Quantify the Impact of Oceanographic and Maneuverability Factors on Vessel Shaft Power Using Explainable Tree-Based Model. Sensors 2023, 23, 1072. https://doi.org/10.3390/s23031072

AMA Style

Kim D, Handayani MP, Lee S, Lee J. Feature Attribution Analysis to Quantify the Impact of Oceanographic and Maneuverability Factors on Vessel Shaft Power Using Explainable Tree-Based Model. Sensors. 2023; 23(3):1072. https://doi.org/10.3390/s23031072

Chicago/Turabian Style

Kim, Donghyun, Melia Putri Handayani, Sangbong Lee, and Jihwan Lee. 2023. "Feature Attribution Analysis to Quantify the Impact of Oceanographic and Maneuverability Factors on Vessel Shaft Power Using Explainable Tree-Based Model" Sensors 23, no. 3: 1072. https://doi.org/10.3390/s23031072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop