Next Article in Journal
Sensitivity Analysis of Factors Affecting the Stability of Deep Buried Tunnel
Next Article in Special Issue
A Research on Autonomous Collision Avoidance under the Constraint of COLREGs
Previous Article in Journal
Heat Transfer Studies on Solar Parabolic trough Collector Using Corrugated Tube Receiver with Conical Strip Inserts
Previous Article in Special Issue
Simulating the Impact of the Sustained Melting Arctic on the Global Container Sea–Rail Intermodal Shipping
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Regression Analysis Using Broad Learning System for Time-Series Forecast of Ship Fuel Consumption

1
Navigation College, Dalian Maritime University, Dalian 116026, China
2
Maritime Big Data & Artificial Intelligent Application Centre, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(1), 380; https://doi.org/10.3390/su15010380
Submission received: 11 November 2022 / Revised: 6 December 2022 / Accepted: 15 December 2022 / Published: 26 December 2022
(This article belongs to the Special Issue Green Maritime Logistics and Sustainable Port Development)

Abstract

:
Accurately forecasting the fuel consumption of ships is critical for improving their energy efficiency. However, the environmental factors that affect ship fuel consumption have not been researched comprehensively, and most of the relevant studies continue to present efficiency and accuracy issues. In view of such problems, a time-series forecasting model of ship fuel consumption based on a novel regression analysis using broad learning system (BLS) was developed in this study. The BLS was compared to a diverse set of fuel consumption forecasting models based on time-series analyses and machine learning techniques, including autoregressive integrated moving average model with exogenous inputs (ARIMAX), support vector regression (SVR), recurrent neural network (RNN), long short-term memory network (LSTM), and extreme learning machines (ELM). In the experiment, two types of passenger roll-on roll-off (ro-ro) ship and liquefied petroleum gas (LPG) carrierwere used as research objects to verify the proposed method’s generalizability, with data divided among two groups (RM, RB). The experimental results showed that the BLS model is the best choice to forecast fuel consumption in actual navigation, with mean absolute error (MAE) values of 0.0140 and 0.0115 on RM and RB, respectively. For the LPG carrier, it has also been proven that the forecast effect is improved when factoring the sea condition, with MAE reaching 0.0108 and 0.0142 under ballast and laden conditions, respectively. Furthermore, the BLS features the advantages of low computing complexity and short forecast time, making it more suitable for real-world applications. The results of this study can therefore effectively improve the energy efficiency of ships by reducing operating costs and emissions.

1. Introduction

With the recently growing demand and promotion of shipping transportation, the carbon emissions of ships have increased significantly, and the sustainability of the shipping industry has attracted considerable attention from public and the governmental entities [1]. Subsequently, few shipping corporations have also focused on this issue to protect the environment with the objective of carbon neutrality, and promoting green development within the shipping industry to minimize pollution [2]. To develop an effective method to improve energy efficiency and reduce emissions reductions in the shipping industry, accurate model must be designed to forecast ship fuel consumption. In particular, it may prove very helpful to discover the principal factors affecting fuel consumption in existing ships, and extract corresponding relationships with fuel cost [3].
Many studies have been devoted to the use of time-series analyses and machine learning techniques to forecast ship fuel consumption. In a regression analysis of ship fuel consumption, Wang et al. considered the convexity, non-negativity, and univariate properties of the fuel speed function, and proposed an exact outer-approximation algorithm to address the speed optimization problem by obtaining an ε-optimal solution [4]. Gao et al. considered the influence of ship speed and ship payload on fuel consumption, and then built a mixed integer nonlinear rule for container transshipment [5]. An increasing number of factors related to fuel consumption have been collected. To effectively utilize these factors, Erto et al. proposed a method based on the multiple linear regression model to monitor fuel consumption, producing superior results to those obtained by the traditional speed–power curves method [6]. Wang et al. proposed an estimation model for fuel consumption based on the least absolute shrinkage and selection operator (LASSO). This model, which can integrate onboard selection for feature variables, was compared with several existing approaches for fuel consumption forecasting [7]. Given the increasing volume of fuel consumption-related data, as well as problems such as nonlinearity and discreteness, the linear regression (LR) approach cannot meet the demand for accurate forecasting. Accordingly, Zhu et al. proposed an integrated model and compared the performance of various machine learning methods —including LR, support vector regression (SVR), and artificial neural networks (ANN)—to obtain accurate fuel consumption forecasting results [8]. Likewise, Peng et al. compared five machine learning models—including gradient boosting regression, LR, k-nearest neighbor regression and ANN—to forecast the energy consumption of ships [9]. Yuan et al. constructed a forecast model of fuel consumption rate and speed based on long short-term memory (LSTM) to optimize resource cost [10] (Appendix A, Table A1). Although existing studies have yielded impressive results, many issues for expansion in terms of effectiveness and efficiency remain. Considering that the collected data represent time-series data, the quantity and quality of the data have a significant impact on forecasting performance. Therefore, an appropriate model must be developed for the characteristics of time-series data. Furthermore, most prior studies only considered onboard fuel consumption impact variables, resulting in insufficient data volume and sources. Given the complexity of the shipping process, more consideration should be given to the marine environment and weather conditions.
In response to the aforementioned problems, this study was conducted to examine passenger roll-on roll-off (ro-ro) ship and liquefied petroleum gas (LPG) carrier as experimental targets, consider of real navigating conditions, and builds a forecast model of fuel consumption based on a novel regression analysis using a broad learning system (BLS) [11]. This study fills a gap in the application of BLS to time-series analyses for fuel consumption. To determine a model that demands minimal time and is more accurate with different structures and complexity, BLS was compared with autoregressive integrated moving average model with exogenous inputs (ARIMAX) [12], SVR [8], recurrent neural network (RNN), LSTM [10] and extreme learning machines (ELM) [13], which are models based on time-series analyses and machine learning techniques. According to the characteristics of different models to demonstrated the applicability and accuracy by systematic verification analysis. Two types of ships were considered in this study to verify the models’ generalizability. Furthermore, to enable the forecasting model’s operation in the real navigation environment and adaptability to possible changes, this study considered the internal and external factors relating to fuel consumption. According to the sea state and weather conditions, the data of the two ships are divided into two groups for separate discussion (Appendix A, Table A2). Passenger ro-ro ship data were allocated among two different groups— R M and R B —according to the route for analysis. The two groups obtained mean absolute error (MAE) values of 0.0140 and 0.0115, respectively. When the BLS model was applied to the LPG carrier under consideration of sea conditions ( C 1 ), the MAE under ballast and laden reached values 0.0108 and 0.0142, respectively. These values represent decreases of 28.00% and 32.70% compared with the corresponding MAE results when sea conditions were not considered ( C 2 ). We therefore conclude that the sea conditions must be considered for LPG carriers in either state. Furthermore, abundant experimental results indicate that all the forecast consumption times incurred by the BLS model for both cases were superior to those obtained by other models.
The remainder of this paper is organized as follows. Section 2 briefly introduces ship fuel consumption forecasting based on time-series analyses and machine learning techniques. The proposed approach and corresponding data are presented in Section 3. Comparative experimental results are discussed in Section 4. Finally, Section 5 presents our conclusions.

2. Literature Review of Forecasting Methods

2.1. Time-Series Analyses

The development of fuel consumption forecast models based on time-series analyses uses historical data as a forecasting tool through the inherent correlation between time- and fuel consumption-related data in various dimensions. Traditional approaches primarily solve the model parameters by determining the time-series parameter model, and uses the solved model to complete the forecast. Conventional models, including the autoregressive average (AR), moving average (MA), autoregressive integrated moving average (ARIMA) [14], and ARIMAX models, are introduced in this section.
Suppose that at time t , x t represents the fuel consumption of the time series, and u t and ε t denote the error term and the white noise, respectively.
The AR model is a linear model. When u t = ε t , p represents the dimension of the time window and indicates that x t is forecasted using the past fuel consumption x t j and ε t of period p . This relation is expressed in Equation (1), where c and ϕ j   denote the model parameter and weight coefficient, respectively [14].
x t = c + j = 1 p ϕ j x t j + ε t
When x t depends only on the linear combination of historical white noise, it can be expressed as x t = u t . From this, the formula for MA is as shown in Equation (2), where θ h denotes the weight coefficient.
x t = c + h = 1 q θ h ε t h + ε t
The ARIMA model was obtained by combining AR and MA using the difference method, as expressed in Equation (3).
x t = μ + j = 1 p ϕ j x t j + h = 1 q θ h ε t h + ε t
For multivariate time-series forecasting, the ARIMAX model is a generalization of ARIMA model, which is capable of incorporating an exogenous input variable [12], as shown in Equation (4).
y t = μ + i = 1 k Θ i ( B ) Φ i ( B ) B l i x t i + ε t
where y t and x t i are the output variable and the i t h input variable at time t , respectively; B is the backshift operator; Φ i ( B ) is the autoregressive coefficient polynomial of the i t h input variable; Θ i ( B ) is the moving average coefficient polynomial of the i t h input variable; l i is the delay order of the i t h input variable.

2.2. Support Vector Regression

As branch of support vector machines (SVM), SVR is a machine learning method based on statistical theory that can solve regression problems [8]. In SVR model, the ship fuel consumption variable X is mapped to a higher-dimensional space using a kernel function φ for LR estimation, as shown in Equation (5) [15]. Y and Y represent real and forecasted fuel consumption, respectively. Some common kernel functions are the linear kernel, polynomial (Poly) kernel and radial basis function (RBF).
Y = W φ ( X ) + b
In the high-dimensional space, a hyperplane that minimizes the maximum distance between the mapping points and hyperplane—which is guaranteed to be less than ε —determined using the model parameters W and b . The optimization objective of SVR is shown in Equation (6) and Figure 1 [15], where f ( x ) = y ,   y i Y , y i Y , i , i = 1 , 2 , , m . Model training is guided by the error between the forecasted and real fuel consumption values.
{ min W , b 1 2 W 2 s . t . | y i y i | ε

2.3. Recurrent Neural Network

RNN is a type of recursive neural network, which is one of the most commonly used models when handling time-series problems based on machine learning methods [10]. The structure of an RNN-based fuel consumption model is illustrated in Figure 2.
As shown in the figure, the left side is folded whereas the right side is unfolded. Dashed arrows represent recurrences in the model structure embodied in the hidden layers. The activation process from the input layer to the hidden layer at time t can be expressed by Equation (7).
h t = f 1 ( U x t + W h t 1 )
where h t represents the output of neurons in the hidden layer at the time t ; U is the connection weight matrix from the input layer or previous hidden layer to the current hidden layer; W is the connection weight matrix between the previous and current time-series hidden layers; and f 1 ( U x t + W h t 1 ) is the activation function of the hidden layer, generally set to t a n h function.
The activation process from the hidden layer to the output layer is expressed by Equation (8).
o t = f 2 ( V h t )
where o t represents the output, V is the connection weight matrix from the hidden layer to the output layer, and f 2 ( x ) is the activation function of the output layer, generally set to the sigmoid function.
The RNN uses the hidden node state h t 1 at the last time point as the input to the neural network unit. In other words, weight connections are established between different layers’ neurons. This may explain the excellent performance exhibited by RNNs when handling time-series data. Neurons in the hidden layer also have weights, with corresponding weight connections W ,   U ,   V [16].

2.4. Long Short-Term Memory

LSTM is a special type of RNN that incorporates a memory cell state to memorize the long-term state of information [10]. The underlying process of an LSTM-based fuel consumption model is illustrated in Figure 3.
As shown in the figure, LSTM controls the circulation and loss of features through a gate mechanism.
For the forget gate f t , the cell state can be represented as c t 1 , c t , c t + 1 . It combines the state value h t 1 of the previous hidden layer with the current input x t , and outputs an f t through the s i g m o i d function σ to determine what information to discard from the cell state c t 1 . The process is expressed by Equation (9), where σ is the activation function, and W f and b f are the weight and bias, respectively.
f t = σ ( W f [ h t 1 ,   x t ] + b f )
Next, the input gate i t and t a n h function determine the new information that can be added to the cell state from the activation value h t 1 of the hidden layer at the previous moment and current fuel consumption input variable x t , thereby obtaining the candidate value c t ˜ . This procedure is expressed by Equations (10) and (11):
i t = σ ( W i [ h t 1 ,   x t ] + b i )
c t ˜ = t a n h ( W c [ h t 1 ,   x t ] + b c )
The next step updates the cell state. As shown in Equation (12), the forget gate f t and the input gate i t are combined, and appropriate information is saved and discarded to obtain the cell state c t at the current moment.
c t = f t c t 1 + i t c t ˜
Finally, the output gate o t combines with t a n h to determine which information of h t 1 , x t , c t is the output to the hidden layer state h t at this moment. The specific method is expressed by Equations (13) and (14), wherein the σ activation function is used to determine the part of the cell state that needs to be output, and the t a n h function is used to obtain the backup output 16.
o t = σ ( W o [ h t 1 ,   x t ] + b o )
h t = o t t a n h ( c t )

2.5. Extreme Learning Machine

The ELM is a special single-hidden layer feedforward neural networks (SLFNs) proposed by Huang et al., comprising an input layer, hidden layer and output layer [13]. In the ELM training process, the weight and bias of the hidden layer are generated randomly and do not require updating. The weight of the output layer is calculated to complete the training process. Consequently, ELM is characterized by a short learning time and superior generalization performance. The output function of ELM is given by Equation (15).
f L ( x ) = i = 1 L β i G ( w i , b i , x )
where L indicates the number of hidden layer nodes and β i is the connection weight vector between the i t h hidden layer neuron and the output neuron. G ( w i , b i , x ) represents the output of the i t h hidden layer neuron corresponding to the input x .
G ( w i , b i , x ) = g ( w i · x + b i )
where g represents the activation function, w i is the weight vector between the i t h hidden layer neuron and input neuron, and b i is the bias of the i t h hidden layer neuron. Although the overall training process of ELM is similar to that of the traditional back-propagation neural network, the weight matrix between the hidden layer and the output is a Moore–Penrose generalized inverse matrix. Equation (17) can be abbreviated as
T = H β
where H denotes the hidden layer output matrix, and T denotes the training data target matrix. The next step is to calculate the weight β by minimizing the square approximation error, as shown in Equation (18).
min β H β T 2
The following appropriate solution can be obtained for Equation (18), where H is the Moore–Penrose generalized inverse matrix of H [17].
β * = H T

3. Material and Method

3.1. Data Source

3.1.1. Passenger Ro-Ro Ship

All passenger ro-ro ship data used in this study were provided by the Danish University of Technology, with main specifications listed in Table 1 [18]. The collected raw dataset was recorded for two consecutive months from February–April 2010, and includes the main route ( R M ) and the backup route ( R B ) adopted under bad weather conditions for R M . The following sections discuss the data characteristics of R M and R B individually.
Following preprocessing and feature selection, the number of data points for R M and R B were reduced to approximately 19,200 and 5700, respectively. Fuel consumption was designated as the output, and 11 feature variables related to fuel consumption were designated as the input. More information on the feature variables for the passenger ro-ro ship can be seen in Table A3 of Appendix A. For convenience, the names of the feature variables are abbreviated according to Table 2.

3.1.2. Liquefied Petroleum Gas Carrier

All LPG carrier data used in this study were obtained from onboard data, as well as open real-time oceanographic and meteorological data, with principal particulars listed in Table 3 [19]. The raw dataset comprises data recorded over seven consecutive months from December 2016 to July 2017, including internal and external factors related to the ship. Onboard data sources include noon reports, the ship automation system (SAS), and the electronic chart display and information system (ECDIS).
The draught of the LPG carrier did not change continuously. Specifically, it can be considered as binary, alternating between ballast or laden. Accordingly, the following chapters discuss the ballast and laden data separately.
After preprocessing and feature selection, the number of data points for the ballast and laden sets were reduced to approximately 9900 and 7400, respectively. Fuel consumption was designated as the output, and 17 feature variables related to fuel consumption were designated as the input. More information on the feature variables for the LPG carrier can be seen in Table A4 of Appendix A. The names of all feature variables were abbreviated, as listed in Table 4. In addition, the other acronyms used in the paper are shown in Table A5 of Appendix A.

3.2. Data Structure

Data corresponding to both case ships required normalization prior to the ship fuel consumption model training. Normalization essentially represents data feature scaling. In this study, the min–max scaling method was employed for data transformation. Let x i M i n and x i M a x be the minimum and maximum values of the i t h feature variable, respectively. When 1 i n , n denotes the number of feature variables. The original value x t i of the i t h feature variable at time t is mapped to the [0, 1] interval to obtain the normalized value x t i , as shown in Equation (20), where 1 t m , and m is the total number of samples at all time points. The probability density distribution of the normalized datasets for the two ships are shown in Figure 4 and Figure 5.
x t i = x t i x i M i n x i M a x x i M i n
Consequently, the input matrix X m × n and output matrix Y m × 1 of the ship fuel consumption forecast model are given by Equation (21).
X = [ X 1 X 2 X t X m ] = [ x 11 x 21 x m 1 x 12 x 22 x m 2 x t i x 1 n x 2 n x m n ] , = [ y 1 + k y 2 + k y t + k y m + k ]
where X t represents the total number n of features required to calculate ship fuel consumption at time t , x t i represents the normalized value of the i t h feature variable at time t , and y t + k denotes the real value of fuel consumption at time t + k . Note that t = 1 ,   2 ,   ,   m , i = 1 ,   2 ,   ,   n , and k = 1 ,   2 ,   ,   10 . Both case ships’ data were allocated among training (70%), validation (10%) and test (20%) sets by using a randomly sampling approach. For Ro-Ro ship, the training set, validation set and test set contain about 13,440, 1920 and 3840 data points respectively on R M , and 3990, 570 and 1140 data points on R B ; For LPG carrier, training set, validation set and test set contain about 6930, 990 and 1980 data points respectively under ballast and 5180, 740 and 1480 data points under laden.

3.3. Broad Learning System

3.3.1. Basic Concepts of BLS

Whereas all aforementioned models are constructed in the depth direction, BLS is employed to construct its network in the breadth direction. That is, the entire network is scaled horizontally, which effectively reduces network complexity [11]. The BLS model consists of four components, as shown in Figure 6: an input layer, a feature mapping layer, an enhancement node layer, and an output layer.
The specific algorithm flow corresponding to each BLS layer is as follows:
In the input layer, the input matrix X is entered into the model, and the BLS network performs feature mapping. Through the feature mapping function ϕ g ( * ) , X   is mapped to the mapping nodes of t groups of different features, and each group of mappings is set to generate k nodes; the calculation formula of the g t h group feature map node Z g is given by
Z g = ϕ g ( X W e g + β e g )
The parallel representation of Z g yields Z j = ( Z 1 , Z 2 , , Z j ) , where Z j m × j k . The mapped features are regarded as a single layer; that is, the feature-mapping layer. Here ϕ g ( * ) is a linear mapping function, W e g n × k is a randomly generated weight matrix, and β e g is the corresponding bias term.
The t group feature map node Z g is connected to the enhancement node layer. The weight coefficient matrix W h s and bias term β h s are generated randomly, and Z j is mapped to the p group of randomly generated weight enhancement nodes through the nonlinear function ξ s ( * ) , expressed as H p = ( H 1 , H 2 , , H p ) . Here, H p m × p q constitutes the enhanced node layer. Let each group of mappings generate q nodes and H s represents the s t h group of enhanced nodes, as expressed in Equation (23).
H s = ξ s ( Z j W h s + β h s )
The parallel connection between Z j and H p , expressed as A , is directly connected to the output terminal, also known as the output layer. Therefore, the representation of the output matrix is given by Equation (24):
Y = ( Z j | H p ) W = A W
Here, W is the connection weight matrix of the output layer, calculated according to Equation (25).
W = A 1 Y
where A 1 , representing the pseudoinverse matrix of matrix A , is calculated using the approximate solution of the ridge regression of A 1 . Subsequently, the connection weight matrix W is obtained, and model training is completed. This is expressed in Equation (26), where I represents the identity matrix and A T is the transpose matrix of A [11].
A 1 = lim λ 0 ( λ I + A T A ) 1 A T

3.3.2. Regression Analysis using BLS

As the most popular machine learning model among multiple fields, BLS has recently attracted significant attention from academia and industry. Owing to its efficiency and effectiveness, BLS has numerous applications in classification and time-series forecasting [20]. The application of BLS in classification achieved good results, such as hyperspectral imagery classification [21], zero-shot image classification [22] and event-based object classification [23]. As for time-series forecasting, Han et al. developed a maximum information exploitation BLS to solve the issue of extreme information utilization for large-scale chaotic time-series. The chaotic systems linear information is captured through an improved leaky integrator dynamical reservoir. In addition, the structural characteristics of BLS can be combined to effectively utilize nonlinear information. Finally, comparative experiments verified that MIE-BLS achieve better results in large-scale dynamical system modeling [24]. Liu et al. applied BLS in the study of a fast architecture for real-time training of the traffic predictor. Compared with LASSO regression, RNN and other machine learning models, BLS was concluded to significantly improve the forecast speed [25]. Yang et al. proposed a novel time-series model based on BLS to forecast port throughput, which exhibited superior performance to that of classical-time series models [26].
In addition to the aforementioned applications, BLS also can also be utilized to address problems related to time series across different domains, as in [27,28,29,30]. Owing to the impressive generalizability and remarkable efficiency, the present study extended the applicability of emerging BLS models to the field of fuel consumption forecasting.

4. Numerical Experiments

4.1. Measuring Critera

The MAE, one of the most common evaluation metrics for model performance, represents the mean of the absolute error between the forecasted and real fuel consumption values, as expressed in Equation (27).
M A E = 1 m i = 1 m | y i y i |
where m denotes the volume of test data. To efficiently forecast fuel consumption in practice, this study also considers the time consumed by the model. Experimental results obtained in this study are uniformly compared and analyzed in Section 4.3.

4.2. Comparison with Baseline Methods

As described in Section 3.1, the fuel consumption variable data were allocated among training and testing sets. Each forecast model was trained with the former, and subsequently verified with the latter. To objectively compare forecast performance between different models, default hyperparameters in the Python programming language were employed for simulation training. The hyperparameters settings of the ARIMAX, SVR, RNN, LSTM, ELM and BLS models are presented in Table 5. In addition, the time step of the SVR, RNN, LSTM, ELM, and BLS models were set to 5.
The real and forecast fuel consumption values were compared for each model. Figure 7 and Figure 8 represent comparative test results for the passenger ro-ro ship on R M and R B , respectively. Figure 9 and Figure 10 correspond to comparative test results for the LPG carrier under the ballast and laden states, respectively. The red curve in each figure represents real fuel consumption values, and the other curves measure the fuel consumption values obtained by different forecast models.
As shown in Figure 7, the real fuel consumption value was distributed in stages. The data difference between each stage was large, and the data difference within each stage was small. Before the 700th data point, the prediction deviation of each model was relatively large. However, from the overall point of view, the prediction deviation of each model was small. The forecast result of BLS model was the closest to reality, and the MAE of the BLS on R M reached 0.0345, while the deviations of the ARIMAX and SVR models were relatively large.
As shown in Figure 8, the real fuel consumption had great overall fluctuation. In particular, after the 450th data point, the data fluctuated greatly and the forecast deviation was large. The BLS model had the best forecast result, and the MAE of the BLS on R B reached 0.0397. The results of the ARIMAX and SVR models were relatively poor.
As shown in Figure 9, the data as a whole were less volatile. On the whole, the forecast of the region showed a stable trend after the 1000th data point and was relatively accurate, while the forecast of the region before the 1000th data point showed large fluctuation and error was relatively large. The forecast result of BLS was closer to the real value than other models, and the MAE of BLS reached 0.0272 under ballast. In addition, ARIMAX failed to perform well in the forecast.
It can be seen from Figure 10 that the real value of fuel consumption was less volatile overall. The forecast error of the regions with large fluctuations showed a trend in each model that was relatively large before the 400th data point and after the 1200th data point, while the forecast was relatively close to the real value between the 400th and 1200th data points. The forecasted result of the BLS model was closest to the real value, and the MAE of the BLS under laden reached 0.0759, while the overall deviations of the SVR and ARIMAX models were large.

4.3. Optimization of Parameters and Discussion

4.3.1. Optimization of Parameters

According to the observed data, each hyperparameter has a relatively suitable value. Furthermore, the hyperparameters will have an impact on the forecast accuracy of the model. Hence, a series of hyperparameters need to be tuned. However, iterated experiments for all possible hyperparameter combinations would incur a long computation time. Therefore, this study only considered several important hyperparameters. The optimal setting for each hyperparameter was determined according to the MAE on the test set. All parameters were adjusted according to the regression principles of ARIMAX, SVR, RNN, LSTM, ELM and BLS, as well as prior studies. The results of each model were the average results on 100 runs to avoid overtraining and validate generalization.
For the ARIMAX model, d was set to the default value following the stability test. The influence of p and q on forecast accuracy is discussed later in the section [12]. Results of four groups of cases corresponding to the two case ships are shown in Figure 11. It can be seen from observation that the optimal parameters ( p ,   q ) corresponding to Figure 11a–d are (1, 3), (2, 5), (3, 4), and (2, 1).
For the SVR model, four commonly used kernel functions were compared in this section: linear, RBF, sigmoid and poly [8]. The experimental results shown in Figure 12a that the sigmoid and RBF kernels produced the best performance for R M and R B , respectively. In the case of the LPG carrier, the poly and linear kernel functions yielded the best respective performance under state of ballast and laden states. The optimal parameters obtained for each dataset were used to adjust the time step from 1 to 10. As shown in Figure 12b, the optimal time step parameters of R M , R B , ballast and laden were 1, 4, 4 and 5, respectively.
For the RNN and LSTM models, we primarily considered the number of hidden layers and the time step [10]. The number of neurons in the first and second hidden layers was set to 100 and 80, respectively. Starting from the third hidden layer, the number of neurons in each layer was set to half of the previous layer. In all datasets, the optimal parameters determined in the final parameter adjustment process were selected, and the time-step parameters were set between 1–10 for optimization. Early stopping method was employed to avoid overfitting by terminating training when the training loss did not fall within 50 epochs.
Figure 13 and Figure 14 show the forecasted MAE results using RNN and LSTM with different numbers of hidden layers for the two case ships, respectively. For R M and R B of the passenger ro-ro ship, optimal performances was obtained under hidden layer parameters of (100,80). Furthermore, RNN yielded its best performance at 8 time step, whereas LSTM produced its best performance at 4 time step. For the ballast and laden states of the LPG carrier, the optimal hidden layer parameters corresponding to the RNN were (100,80), whereas those corresponding to LSTM were (100,80,40). Likewise, the best performance time step parameters of the RNN and LSTM models were 7 and 3, respectively.
For the ELM model, the parameter that had the greatest influence on model performance was the number of hidden layer nodes, which was adjusted in a range between 10 and 100 for optimization [17]. As shown in Figure 15a, optimal performance was achieved for the passenger ro-ro ship with 80 and 90 hidden layer nodes on R M and R B , respectively. For the LPG carrier, performance under ballast and laden states was optimized with 70 and 40 hidden layer nodes, respectively. The best parameters for each dataset were subsequently used to adjust the time step from 1 to 10. As shown in Figure 15b, the optimal time step parameters of R M , R B , ballast and laden were 2, 2, 5 and 8, respectively.
In the BLS forecast model, the most significant parameters are the number of feature nodes N1, number of feature windows N2, and number of enhanced nodes N3. Therefore, a grid search method was applied to traverse these parameters. Specifically, N1 was first optimized with N2, N3 to 10. To optimize N2, N1 was subsequently set to the optimal parameter while N3 remained at 10. Finally, both N1 and N2 was subsequently set to the optimal parameter while N3 remained at 10. Finally, both N1 and N2 were set to their respective optimal values to optimize N3. The results are shown in Figure 16.
Figure 16a shows that the best performance parameters of the BLS model on R M and R B were (N1 = 13, N2 = 19, N3 = 16) and (N1 = 3, N2 = 7, N3 = 3), respectively. Likewise, Figure 16b indicates that the best performance parameters of the BLS model under the ballast and laden states were (N1 = 6, N2 = 5, N3 = 2) and (N1 = 19, N2 = 9, N3 = 19), respectively. The optimal parameters for each dataset were used to adjust the time step from 1 to 10. As shown in Figure 16c, the optimal time step parameters of R M , R B , ballast, and laden were 2, 4, 2 and 6, respectively.

4.3.2. Discussion of Passenger Ro-Ro Ship

The figures in the previous section reflect the MAE changes of each fuel consumption forecast model on the four datasets during the parameter optimization. It is still necessary to compare the test sets accuracy obtained by the models using these optimized parameters. Two evaluation indicators—MAE and time—were considered in the comparative analysis.
For passenger ro-ro ships, the comparative experimental results obtained by each model with optimized hyperparameters are listed in Table 6. It is apparent that the BLS model obtained the highest accuracy for both R M and R B , with the MAE reaching 0.0140 and 0.0115, respectively. These figures represent a decrease by 59.42% and 71.03%, respectively, compared with the results obtained under default parameters. Furthermore, the MAE of BLS on R M was 59.18%, 55.13%, 7.89%, 6.04%, and 19.08% lower than those of ARIMAX, SVR, RNN, LSTM and ELM, respectively. For R B , the MAE likewise decreased by 90.12%, 56.60%, 33.91%, 22.29%, and 22.82%, respectively. It can be concluded that BLS yields a significant improvement over the baseline models in forecast performance for the passenger ro-ro ship. In terms of time consumption, the result obtained by BLS on R M is better than that of ELM, and approximately half that of SVR. Meanwhile, BLS has absolute advantage over the ARIMAX, RNN and LSTM models, with approximately 100-fold decreases in time consumption. On R B , BLS is faster than SVR, approximately four times as fast as ELM, and hundreds of times faster than ARIMAX, RNN and LSTM. Meanwhile, BLS has absolute advantage over the ARIMAX, RNN, and LSTM models, with approximately 100-fold decreases in time consumption. On R B , BLS is faster than SVR, approximately four times as fast as ELM, and hundreds of times faster than ARIMAX, RNN, and LSTM.

4.3.3. Discussion of the Liquefied Petroleum Gas Carrier

The following section discusses the impact of sea conditions on the fuel consumption forecast for the LPG carrier. Data corresponding to the LPG carrier were allocated between a dataset considering sea condition ( C 1 ), and a dataset without considering sea condition ( C 2 ) under the two loading situations. C 1 comprises the feature variables listed in Table 4, whereas C 2 omitted the SWH, WD, WP, SD and SCS variables. Under the ballast and laden conditions, C 1 and C 2 were forecasted using six models: ARIMAX, SVR, RNN, LSTM, ELM, and BLS. The experimental results are listed in Table 7.
For all datasets in Table 7, it is apparent that the BLS model achieved the highest forecast accuracy, with the MAE of C 1 under the ballast and laden conditions decreasing by 60.29% and 81.29%, respectively, compared with that under the default parameters. Moreover, the MAE of BLS for C 1 and C 2 under ballast and laden can reach 0.0108, 0.0150, 0.0142 and 0.0211, respectively. The MAE of C 1 under the ballast state decreased by 28.00% compared with that of C 2 , while the MAE of C 1 under the laden state was 32.70% lower than that of C 2 . In terms of time consumption, BLS performed better than SVR under the ballast state, and is four times faster than ELM. Furthermore, BLS performs better than ELM under laden state, and is five times and three times faster than SVR of C 1 and C 2 , respectively. In addition, BLS performs significantly better than ARIMAX, RNN and LSTM, with 10-fold to 100-fold improvements in speed for all datasets.
In summary, the BLS model obtained the optimal practical fuel consumption forecasting results for both ships. Furthermore, the forecast performance of all six models was better under C 1 than under C 2 . Therefore, the sea condition factors must be considered to achieve optimal forecast performance.

5. Conclusions

The key to maximizing the energy efficiency and minimizing the consumption costs associated with shipping is the accurate forecasting of fuel consumption. Currently, the accuracy and time performance of existing ship fuel consumption forecast models require improvement. In view of this problem, this paper presents a study comparing BLS with the ARIMAX, SVR, RNN LSTM and ELM models with actual navigation data obtained from two types of ships.
For the LPG carrier, data were further allocated among the C 1 and C 2 datasets to examine the significance of sea conditions on fuel consumption. Finally, conclusions were drawn based on an analysis of experimental results. Specifically, the BLS model achieved optimal forecasting performance for both ships. These results correlate with the unique characteristics of BLS, which includes its simple structure and short training time, making it suitable for applications in real-time navigation scenarios. Furthermore, results on the LPG carrier data indicate that sea conditions must be accounted for to maximize forecast performance.
The optimal time-series forecasting model based on a novel regression analysis using BLS in this study can be employed for the real-time monitoring of fuel consumption, thereby enabling energy- and environmentally friendly shipping conditions. This is of great significance in the problems of route, trim and ship speed optimization. Furthermore, additional sailing data corresponding to different ships will be collected in the future to verify the model’s generalizability.

Author Contributions

Writing—review and editing, methodology, and formal analysis, Y.Z.; data curation, and writing—original draft preparation, X.L.; validation, J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (grant nos. 52131101, 51939001, and U1813203), the Liao Ning Revitalization Talents Program (grant nos. XLYC1807046, and XLYC1908018), the Natural Science Foundation of Liaoning Province (grant no. 2020-HYLH-26), and the Science and Technology Fund for Distinguished Young Scholars of Dalian (grant no. 2021RJ08).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The fuel consumption related data of the passenger ro-ro ship can be found at http://cogsys.imm.dtu.dk/propulsionmodelling/data.html (accessed on 8 November 2022); the fuel consumption related data of the LPG carrier can be found at http://langnet.uniri.hr/resources.html (accessed on 8 November 2022).

Conflicts of Interest

The authors declare that there are no conflict of interest regarding the publication of this paper.

Appendix A

Table A1. Summary of this paper and related literature review.
Table A1. Summary of this paper and related literature review.
LiteratureCase ShipData SourceForecast ModelMeasuring Criterion
[4]Container shipsShip operational data LR R 2
[6]Cruise shipNavigation data, engine data MLR R 2
[7]Container shipThe dataset derived from the fleet management system LASSOMAE, RMSD, CS
[8]Passenger ro-ro shipShip operational dataLR, SVR, ANNRMSE
[9]More than 30 types of shipsCollected by Jingtang PortGBR, RF, ANN, LR, KNNEVS, MAE, MSE, R 2
[10]Cargo shipThe monitoring data of ship sailingANN, RNN, LSTMRMSE, MAE, R 2
This paperPassenger ro-ro ship, LPG carrierOnboard data, available oceanographic and meteorological dataARIMAX, SVR, RNN, LSTM, ELM, BLSMAE, Time
Table A2. Summary of the academic contributions of this paper.
Table A2. Summary of the academic contributions of this paper.
HighlightsAcademic Contributions
1The application of emerging BLS model in the field of time series forecast for ship fuel consumption is expanded.
2Using two different types of case ships as research objects to verify the generalization of the results.
3According to the sea state and weather conditions, the data of the two case ships are divided into two groups for separate discussion, so as to restore the actual navigation environment to the maximum extent and improve the reliability of the forecast model.
4The model which consume less time and more accurately is obtained by systematic verification analysis on the applicability and accuracy of different time series analysis and machine learning techniques for ship fuel consumption forecast.
Table A3. The description about the feature variables for the passenger ro-ro ship.
Table A3. The description about the feature variables for the passenger ro-ro ship.
Feature VariablesDescription
SOGSOG is the speed of the ship relative to the ground or any other fixed object, such as island or buoy.
TrimThe trim indicates its floating position in the direction of length, i.e., whether the bow or stern is deeper submerged into the water.
PitchThe distance traveled by the propeller in one rotation.
DraughtThe draught describes the vertical distance between the waterline and the bottom of the hull.
HeadwindThe headwind means the direction of travel is the opposite direction of the wind.
CrosswindThe crosswind means the direction of travel is perpendicular to the wind direction.
RudderThe rudder is a piece of blade installed aft in the submerged part of the ship to provide directional control for ship.
Table A4. The description about the feature variables for R B .
Table A4. The description about the feature variables for R B .
Feature VariablesDescription
RPMRPM is the number of complete revolutions or rotations of a propeller in one minute.
COGCOG refers to the actual direction of progress of the ship, between two points, relative to the surface of the earth.
TLTL represents the need for liquefaction of cargo on the case of LPG.
ListList refers to a floating state in which the ship tilts to starboard or port with unequal draft on the starboard or port sides.
UwindUwind is the component of wind speed in the horizontal direction.
VwindVwind is the vertical component of wind speed
WGWG means a brief and sudden increase in speed of the wind.
SWHSWH is the actual wave height measured according to certain rules.
WDThe line along which a wave travels with respect to its compass heading.
WPWP is the time for a particle on a medium to make one complete vibrational cycle.
SDSD indicates the direction they are headed for or where the current is flowing towards.
Table A5. The list of acronyms used in the paper.
Table A5. The list of acronyms used in the paper.
ExplanationAbbreviationExplanationAbbreviation
Broad Learning SystemBLSAutoregressive Integrated Moving Average Model with Exogenous InputsARIMAX
Support Vector RegressionSVRRecurrent Neural NetworkRNN
Long Short-term Memory NetworkLSTMExtreme Learning MachinesELM
Roll-on Roll-offRo-RoLiquefied Petroleum GasLPG
Least Absolute Shrinkage and Selection OperatorLASSOLinear RegressionLR
Artificial Neural NetworksANNAutoregressive AverageAR
Moving AverageMAAutoregressive Integrated Moving AverageARIMA
Support Vector MachinesSVMPolynomialPoly
Radial Basis FunctionRBFSingle-hidden Layer Feedforward Neural NetworksSLFNs
Ship Automation SystemSASElectronic Chart Display and Information SystemECDIS
Mean Absolute ErrorMAEMultiple linear regressionMLR
Gradient Boosting RegressionGBRRandom Forest RegressionRF
K-Nearest Neighbor RegressionKNNExplained Variance ScoreEVS
Mean Squared ErrorMSECumulative ScoreCS
Root Mean Square DeviationRMSD

References

  1. Wang, S.; Psaraftis, H.N.; Qi, J. Paradox of international maritime organization’s carbon intensity indicator. Commun. Transp. Res. 2021, 1, 100005. [Google Scholar] [CrossRef]
  2. Wan, Z.; El Makhloufi, A.; Chen, Y.; Tang, J. Decarbonizing the international shipping industry: Solutions and policy recommendations. Mar. Pollut. Bull. 2018, 126, 428–435. [Google Scholar] [CrossRef] [Green Version]
  3. Yan, R.; Wang, S.; Psaraftis, H. Data analytics for fuel consumption management in maritime transportation: Status and perspectives. Transp. Res. Pt. e-Logist. Transp. Rev. 2021, 155, 102489. [Google Scholar] [CrossRef]
  4. Wang, S.; Meng, Q. Sailing speed optimization for container ships in a liner shipping network. Transp. Res. Pt. e-Logist. Transp. Rev. 2012, 48, 701–714. [Google Scholar] [CrossRef]
  5. Gao, C.; Hu, Z. Speed optimization for container ship fleet deployment considering fuel consumption. Sustainability 2021, 13, 5242. [Google Scholar] [CrossRef]
  6. Erto, P.; Lepore, A.; Palumbo, B.; Vitiello, L. A procedure for predicting and controlling the ship fuel consumption: Its implementation and test. Qual. Reliab. Eng. Int. 2015, 31, 1177–1184. [Google Scholar] [CrossRef]
  7. Wang, S.; Ji, B.; Zhao, J.; Liu, W.; Xu, T. Predicting ship fuel consumption based on LASSO regression. Transport. Res. Part D-Transport. Environ. 2018, 65, 817–824. [Google Scholar] [CrossRef]
  8. Zhu, Y.; Zuo, Y.; Li, T. Modeling of ship fuel consumption based on multisource and heterogeneous data: Case study of passenger ship. J. Mar. Sci. Eng. 2021, 9, 273. [Google Scholar] [CrossRef]
  9. Peng, Y.; Liu, H.; Li, X.; Huang, J.; Wang, W. Machine learning method for energy consumption prediction of ships in port considering green ports. J. Clean Prod. 2020, 264, 121564. [Google Scholar] [CrossRef]
  10. Yuan, Z.; Liu, J.; Zhang, Q.; Liu, Y.; Yuan, Y.; Li, Z. Prediction and optimisation of fuel consumption for inland ships considering real-time status and environmental factors. Ocean Eng. 2021, 221, 108530. [Google Scholar] [CrossRef]
  11. Chen, C.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 10–24. [Google Scholar] [CrossRef] [PubMed]
  12. Alam, W.; Ray, M.; Kumar, R.R.; Sinha, K.; Rathod, S.; Singh, K.N. Improved ARIMAX modal based on ANN and SVM approaches for forecasting rice yield using weather variables. Indian J. Agric. Sci. 2018, 88, 1909–1913. [Google Scholar]
  13. Huang, G.; Zhu, Q.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  14. Mohammad, V.; Mohammad, E.; Seyyed, M. Comparison of the ARMA, ARIMA, and the autoregressive artificial neural network models in forecasting the monthly inflow of Dez dam reservoir. J. Hydrol. 2013, 476, 433–441. [Google Scholar]
  15. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef] [Green Version]
  16. Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Physica D 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  17. Huang, G.; Huang, G.B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef] [PubMed]
  18. Petersen, J.P. Mining of Ship Operation Data for Energy Conservation. Ph.D Thesis, Technical University of Denmark, Lyngby, Denmark, 2011. [Google Scholar]
  19. Vorkapić, A.; Radonja, R.; Martinčić-Ipšić, S. Predicting seagoing ship energy efficiency from the operational data. Sensors 2021, 21, 2832. [Google Scholar] [CrossRef]
  20. Gong, X.; Zhang, T.; Chen, C.L.P.; Liu, Z. Research review for broad learning system: Algorithms, theory, and applications. IEEE Trans. Cybern. 2021, 52, 8922–8950. [Google Scholar] [CrossRef]
  21. Kong, Y.; Wang, X.; Cheng, Y.; Chen, C.P. Hyperspectral imagery classification based on semi-supervised broad learning system. Remote Sens. 2018, 10, 685. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, J.; Wang, X.; Cheng, Y. Broad attribute prediction model with enhanced attribute and feature. IEEE Access. 2019, 7, 124606–124620. [Google Scholar] [CrossRef]
  23. Gao, S.; Guo, G.; Huang, H.; Cheng, X.; Chen, C.P. An end-to-end broad learning system for event-based object classification. IEEE Access. 2020, 8, 45974–45984. [Google Scholar] [CrossRef]
  24. Han, M.; Li, W.; Feng, S.; Qiu, T.; Chen, C.L.P. Maximum information exploitation using broad learning system for large-scale chaotic time-series prediction. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2320–2329. [Google Scholar] [CrossRef]
  25. Liu, D.; Baldi, S.; Yu, W.; Cao, J.; Huang, W. On training traffic predictors via broad learning structures: A benchmark study. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 749–758. [Google Scholar] [CrossRef]
  26. Yang, J.; Li, T.; Zuo, Y.; Tian, Y.; Cao, Y.; Yang, H.; Chen, C.L.P. Forecast application of time series model based on BLS in port Cargo throughput. In Proceedings of the 2018 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), Jinan, China, 14–17 December 2018. [Google Scholar]
  27. Feng, S.; Ren, W.; Han, M.; Chen, Y.W. Robust manifold broad learning system for large-scale noisy chaotic time series prediction: A perturbation perspective. Neural Netw. 2019, 117, 179–190. [Google Scholar] [CrossRef] [PubMed]
  28. Guo, J.; Wang, L.; Fan, K.; Yang, B. An efficient model for predicting setting time of cement based on broad learning system. Appl. Soft. Comput. 2020, 96, 106698. [Google Scholar] [CrossRef]
  29. Kuok, S.C.; Yuen, K.V. Broad learning for nonparametric spatial modeling with application to seismic attenuation. Appl. Soft. Comput. 2020, 35, 203–218. [Google Scholar] [CrossRef]
  30. Li, X.; Zhu, Y.; Zuo, Y.; Li, T.; Chen, C.L.P. Prediction of ship fuel consumption based on broad learning system. In Proceedings of the 2019 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), Guangzhou, China, 20–23 December 2019. [Google Scholar]
Figure 1. Modeling of fuel consumption based on SVR.
Figure 1. Modeling of fuel consumption based on SVR.
Sustainability 15 00380 g001
Figure 2. Modeling of fuel consumption based on RNN.
Figure 2. Modeling of fuel consumption based on RNN.
Sustainability 15 00380 g002
Figure 3. Modeling of fuel consumption based on LSTM.
Figure 3. Modeling of fuel consumption based on LSTM.
Sustainability 15 00380 g003
Figure 4. Distribution of normalized datasets for passenger ro-ro ship.
Figure 4. Distribution of normalized datasets for passenger ro-ro ship.
Sustainability 15 00380 g004
Figure 5. Distribution of normalized datasets for LPG carrier.
Figure 5. Distribution of normalized datasets for LPG carrier.
Sustainability 15 00380 g005
Figure 6. Modeling of fuel consumption based on BLS.
Figure 6. Modeling of fuel consumption based on BLS.
Sustainability 15 00380 g006
Figure 7. Comparison between real value and forecasted value ( R M ).
Figure 7. Comparison between real value and forecasted value ( R M ).
Sustainability 15 00380 g007
Figure 8. Comparison between real value and forecasted value ( R B ).
Figure 8. Comparison between real value and forecasted value ( R B ).
Sustainability 15 00380 g008
Figure 9. Comparison between real value and forecasted value (ballast).
Figure 9. Comparison between real value and forecasted value (ballast).
Sustainability 15 00380 g009
Figure 10. Comparison between real value and forecasted value (laden).
Figure 10. Comparison between real value and forecasted value (laden).
Sustainability 15 00380 g010
Figure 11. Results of ARIMAX forecast model parameter tuning: (a) R M , (b) R B , (c) ballast, (d) laden.
Figure 11. Results of ARIMAX forecast model parameter tuning: (a) R M , (b) R B , (c) ballast, (d) laden.
Sustainability 15 00380 g011
Figure 12. Results of SVR forecast model parameter tuning: (a) kernel function, (b) time step.
Figure 12. Results of SVR forecast model parameter tuning: (a) kernel function, (b) time step.
Sustainability 15 00380 g012
Figure 13. Results of RNN forecast model parameter tuning: (a) number of hidden layers, (b) time step.
Figure 13. Results of RNN forecast model parameter tuning: (a) number of hidden layers, (b) time step.
Sustainability 15 00380 g013
Figure 14. Results of LSTM forecast model parameter tuning: (a) number of hidden layers, (b) time step.
Figure 14. Results of LSTM forecast model parameter tuning: (a) number of hidden layers, (b) time step.
Sustainability 15 00380 g014
Figure 15. Results of ELM forecast model parameter tuning: (a) number of hidden layer nodes, (b) time step.
Figure 15. Results of ELM forecast model parameter tuning: (a) number of hidden layer nodes, (b) time step.
Sustainability 15 00380 g015
Figure 16. Results of BLS forecast model parameter tuning: (a) passenger ro-ro ship, (b) LPG carrier, (c) time step.
Figure 16. Results of BLS forecast model parameter tuning: (a) passenger ro-ro ship, (b) LPG carrier, (c) time step.
Sustainability 15 00380 g016
Table 1. Principal particulars of passenger ro-ro ship.
Table 1. Principal particulars of passenger ro-ro ship.
ParametersValue
Length Overall135 m
Breadth Molded22.7 m
Depth to Main Deck8.1 m
Designed Speed21 knots
Main Engine Power3360 kW
Table 2. Feature variables of passenger ro-ro ship.
Table 2. Feature variables of passenger ro-ro ship.
NoFeature VariablesAbbreviation
1ME Total Fuel Consumption METF
2Speed Over GroundSOG
3TrimTrim
4Port PitchPPitch
5Starboard PitchSPitch
6Port DraughtPDraught
7Starboard DraughtSDraught
8HeadwindHeadwind
9CrosswindCrosswind
10Port RudderPRudder
11Starboard RudderSRudder
Table 3. Principal particulars of LPG carrier.
Table 3. Principal particulars of LPG carrier.
ParametersValue
Capacity54340 DWT
Length Overall225 m
Breadth Molded37 m
Designed Speed16.8 knots
Main Engine Power12,400 kW
Propeller Diameter7400 mm
Propeller Pitch5971.06 mm
Table 4. Feature variables of the LPG carrier.
Table 4. Feature variables of the LPG carrier.
NoFeature VariablesAbbreviation
1ME Total Fuel Consumption METF
2Revolution Per MinuteRPM
3Speed Over GroundSOG
4Course Over GroundCOG
5Total LoadTL
6Ambient Air TemperatureAT
7Sea Water TemperatureSWT
8TrimTrim
9ListList
10UwindUwind
11VwindVwind
12Wind GustWG
13Significant Wave HeightSWH
14Wave DirectionWD
15Wave PeriodWP
16Sea DirectionSD
17Sea Current SpeedSCS
Table 5. The model hyperparameter settings.
Table 5. The model hyperparameter settings.
Forecast ModelHyperparameter Setting
ARIMAXp = 1
q = 1
d = ’None’
SVRC = 1.0
Kernel = ‘Linear’
RNNNumber of Hidden Layers = 1
Number of Neurons = 100
Activation Function = Tanh
Batch Size = 1
Dropout = 0.2
LSTMNumber of Hidden Layers = 1
Number of Neurons = 100
Activation Function = Tanh
Batch Size = 1
Dropout = 0.2
ELMNumber of Hidden Layer Nodes = 50
BLSN1 = 10
N2 = 10
N3 = 10
Table 6. Comparison of various fuel consumption forecast models (passenger ro-ro ship).
Table 6. Comparison of various fuel consumption forecast models (passenger ro-ro ship).
DatasetForecast ModelMAETime(s)
R M ARIMAX0.034370.0374
SVR0.03122.9718
RNN0.0152100.3603
LSTM0.0149174.8111
ELM0.01731.3166
BLS0.0141.2559
R B ARIMAX0.116541.6143
SVR0.02650.1039
RNN0.017441.9314
LSTM0.014858.6204
ELM0.01490.4194
BLS0.01150.0967
Table 7. Comparison of various fuel consumption forecast models (LPG carrier).
Table 7. Comparison of various fuel consumption forecast models (LPG carrier).
DatasetForecast ModelMAETime(s)
Ballast- C 1 ARIMAX0.129528.1713
SVR0.04380.2886
RNN0.024865.7348
LSTM0.0172111.9017
ELM0.02130.4667
BLS0.01080.1716
Ballast- C 2 ARIMAX0.294525.0112
SVR0.07280.2398
RNN0.035665.737
LSTM0.0221112.4083
ELM0.02590.4633
BLS0.0150.1574
Laden- C 1 ARIMAX0.272215.0424
SVR0.04152.3295
RNN0.040167.2778
LSTM0.025588.4812
ELM0.02790.4832
BLS0.01420.4638
Laden- C 2 ARIMAX0.743413.0898
SVR0.04461.3082
RNN0.049754.3062
LSTM0.02789.382
ELM0.0410.449
BLS0.02110.434
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Zuo, Y.; Jiang, J. Application of Regression Analysis Using Broad Learning System for Time-Series Forecast of Ship Fuel Consumption. Sustainability 2023, 15, 380. https://doi.org/10.3390/su15010380

AMA Style

Li X, Zuo Y, Jiang J. Application of Regression Analysis Using Broad Learning System for Time-Series Forecast of Ship Fuel Consumption. Sustainability. 2023; 15(1):380. https://doi.org/10.3390/su15010380

Chicago/Turabian Style

Li, Xinyu, Yi Zuo, and Junhao Jiang. 2023. "Application of Regression Analysis Using Broad Learning System for Time-Series Forecast of Ship Fuel Consumption" Sustainability 15, no. 1: 380. https://doi.org/10.3390/su15010380

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop