Next Article in Journal
A Pedestrian Detection Network Based on an Attention Mechanism and Pose Information
Next Article in Special Issue
A Local Discrete Feature Histogram for Point Cloud Feature Representation
Previous Article in Journal
Quantitative Assessment of Rock Burst Risk in Roadway Tunneling Considering Variation of Coal Mass Parameters
Previous Article in Special Issue
Lagrange Relaxation for the Capacitated Multi-Item Lot-Sizing Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Ship-Unloading Time Using Neural Networks

by
Zhen Gao
1,2,*,
Danning Li
1,2,
Danni Wang
1,2,
Zengcai Yu
1,2,
Witold Pedrycz
3 and
Xinhai Wang
4
1
National Frontiers Science Center for Industrial Intelligence and Systems Optimization, Northeastern University, Shenyang 110819, China
2
Key Laboratory of Data Analytics and Optimization for Smart Industry (Northeastern University), Ministry of Education, Shenyang 110819, China
3
The Department of Electrical and Computer Engineering, University of Alberta, Edmonton, AB T6R 2V4, Canada
4
College of Sciences, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(18), 8213; https://doi.org/10.3390/app14188213
Submission received: 8 August 2024 / Revised: 6 September 2024 / Accepted: 11 September 2024 / Published: 12 September 2024

Abstract

:
The prediction of unloading times is crucial for reducing demurrage costs and ensuring the smooth scheduling of downstream processes in a steel plant. The duration of unloading a cargo ship is primarily determined by the unloading schedule established at the raw materials terminal and the storage operation schedule implemented in the stockyard. This study aims to provide an accurate forecast of unloading times for incoming ships at the raw materials terminal of a steel plant. We propose three neural network-based methods: the Backpropagation Neural Network (BP), the Random Vector Functional Link (RVFL), and the Stochastic Configurations Network (SCN) for this prediction. This issue has not been previously researched using similar methods, particularly in the context of large-scale steel plants. The performance of these three methods is evaluated based on several indices: the Root Mean Square Error (RMSE), the quality of the best solution, convergence, and stability, which are employed for predicting unloading times. The prediction accuracies achieved by the BP, RVFL, and SCN were 76%, 85%, and 87%, respectively. These results demonstrate the effectiveness and potential applications of the proposed methods.

1. Introduction

The prediction of unloading times is crucial for reducing delay costs and ensuring the smooth scheduling of downstream processes in a large steel plant [1,2]. In fact, accurately predicting unloading times can create favorable conditions for the more efficient allocation of unloading resources for subsequently arriving ships. Additionally, it can provide a buffer for subsequent storage operations in the storage yard. In a large-scale iron and steel enterprise, hundreds or even thousands of ships transport tens of millions of tons of raw materials to the terminal each year. These ships must be unloaded promptly to minimize delay costs and meet production demands for raw materials. However, significant demurrage costs—amounting to hundreds of millions annually—arise when many ships fail to unload on time due to various factors.
Two key factors influence the unloading time of ships: the ship-unloading schedule established at the raw materials terminal and the raw materials storage operation schedule implemented in the stockyard. The ship-unloading system of a steel plant is illustrated in Figure 1. Typically, this system consists of several berths, a limited number of ship unloaders, and a conveyor transmission system. The berths are designated for docking raw material ships, while the ship unloaders are responsible for unloading the raw materials. These unloaders can move along a fixed track. For smaller vessels, typically those under 50,000 tons, a single ship unloader is enough to meet unloading requirements. However, for larger vessels, generally exceeding 150,000 tons, it is necessary to allocate two to three ship unloaders to facilitate the unloading operations. The conveyor system is installed throughout the entire port area, the stockyard, and certain production areas to enable the bidirectional transfer of raw materials and products among these locations. When a ship arrives at the raw materials terminal, a berth and one or more ship unloaders are assigned to it for the unloading operation. Simultaneously, storage space must be designated for the unloaded raw materials in the stockyard. Accurate prediction of ship-unloading times can create buffer space for subsequent dock-unloading schedules and storage space allocation. This improvement facilitates the execution of related operations, reduces ship delay costs, and enhances economic efficiency, making it a critical factor in port operations.
The prediction methods have become a prominent area of study in both industrial applications [3,4] and academic research [5,6]. These methods can be broadly categorized into three groups: statistical methods [7,8,9], mechanistic methods [10,11], and data-driven methods [12,13]. Statistical methods operate under the assumption that the observations of the predicted object follow specific statistical distributions, utilizing statistical theory to derive prediction results. In contrast, data-driven methods rely entirely on data, employing machine learning techniques to generate predictions. Mechanistic methods assume that the predicted object adheres to certain physical or chemical laws, including thermodynamics, dynamics, and fluid mechanics, and they derive predictions through the application of relevant mechanistic equations. However, data-driven methods have gained popularity due to their user-friendliness and lower implementation costs in industrial settings.
As far as we know, unloading time prediction is rarely studied for bulk carriers. However, there are several analogous studies in other industries. For instance, Zhang et al. [14] developed a fuzzy sequence-to-sequence network for forecasting unloading times in earthmoving projects. Liang and Wang [15] developed a modularized simulation method (MSM) for forecasting delivery times of priority lots in a 300 mm silicon wafer fabrication production line. Xu et al. [16] developed a robust berth-scheduling algorithm to address the uncertainty of vessel arrival delay and handling time. In response to this research gap, this study proposes a method that employs three neural network-based approaches: a Backpropagation Neural Network (BP), a Random Vector Functional Link Network (RVFL), and a Stochastic Configuration Network (SCN) model to predict unloading time at a steel plant raw materials terminal. There are several compelling reasons for selecting this prediction method:
Neural networks have been widely applied to prediction and classification and exhibited a great deal of success.
  • BP as a mature neural network and the ensuing learning method has been widely applied [17,18].
  • RVFL introduces an efficient way to find out the optimal network parameters and represents a richer structure variety of neural networks [19].
  • SCN exhibits a concise network structure and demonstrates a better application potential [20].
The primary contributions of this paper are:
  • Three neural network models-Backpropagation (BP), Random Vector Functional Link (RVFL), and Stochastic Configuration (SCN) have been developed for predictive applications.
  • Three prediction models are comprehensively evaluated based on their predictive accuracy, convergence, and stability, and are applied to a real-world scenario involving unloading-time prediction.
The paper is organized as follows. Section 2 covers the relevant studies. Section 3 includes a description of the relevant methods. Section 4 elaborates on the experimental results produced with the proposed methods. Section 5 includes the conclusions and identifies some future directions.

2. Literature Review

There are several studies related to predicting unloading time, which can be summarized from three perspectives: ship-unloading scheduling, stockyard storage operations, and the application of neural networks for prediction.
Ship-unloading scheduling: Dhingra et al. [21] proposed a two-level stochastic model to estimate a container ship’s handling time. The higher-level model employed a continuous-time Markov chain to estimate the handling time for loading and unloading, while the lower-level model utilized a closed queuing network to provide the transition matrix as input for the higher-level model. Bish [22] addressed a container loading and unloading scheduling problem involving a set of ships, considering the storage location of each container, the assignment of vehicles to containers, and the scheduling of cranes for loading and unloading operations. The objective was to minimize the maximum time required to serve the set of ships. A heuristic approach was proposed to solve the problem, and the effectiveness of this heuristic was subsequently analyzed. Al-Dhaheri et al. [23] introduced a stochastic mixed-integer programming model aimed at minimizing container ship handling time. A genetic algorithm was developed to construct the crane schedule while adhering to the operational rules governing crane usage. Sun et al. [24] investigated a quay crane scheduling problem aimed at minimizing the completion time of loading and unloading operations for container ships. They formulated a mathematical programming model and proposed a Benders decomposition method to solve the problem. Sammarra et al. [25] introduced a tabu search heuristic for quay crane scheduling, focusing on minimizing the completion time of loading and unloading operations for container ships. They considered specific operational constraints for the quay cranes, decomposing the problem into a routing problem and a scheduling problem, which were addressed using a tabu search heuristic and a local search technique, respectively. Tang et al. [26] proposed a joint scheduling approach for quay cranes and trucks servicing a container ship. They developed a particle swarm optimization (PSO) method to tackle the scheduling problem. Kao et al. [27] presented a knowledge-based approach for ship unloading, aimed at minimizing total demurrage costs. Their knowledge base was designed to manage the sequencing of waiting ships and the transfer of ships between two docks. Kao et al. [28] proposed a heuristic approach for scheduling ship discharging. Their method considered the sequencing of holds and the assignment of ship unloaders, belt conveyor units, and stackers. Kao and Lee [29] introduced an integrated system for ship unloading, which coordinates two operations—dock assignment and ship discharging—to enhance the efficiency of unloading operations. Kim and Moon [30] developed a mixed-integer programming (MIP) model to address a berth-planning problem. In their approach, each ship was expected to remain for a fixed duration, irrespective of the berthing locations. A simulated annealing algorithm was devised to solve this problem. Kim et al. [31] investigated a ship-unloading issue for a large-scale steel plant, aiming to minimize the total flow time of all ships at the seaport. They developed a heuristic to address this challenge. Gao et al. [1] examined the scheduling of ship unloading in a large-scale steel plant. They formulated a mathematical model and proposed a column-generation algorithm to solve the problem. Additionally, Gao et al. [2] introduced a differential evolution algorithm for ship-unloading scheduling, with special attention given to the assignment of belt conveyors.
Stockyard storage operations: Lee et al. [32] proposed a mixed-integer programming (MIP) model for yard truck scheduling and storage allocation. Their objective was to minimize the weighted total cost, which included penalties for total delays and the costs associated with total travel time. A constructive heuristic was developed for the MIP model, resulting in a solution gap of 10.27% when compared to CPLEX. The solution time for the heuristic was at the second level, while CPLEX required 30 to 40 h to find a solution. Tang et al. [33] investigated a stockyard storage space-allocation problem at a large iron ore terminal. They formulated a mixed-integer linear programming model aimed at minimizing the total travel distance for all incoming iron ores. The primary constraints included rules governing space allocation and the operations of stacker-reclaimers. A genetic algorithm-based heuristic was employed to solve the model. For smaller problems, the heuristic can achieve optimal solutions in seconds; however, for larger-scale problems, CPLEX is unable to find an optimal solution within 2 h, while the heuristic can provide near-optimal solutions in seconds. Li and Tang [34] addressed a storage space-allocation problem in an iron-steel stockyard by developing a nonlinear programming model. The objective was to minimize transportation and penalty costs, with constraints related to geometric factors (length, width, and height) and differences in material types for spatial allocation. This problem was solved using an improved tabu search algorithm. Kim et al. [35] examined a raw materials storage-allocation problem for a large-scale steelworks plant. They proposed a mixed-integer linear programming model, which was solved using CPLEX 9.02. The model’s constraints included maintaining a safety distance between two stockpiles and ensuring materials balance. Their method enhanced current production practices.
Neural networks in prediction: Various neural network variants have been employed for prediction and classification tasks. Zhang and Shin [36] proposed a probabilistic neural network for monitoring manufacturing processes. This network featured Gaussian-mixture distributed parameters, which enhanced computational efficiency. Li et al. [37] developed a long short-term memory (LSTM) neural network for time-series prediction, utilizing partial least squares (PLS) to simplify the network architecture. The streamlined LSTM model effectively balanced strong generalization capabilities with a compact structure. Xiao et al. [12] employed a Backpropagation Neural Network (BP) combined with rough set theory for power load forecasting, using rough sets for dimensionality reduction. This approach yielded improved prediction results. Adelia and Panakkat [13] introduced a probabilistic neural network for predicting earthquake magnitudes. Kosanoglu [38] proposed an ensemble model that integrates time-series clustering techniques with deep learning methods for wind speed forecasting. The model utilized a Dirichlet mixture model and dynamic time warping to cluster features from the time-series data. The results indicated that this feature-clustering approach is a promising framework for forecasting. Hussein et al. [39] developed a machine learning method that combines the random vector functional link (RVFL) network with a moth search algorithm to predict missing values of total algal counts during water quality monitoring. The moth search algorithm optimized the input features for the RVFL network, resulting in predicted algal values that closely matched real observations. El-Said et al. [40] proposed four machine learning algorithms: RVFL, support vector machine, social media optimization, and k-nearest neighbors, to assess the impact of air injection and transverse baffles on the thermohydraulic performance of shell and tube heat exchangers. The results demonstrated that the RVFL model effectively identified nonlinear relationships between operating conditions and process responses. Wang and Wang [41] introduced an SCN model for predicting component concentrations in sodium aluminate liquor. This mechanistic model was integrated to elucidate the relationships among conductivity, temperature, and component concentrations in Bayer alumina production, with experimental results indicating high prediction accuracy. Li et al. [42] presented an improved SCN model for predicting ammonia nitrogen concentrations in water quality monitoring. They introduced a new inequality during the network construction process and proposed a node-selection method. The experimental results confirmed the effectiveness of the enhanced SCN model.
In summary, while there is a substantial amount of research on predicting ship-unloading times, most studies focus on container ships rather than bulk carriers. Furthermore, existing research on bulk carriers primarily addresses unloading scheduling rather than unloading-time prediction, which does not adequately reflect the significance of this issue. The study aims to address this gap.

3. Related Methods

3.1. BP Neural Network

BP is a multilayer neural network whose effectiveness has been demonstrated in numerous studies [17]. Typically, BP consists of one input layer, one or more hidden layers, and one output layer, with full connectivity between adjacent layers. The BP learning algorithm operates iteratively. The network parameters, which include weights and biases, are propagated forward through the network and adjusted backward to minimize error.
Given a group of samples, { ( x i , y i ) ,   i = 1 , , N ,   x i R d , y i R l } , assuming there is one hidden layer in the BP (Backpropagation) network. containing J nodes, denoted as v 1 , v 2 , , v J . The input weights, biases for the hidden nodes, output weights of the hidden nodes, and thresholds for the output nodes are w i j , b j , w j k , θ k , i = 1 , , d ,   j = 1 , , J ,   k = 1 , , l , respectively. Then,
  • Hidden layer outputs
v j = f 1 ( i = 0 d w i j x i ) ,   j = 0 , 1 , , J ,
f1 serves as the activation function for hidden node j, x0 = −1, w0j = bj.
2.
Output layer results
y - k = f 2 ( j = 0 J w j k v j ) ,   k = 0 , 1 , , l ,
f2 serves as the activation function for output node k, v0 = −1, w0k = θk.
3.
Updating of weights
Δ w j k = η g k v j , j = 0 , , J ,   k = 1 , , l ,   g k   y - k ( 1 y - k ) ( y k y - k ) .
Δ w i j = η e j x i , i = 0 , , d ,   j = 1 , , J , e j v j ( 1 v j ) k = 1 l w j k g k ,   g k y - k ( 1 y - k ) ( y k y - k ) .
However, the Backpropagation (BP) algorithm is inherently gradient based, which means there is a possibility of converging to local minima. Nevertheless, it remains a robust method due to various strategies designed to mitigate the risk of local optimality. These strategies include the momentum method, adaptive learning rates, and the use of first-order approximations, such as the Newton and quasi-Newton methods, instead of relying solely on the gradient method. The BP algorithm is detailed in Algorithm 1.
Algorithm 1: BP algorithm
Input:    samples {xi, yi}, xiRd, yiRl, i = 1, …, N; J
Output:  wij, wjk, i = 1, …, d; j=1, …, J; k = 1,…,l
Initialization:
  For all i, j
    wij = generateRandomNumber (−1, 1); //generate random numbers in interval [−1, 1];
  For all j, k
    wjk = generateRandomNumber (−1, 1); //generate random numbers in interval [−1, 1];
  η = 0.8; //learning rate
  Lmax = 5000; L = 0;
  ε = 0.001; //error bound

While L L m a x and E F > ε , do
  1.  Calculate hidden layer outputs with Formula (1);
  2.  Calculate prediction values with Formula (2);
  3.  Calculate error: E = 1 2 k = 1 l ( y - k y k ) 2 ;
  4.  Update weights with Formulas (3) and (4);
  5.  L++;
End While
Return wij, wjk;

3.2. Random Vector Functional Link Network (RVFL)

RVFL is a single hidden layer feedforward neural network (SLFNN) that features direct connections between the input layer and the output layer nodes. The input weights from the input layer to the hidden layer are generated randomly within specified intervals, while the output weights are computed analytically using either the least-square method or the pseudo-inverse method.
Formally, the RVFL model with L hidden nodes can be described as:
f ( x i ) = k = 1 d β k x i k + k = d + 1 k + L β k g ( α k T X + b k ) ,   i = 1 , , N .
The optimization problem of the standard RVFL model can be written as:
min β R d + L   1 2 β 2 + 1 2 λ H β - Y 2 ,
where, H = [H1 H2]N×(d+L),
H 1 = | x 11 x 12 x 1 d x N 1 x N 2 x N d | ,  
H 2 = | θ ( α 1 x 1 + σ 1 ) θ ( α 2 x 1 + σ 2 ) θ ( α L x 1 + σ L ) θ ( α 1 x N + σ 1 ) θ ( α 2 x N + σ 2 ) θ ( α L x N + σ L ) | ,
β = [ β 1 β 2 β d + L ] ( d + L ) × m , Y = [ y 1 y 2 y N ] N × m .
where, βk = [βk1, βk2, …, βkm] is the output weight vector connecting the kth input (hidden) node to the output nodes, 1 ≤ kd + L, and αj = [αj1, αj2, …, αjd] is the weight vector connecting the jth hidden node to the input nodes, 1 ≤ jL. Additionally, xi = [xi1, xi2, …, xid] is the ith sample. For the target matrix, yi = [yi1, yi2, …, yim], 1 ≤ iN. Moreover, θ (·) and σi are the non-constant activation function and the bias term of ith hidden node, respectively. λ is a regularization coefficient.
The optimal solution to the problem (6) when δ = 1 λ = 0 reads as follows:
β = H+ Y,
where H+ represents the Moore-Penrose generalized inverse of the matrix H. The regularization term is employed to avoid the over-fitting issue. The RVFL algorithm is described in Algorithm 2.
Algorithm 2: RVFL algorithm
Inputs: Input X = { x 1 ,   , x N } , x i R d and output Y = { y 1 ,   , y N } , y i   R m ; Bias b = [ β 1 ,   , β L ] ; The number of nodes in hidden layer L ; the range of weights σ ; regularization parameter λ .
Outputs: W 0 .
1: Initialize: randomly selecting the initial weights in range of [ σ , σ ] .   W i n i t i a l   [ σ , σ ] L × d ;
2: Calculate activation values of the hidden layer: Z p r e = f a ( W i n i t i a l   X + b ) , where f a is an activation function ;
3: Establish enhancement layer to implement direct link: H = [ Z p r e ,   X ] ;
4: Calculate output weights through regularized least squares method: W 0 = Y × H T × ( λ I   + H H T   ) 1 ;    
5: Return W 0 , there by Y ^   = W 0 × H     can be calculated.

3.3. Stochastic Configuration Networks

Like RVFL in structure, the SCN is also a single hidden-layer neural network; however, the number of nodes in its hidden layer is not fixed but is dynamically increased through a supervisory mechanism [20]. Additionally, the input parameters between the input layer and the hidden layer are generated randomly within a specified domain, and the output parameters from the hidden layer are computed using least squares. The key to the SCN algorithm is the introduction of a supervisory mechanism that facilitates the addition of hidden nodes, helping to minimize the loss function and ultimately achieve convergence.
Given a set of N samples {( x i ,   y i ), xiRd, yiRm, i = 1, …, N}. An SCN learning network with L − 1 hidden nodes can be represented as follows:
f L 1 ( X ) = j = 1 L 1 β j g j ( w j , b j , X ) ,
and the current residual error is:
e L 1 = f f L 1 = [ e L 1 , 1 , e L 1 , 2 , , e L 1 , m ] .
where, g j ( ) is the activation function of hidden node j, βj is the output weight vector connecting the jth hidden node and w j and b j are input weights randomly generated in specified intervals [−λ, +λ]d and [−λ, +λ], λ is a positive scalar.
While a new hidden node is added to the network, the algorithm is considered convergent if the resulting sequence of residual errors is monotonically decreasing.
f L = f L 1 + β L g L ( w L X + b L ) ,   e L = f f L ,   e L e L 1 .
A result obtained from reference [20] satisfies the Formula (12), where:
β L = e L 1 , g L g L 2 .
where,
{ e L 1 , g L 2 b g 2 δ L , 0 < g L < b g , δ L = ( 1 r u L ) e L 1 2 , r > 0 ,   u L > 0 ,   1 r u L > 0 , lim L + u L = 0 .
The learning problem (11) can be converted to the following optimization task:
[ β 1 * , β 2 * , , β L * ] = arg min β f j = 1 L β j g j .
Its equivalent matrix form reads as:
β * = arg min β G L β Y 2 = G L + Y ,
where GL = [g1, g2, …, gL], G L + is the Moore-Penrose generalized inverse of the matrix G L .
The SCN algorithm is detailed in Algorithm 3. For clarity, several formulas are introduced below.
g L ( w L , b L , X ) = [ g L ( w L T x 1 + b L ) ,   g L ( w L T x 2 + b L ) , , g L ( w L T x N + b L ) ] T ,
β L , q = e L 1 , q ( X ) T g L ( w L , b L , X ) g L ( w L , b L , X ) T g L ( w L , b L , X ) ,   q = 1 , 2 , , m ,
and let
ξ L , q = e L 1 , q e L , q = β L , q ( 1 r u L ) e L 1 , q ( X ) T e L 1 , q ( X ) .
Algorithm 3: SCN algorithm
Input: X = {x1, x2, …, xN }, xiRd and outputs Y = {y1, y2, …, yN }, yiRm; Lmax, ε, Tmax; Λ = {λmin: ∆λ: λmax};
Output: β * ,   w * ,   b *
1. Initialize e0: = [y1, y2, …, yN]T, 0 < r < 1, two empty sets Ω and W;
2. While LLmax AND ‖e0‖ > ε, Do
Phase 1: Stochastic parameters configuration (steps 3–17):
3.   For λΛ
4.    For k = 1, 2, …, Tmax, Do
5.     Randomly assign vL and bL from [−λ, λ]d and [−λ, λ], respectively;
6.     Calculate gL, ξL,q based on Equations (17) and (19), and µL = (1 − r)/(L + 1);
7.     If min {ξL,1, ξL,2, …, ξL,m} ≥ 0
8.      Save wL and bL in W, ξ L = q = 1 m ξ L , q in Ω, respectively;
9.     Else go back to step 4;
10.   End If
11.  End For (corresponds to Step 4)
12.  If W is not empty
13.   Find wL*, bL* that maximize ξL in Ω, and set GL = [g1*, g2*, …, gL*];
14.   Break (go to step 18);
15.  Else randomly take τ ∈ (0, 1 − r), renew r: = r + τ, return to step 4;
16.  End If
17. End For (corresponds to step 3)
Phase 2: Output weights computation:
18. Obtain GL = [g1*, g2*, …, gL*];
19. Calculate β* = [β1*, β2*, …, βL*]T = GL+·Y;
20. Calculate eL = eL−1 − βL*gL*;
21. Renew e0: =eL; L: =L + 1;
22. End While
23. Return β1*, β2*, …, βL*, w* = [w1*, w2*, …, wL*], b = [b1*, b2*, …, bL*].

3.4. Evaluation Metrics

Three metrics RMSE, MAPE, and R2 are used to evaluate the performance of the proposed models, where RMSE is the root mean square error used to compare the prediction accuracy of different models, and MAPE is the mean absolute percentage error used to avoid positive and negative error cancellation on a prediction accuracy measurement. R2 is the coefficient of determination, the larger the better. All results are computed based on the average values over 100 independent trials.
R M S E = 1 N i = 1 N ( y i y ^ i ) 2 ,
M A P E = 1 N i = 1 N | y ^ i y i y i | ,
R 2 = 1 i = 1 N ( y ^ i y i ) 2 i = 1 N ( y i y - i ) 2 ,
where yi is the actual value of the sample, y - i is the predicted value of the sample, y ^ i is the mean of the actual values of samples, and N is the total number of samples.

4. Experimental Results

To evaluate the performance of the proposed models, we tested the influence of different structures and initial weights on the model performance, stability, convergence, and prediction accuracy of different models. Four datasets-DB1, DB2, DB3, and DB4 were selected for benchmarking, where DB1 and DB2 are single-input-single-output datasets, but DB3 and DB4 have multiple inputs and a single output. All datasets were divided into 70% as training sets and 30% as test sets.
(1)
DB1
f ( x ) = sin ( x ) cos ( x ) .
(2)
DB2
f ( x ) = x 2 x + 13 5 sin ( 2 π x 2 ) .
(3)
DB3 (stock) is a standard dataset for regression prediction from the KEEL website. The data is the daily stock prices of 10 aerospace companies from January 1988 to October 1991. The task is to predict the price of the 10th company, given the prices of the rest of the companies.
(4)
DB4 (concrete) is a standard dataset for regression prediction from the KEEL website. The concrete compressive strength is a nonlinear function of age and ingredients. The task is to predict the concrete compressive strength.

4.1. Network Structure

Network structure is critical for models based on neural networks. Different network architectures can significantly impact a model’s performance. However, the network structure of a Stochastic Configuration Network (SCN) is determined by the algorithm itself and is not predetermined. Therefore, discussions regarding various network structures are relevant only to Backpropagation (BP) and Random Vector Functional Link (RVFL) networks.
For the Backpropagation (BP) model, we have selected a double-hidden-layer architecture as the foundational structure, with a primary focus on determining the optimal number of nodes in each layer. The testing process is conducted repeatedly, varying the number of nodes in the two hidden layers, ensuring that the maximum number of nodes per hidden layer does not exceed 50. The results of the tests are assessed using the root mean square error (RMSE) metric. The findings are presented in Table 1.
For the RVFL, the objective of the test is to determine the optimal number of nodes in the hidden layer and the appropriate range of random weights based on the RMSE measure. The results are presented in Table 2.

4.2. Performance

We evaluated the performance of three models in conjunction with two widely recognized algorithms, Support Vector Regression (SVR) and Random Forest (RF), across four datasets. A total of 100 independent trials were conducted for each model and dataset. The evaluation metrics used include Root Mean Square Error (RMSE), Mean Absolute Percentage Error (MAPE), and R2. The test results are presented in Table 3, while the fitting curves are illustrated in Figure 2, Figure 3, Figure 4 and Figure 5. Based on the evaluation results, we found that the SCN model outperformed all other models across every dataset.

4.3. Convergence

Convergence is a crucial metric for assessing the performance of a model. The loss curves for the Backpropagation (BP) and the SCN were drawn in Figure 6, Figure 7, Figure 8 and Figure 9 for four different data sets. It is important to note that the Random Vector Functional Link (RVFL) curves are not included, as this model does not involve an iterative process. As shown in Figure 6, Figure 7, Figure 8 and Figure 9, the SCN exhibits a steeper decline in loss compared to the BP during iterations, ultimately resulting in a lower loss for the SCN across all data sets. This indicates that the SCN outperforms the BP in terms of model efficacy.

4.4. Stability

Stability is an important measure of an algorithm’s performance. For the same reason, the RVFL is not inspected; only the BP and the SCN are discussed for model stability. Ibid., four data sets are used for testing, and 100 trials are executed. The RMSE is used as the evaluation measure. The computational results are shown in Figure 10, Figure 11, Figure 12 and Figure 13. Compared with the BP, the RMSE curves of the SCN have smaller values and fluctuations, which indicates that the SCN has better stability and prediction accuracy with the algorithm proceeding.

4.5. Performance with Various Initial Weights

The range of initial weights greatly influences the learning capability and generalization property. We select 11 intervals in the range of [−500,500] to find the best range of initial weights. The computational results (RMSE) are shown in Table 4, Table 5, Table 6 and Table 7 and summarized in Table 8. The results show that the BP is the most sensitive to the changes in the range of weights; when and only when the weights are small enough, the BP can maintain stability. Specially, when the range of weights becomes large, the BP becomes extremely unstable. The RVFL performs a little better than the BP, but its accuracy is easily influenced by the range of weights. The SCN gains the best result with the test data, and the change in the accuracy is quite small compared with the other two algorithms.

4.6. Case Study

A real-world application of the unloading-time prediction will be addressed using the three models in this section. The data originates from the raw materials terminal of the studied steel plant, covering the years 2018 to 2020. Each ship-unloading record includes the following fields: BN (Berth Number), SN (Ship Name), RMN (cargo loaded), QRM (Quantity of Raw Materials), UD (Unloading Duration), ST (Starting Time of unloading), ET (End Time of unloading), SL1 (Ship Length), and SL2 (Storage Location in the stockyard). Additionally, there are other relevant fields not listed, such as berthing time, empty berthing time, and delay time, which can be appropriately incorporated into the calculations. Some unloading records are presented in Table 9.
A total of 70% of the collected data was utilized as the training set, while the remaining 30% served as the test set. The computational results are presented in Figure 14, which displays the prediction outcomes for 30, 50, and 100 ships. The experimental findings indicate that the SCN model achieves an average prediction accuracy of 87%, the RVFL model 85%, and the BP model 76%. Additionally, the running times for all three models are in seconds.

5. Conclusions

The prediction of ship-unloading time is important for improving the efficiency of ship-unloading operations and reducing the delay cost for a steel plant. In this paper, three neural network-based methods-BP, RVFL, and SCN are proposed for the prediction purpose. The performance of these models is evaluated and tested using the benchmark data sets, which are from open databases and published literature. Finally, the three models are used for the prediction of ship unloading time. The experiment results indicate that the three models are very effective. The prediction accuracies of the SCN, RVFL, and BP are 87%, 85%, and 76%, respectively. That shows that our methods are effective and exhibit application potential.
In this study, we developed neural networks to predict the unloading time at the raw materials terminal of a steel plant. This method holds significant practical implications, not only for the operation of the steel plant’s raw materials terminal but also for similar industries and industrial processes that are heavily reliant on upstream operation times. We must acknowledge the limitations of our method. While the results presented in this paper are satisfactory, every method has its constraints. As discussed in the algorithm testing section, each algorithm requires meticulous design and debugging to fulfill its intended purpose. Furthermore, varying data sets can also influence the outcomes. Future research will concentrate on integrating data processing technologies with forecasting methods to address practical industrial challenges.

Author Contributions

Conceptualization: Z.G. and D.L.; data curation: D.L.; formal analysis: D.W.; methodology: Z.G. and W.P.; experimental design and testing: D.L.; software: D.W. and Z.Y.; supervision: Z.G.; validation: Z.Y. and X.W.; writing original draft: Z.G.; writing review and editing: W.P. and D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the Major Program of the National Natural Science Foundation of China (72192830, 72192831) and the 111 Project (B16009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gao, Z.; Sun, D.; Zhao, R.; Dong, Y. Ship-unloading scheduling optimization for a steel plant. Inf. Sci. 2021, 544, 214–226. [Google Scholar] [CrossRef]
  2. Gao, Z.; Zhang, M.; Zhang, L. Ship-unloading optimization with differential evolution. Inf. Sci. 2022, 591, 88–102. [Google Scholar] [CrossRef]
  3. Zhang, X.F.; Chen, S.N.; Zhang, J.X. Adaptive sliding mode consensus control based on neural network for singular fractional order multi-agent systems. Appl. Math. Comput. 2022, 434, 127442. [Google Scholar] [CrossRef]
  4. Zhang, X.F.; Chen, Y.Q. Admissibility and robust stabilization of continuous linear singular fractional order systems with the fractional order α: The 0 < α < 1 case. ISA Trans. 2018, 82, 42–50. [Google Scholar] [PubMed]
  5. Yan, H.; Zhang, J.X.; Zhang, X.F. Injected infrared and visible image fusion via L1 decomposition model and guided filtering. IEEE Trans. Comput. Imaging 2022, 8, 162–173. [Google Scholar] [CrossRef]
  6. Tang, L.; Meng, Y. Data analytics and optimization for smart industry. Front. Eng. Manag. 2021, 8, 157–171. [Google Scholar] [CrossRef]
  7. Kong, D.; Zhu, J.; Duan, C.; Liu, L.; Chen, D. Bayesian linear regression for surface roughness prediction. Mech. Syst. Signal Process. 2020, 142, 106770. [Google Scholar] [CrossRef]
  8. An, X.; Jiang, D.; Liu, C.; Zhao, M. Wind farm power prediction based on wavelet decomposition and chaotic time series. Expert Syst. Appl. 2011, 38, 11280–11285. [Google Scholar] [CrossRef]
  9. Rojas, I.; Valenzuela, O.; Rojas, F.; Guillen, A.; Herrera, L.J.; Pomares, H.; Marquez, L.; Pasadas, M. Soft-computing techniques and ARMA model for time series prediction. Neurocomputing 2008, 71, 519–537. [Google Scholar] [CrossRef]
  10. Kundrat, D.M. Conceptual model of the iron blast furnace considering thermodynamic and kinetic constraints on the process. Metall. Trans. B 1989, 20, 205–218. [Google Scholar] [CrossRef]
  11. Gan, L.; Lai, C. A general viscosity model for molten blast furnace slag. Metall. Mater. Trans. B 2014, 45, 875–888. [Google Scholar] [CrossRef]
  12. Xiao, Z.; Ye, S.J.; Zhong, B.; Sun, C.X. BP neural network with rough set for short term load forecasting. Expert Syst. Appl. 2009, 36, 273–279. [Google Scholar] [CrossRef]
  13. Adelia, H.; Panakkat, A. A probabilistic neural network for earthquake magnitude prediction. Neural Netw. 2009, 22, 1018–1024. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, Y.; Wang, X.; Yu, J.; Zeng, T.; Wang, J. Multistep prediction for earthworks unloading duration: A fuzzy Att-Seq2Seq network with optimal partitioning and multi-time granularity modeling. Neural Comput. Appl. 2023, 35, 21023–21042. [Google Scholar] [CrossRef]
  15. Liang, S.K.; Wang, C.N. Modularized simulation for lot delivery time forecast in automatic material handling systems of 300 mm semiconductor manufacturing. Int. J. Adv. Manuf. Technol. 2005, 26, 645–652. [Google Scholar] [CrossRef]
  16. Xu, Y.; Chen, Q.; Quan, X. Robust berth scheduling with uncertain vessel delay and handling time. Ann. Oper. Res. 2012, 192, 123–140. [Google Scholar] [CrossRef]
  17. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  18. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  19. Malik, A.K.; Gao, R.; Ganaie, M.A.; Tanveer, M.; Suganthan, P.N. Random vector functional link network: Recent developments, applications, and future directions. Appl. Soft Comput. 2023, 143, 110377. [Google Scholar]
  20. Wang, D.; Li, M. Stochastic configuration networks: Fundamentals and algorithms. IEEE Trans. Cybern. 2017, 47, 3466–3479. [Google Scholar] [CrossRef]
  21. Dhingra, V.; Roy, D.; de Koster, R.B.M. A cooperative quay crane-based stochastic model to estimate vessel handling time. Flex. Serv. Manuf. J. 2017, 29, 97–124. [Google Scholar] [CrossRef]
  22. Bish, E.K. A multiple-crane-constrained scheduling problem in a container terminal. Eur. J. Oper. Res. 2003, 144, 83–107. [Google Scholar] [CrossRef]
  23. Al-Dhaheri, N.; Jebali, A.; Diabat, A. A simulation-based genetic algorithm approach for the quay crane scheduling under uncertainty. Simul. Model. Pract. Theory 2016, 66, 122–138. [Google Scholar] [CrossRef]
  24. Sun, D.; Tang, L.; Baldacci, R. A benders decomposition-based framework for solving quay crane scheduling problems. Eur. J. Oper. Res. 2019, 273, 504–515. [Google Scholar] [CrossRef]
  25. Sammarra, M.; Cordeau, J.F.; Laporte, G.; Monaco, M.F. A tabu search heuristic for the quay crane scheduling problem. J. Sched. 2007, 10, 327–336. [Google Scholar] [CrossRef]
  26. Tang, L.; Zhao, J.; Liu, J. Modeling and solution of the joint quay crane and truck scheduling problem. Eur. J. Oper. Res. 2014, 236, 978–990. [Google Scholar] [CrossRef]
  27. Kao, C.; Li, D.C.; Wu, C.; Tsai, C.C. Knowledge-based approach to the optimal dock arrangement. Int. J. Syst. Sci. 1990, 21, 2209–2215. [Google Scholar] [CrossRef]
  28. Kao, C.; Wu, C.; Li, D.C.; Lai, C.Y. Scheduling ship discharging via knowledge transformed heuristic evaluation function. Int. J. Syst. Sci. 1992, 23, 631–639. [Google Scholar] [CrossRef]
  29. Kao, C.; Lee, H.T. Coordinated dock operations: Integrating dock arrangement with ship discharging. Comput. Ind. 1996, 28, 113–122. [Google Scholar] [CrossRef]
  30. Kim, K.H.; Moon, K.C. Berth scheduling by simulated annealing. Transp. Res. Part B 2003, 37, 541–560. [Google Scholar] [CrossRef]
  31. Kim, B.I.; Chang, S.Y.; Chang, J.; Han, Y.; Koo, J.; Lim, K.; Shin, J.; Jeong, S.; Kwak, W. Scheduling of raw-material unloading from ships at a steelworks. Prod. Plan. Control 2011, 22, 389–402. [Google Scholar] [CrossRef]
  32. Lee, D.H.; Cao, J.X.; Shi, Q.; Chen, J.H. A heuristic algorithm for yard truck scheduling and storage allocation problems. Transp. Res. Part E 2009, 45, 810–820. [Google Scholar] [CrossRef]
  33. Tang, X.; Jin, J.G.; Shi, X. Stockyard storage space allocation in large iron ore terminals. Comput. Ind. Eng. 2002, 164, 107911. [Google Scholar] [CrossRef]
  34. Li, S.; Tang, L. Improved tabu search algorithms for storage space allocation in integrated iron and steel plant. In Proceedings of the 2005 ICSC Congress on Computational Intelligence Methods and Applications, Istanbul, Turkey, 15–17 December 2005; p. 6. [Google Scholar] [CrossRef]
  35. Kim, B.I.; Koo, J.; Park, B.S. A raw material storage yard allocation problem for a large-scale steelwork. Int. J. Adv. Manuf. Technol. 2009, 41, 880–884. [Google Scholar] [CrossRef]
  36. Zhang, B.; Shin, Y.C. A probabilistic neural network for uncertainty prediction with applications to manufacturing process monitoring. Appl. Soft Comput. 2022, 124, 108995. [Google Scholar] [CrossRef]
  37. Li, W.; Wang, X.X.; Han, H.G.; Qiao, J. A PLS-based pruning algorithm for simplified long-short term memory neural network in time series prediction. Knowl.-Based Syst. 2022, 254, 109608. [Google Scholar] [CrossRef]
  38. Kosanoglu, F. Wind speed forecasting with a clustering-based deep learning model. Appl. Sci. 2022, 12, 13031. [Google Scholar] [CrossRef]
  39. Hussein, A.M.; Elaziz, M.A.; Abdel Wahed, M.S.M.; Sillanpää, M. A new approach to predict the missing values of algae during water quality monitoring programs based on a hybrid moth search algorithm and the random vector functional link network. J. Hydrol. 2019, 575, 852–863. [Google Scholar] [CrossRef]
  40. El-Said, E.M.S.; Elaziz, M.A.; Elsheikh, A.H. Machine learning algorithms for improving the prediction of air injection effect on the thermohydraulic performance of shell and tube heat exchanger. Appl. Therm. Eng. 2021, 185, 116471. [Google Scholar] [CrossRef]
  41. Wang, W.; Wang, D. Prediction of component concentrations in sodium aluminate liquor using stochastic configuration networks. Neural Comput. Appl. 2020, 32, 13625–13638. [Google Scholar] [CrossRef]
  42. Li, K.; Yang, C.; Wang, W.; Qiao, J. An improved stochastic configuration network for concentration prediction in wastewater treatment process. Inf. Sci. 2023, 622, 148–160. [Google Scholar] [CrossRef]
Figure 1. The ship-unloading system of a steel plant.
Figure 1. The ship-unloading system of a steel plant.
Applsci 14 08213 g001
Figure 2. Fitting curves on DB3.
Figure 2. Fitting curves on DB3.
Applsci 14 08213 g002
Figure 3. Fitting curves on DB4.
Figure 3. Fitting curves on DB4.
Applsci 14 08213 g003
Figure 4. Fitting curves on DB1.
Figure 4. Fitting curves on DB1.
Applsci 14 08213 g004
Figure 5. Fitting curves on DB2.
Figure 5. Fitting curves on DB2.
Applsci 14 08213 g005
Figure 6. Loss curve of training (DB1).
Figure 6. Loss curve of training (DB1).
Applsci 14 08213 g006
Figure 7. Loss curve of training (DB2).
Figure 7. Loss curve of training (DB2).
Applsci 14 08213 g007
Figure 8. Loss curve of training (DB3).
Figure 8. Loss curve of training (DB3).
Applsci 14 08213 g008
Figure 9. Loss curve of training (DB4).
Figure 9. Loss curve of training (DB4).
Applsci 14 08213 g009
Figure 10. RMSE within 100 trials (DB1).
Figure 10. RMSE within 100 trials (DB1).
Applsci 14 08213 g010
Figure 11. RMSE within 100 trials (DB2).
Figure 11. RMSE within 100 trials (DB2).
Applsci 14 08213 g011
Figure 12. RMSE within 100 trials (DB3).
Figure 12. RMSE within 100 trials (DB3).
Applsci 14 08213 g012
Figure 13. RMSE within 100 trials (DB4).
Figure 13. RMSE within 100 trials (DB4).
Applsci 14 08213 g013
Figure 14. Unloading time predictions for 30, 50, and 100 ships.
Figure 14. Unloading time predictions for 30, 50, and 100 ships.
Applsci 14 08213 g014aApplsci 14 08213 g014b
Table 1. BP network structure on datasets.
Table 1. BP network structure on datasets.
DatasetNumber 1Number 2Range
DB1530[0, 50]
DB21040[0, 50]
DB35010[0, 50]
DB4305[0, 50]
Remark: Number 1: the number of nodes in the first hidden layer; Number 2: the number of nodes in the second hidden layer; Range: the range of the number of nodes.
Table 2. RVFL network structure on datasets.
Table 2. RVFL network structure on datasets.
DatasetNumberRange-1Range-2
DB180[−40, 40][−50, 50]
DB2100[−20, 20][−50, 50]
DB3100[−5, 5][−50, 50]
DB4100[−10, 10][−50, 50]
Remark: Number: the number of nodes in the hidden layer; Range-1: the range of the random weights; Range-2: the maximum range of the permitted random weights.
Table 3. Performance comparison among different models on four datasets.
Table 3. Performance comparison among different models on four datasets.
DatasetModelRMSEMAPER2
Mean ± (SD)BestMean ± (SD)Best
DB1SVR0.0703 ± 0.00020.070030.1491 ± 0.915228.31370.0045
RF0.0085 ± 0.00070.00767.7704 ± 2.74913.77310.9994
BP0.0122 ± 0.01230.011537.6948 ± 23.317613.75740.9976
RVFL0.0484 ± 0.00060.048466.1857 ± 14.456239.93760.9824
SCN0.0074 ± 0.00240.00245.6983 ± 2.61240.84400.9995
DB2SVR0.1996 ± 0.00090.198221.4102 ± 0.198621.14810.0594
RF0.0064 ± 0.00050.00560.4135 ± 0.01500.39440.9993
BP0.0761 ± 0.01000.06016.2884 ± 1.94446.41210.9063
RVFL0.1084 ± 0.00050.10847.3193 ± 0.02257.31710.8197
SCN0.0008 ± 0.00390.00080.0050 ± 0.00080.00380.9999
DB3SVR0.0536 ± 0.00270.04922.7469 ± 0.15622.51310.9460
RF0.03968 ± 0.00360.03331.9413 ± 0.15551.65620.9649
BP0.0602 ± 0.01550.038610.4826 ± 1.96017.57250.9454
RVFL0.0687 ± 0.00010.068614.2589 ± 0.000114.25780.9170
SCN0.0388 ± 0.00120.03571.8707 ± 0.04831.76500.9681
DB4SVR0.0885 ± 0.00450.082320.8771 ± 1.071019.15710.8147
RF0.0961 ± 0.00800.078518.3751 ± 1.185116.42980.7839
BP0.0911 ± 0.00870.079619.5973 ± 1.731017.18500.8029
RVFL0.1932 ± 0.00010.192241.6601 ± 2.480333.19100.4144
SCN0.0870 ± 0.00600.076721.8238 ± 1.226119.37930.8135
Table 4. Performance with various weights on DB1.
Table 4. Performance with various weights on DB1.
Range of WeightsDB1
RVFLBPSCN
[−1, 1]0.281480.145960.00586
[−5, 5]0.139760.491690.00678
[−10, 10]0.081510.562050.00748
[−20, 20]0.054530.598310.00877
[−30, 30]0.046940.611420.00837
[−40, 40]0.048780.615640.00895
[−50, 50]0.056610.562050.00897
[−60, 60]0.162080.562050.00968
[−80, 80]0.217590.607720.00995
[−100, 100]0.160810.616100.00954
[−500, 500]0.256340.613740.01169
Table 5. Performance with various weights on DB2.
Table 5. Performance with various weights on DB2.
Range of WeightsDB2
RVFLBPSCN
[−1, 1]0.183560.204130.00474
[−5, 5]0.121670.529180.00704
[−10, 10]0.098080.542080.00817
[−20, 20]0.096790.537240.00582
[−30, 30]0.089420.578660.00853
[−40, 40]0.090560.604720.00962
[−50, 50]0.101270.560140.00937
[−60, 60]0.140490.592120.00942
[−80, 80]0.169430.560780.00966
[−100, 100]0.167990.510730.01021
[−500, 500]0.319120.542410.01235
Table 6. Performance with various weights on DB3.
Table 6. Performance with various weights on DB3.
Range of WeightsDB3
RVFLBPSCN
[−1, 1]0.073250.045930.03709
[−5, 5]0.047590.506020.07701
[−10, 10]0.042870.571130.07612
[−20, 20]0.076980.046060.07848
[−30, 30]0.094100.558660.10125
[−40, 40]0.104060.523180.09185
[−50, 50]0.120920.533170.08473
[−60, 60]0.121940.532910.08901
[−80, 80]0.131080.530270.08757
[−100, 100]0.123790.540400.08571
[−500, 500]0.132520.555920.08917
Table 7. Performance with various weights on DB4.
Table 7. Performance with various weights on DB4.
Range of WeightsDB4
RVFLBPSCN
[−1, 1]0.183870.096370.10241
[−5, 5]0.139770.409240.10093
[−10, 10]0.144360.570550.09312
[−20, 20]0.170750.550350.10734
[−30, 30]0.197140.562270.13241
[−40, 40]0.206600.531620.12498
[−50, 50]0.211260.556770.10198
[−60, 60]0.216760.552450.10081
[−80, 80]0.220010.565670.10748
[−100, 100]0.230560.551520.09669
[−500, 500]0.244820.503430.10091
Table 8. Algorithm stability with various initial weights.
Table 8. Algorithm stability with various initial weights.
AlgorithmDB1DB2DB3DB4
MeanStdMeanStdMeanStdMeanStd
RVFL0.136950.086630.143490.067790.097190.032760.196900.03396
BP0.544250.137440.523840.109620.449420.200260.495480.14010
SCN0.008730.001600.008630.002130.081640.016460.106280.01192
Table 9. Ship-unloading records.
Table 9. Ship-unloading records.
BNSNRMNQRMUDSTETSL1SL2
7FS8Ore fines 123,585375 November 2018 15:007 November 2018 4:00173OB24
1MTOYD-S/ONM-N51,500235 November 2018 9:006 November 2018 8:00292OA16
1BYOHY-S90,000347 November 2018 10:008 November 2018 20:00320OA18
2ZC258Ore fines 433,000355 November 2018 9:006 November 2018 20:00189OB32
4BXH6dolomite110035 November 2018 13:005 November 2018 16:0050OC23
5JN9Limestone5005186 November 2018 2:006 November 2018 20:00110OC11
2ZC258Ore fines 412,000136 November 2018 9: 006 November 2018 22:00189OB16
2BYOHY-S90,000347 November 2018 3:308 November 2018 20:00320OA36
4JN9Limestone5005176 November 2018 9:007 November 2018 2:00110OC26
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, Z.; Li, D.; Wang, D.; Yu, Z.; Pedrycz, W.; Wang, X. Prediction of Ship-Unloading Time Using Neural Networks. Appl. Sci. 2024, 14, 8213. https://doi.org/10.3390/app14188213

AMA Style

Gao Z, Li D, Wang D, Yu Z, Pedrycz W, Wang X. Prediction of Ship-Unloading Time Using Neural Networks. Applied Sciences. 2024; 14(18):8213. https://doi.org/10.3390/app14188213

Chicago/Turabian Style

Gao, Zhen, Danning Li, Danni Wang, Zengcai Yu, Witold Pedrycz, and Xinhai Wang. 2024. "Prediction of Ship-Unloading Time Using Neural Networks" Applied Sciences 14, no. 18: 8213. https://doi.org/10.3390/app14188213

APA Style

Gao, Z., Li, D., Wang, D., Yu, Z., Pedrycz, W., & Wang, X. (2024). Prediction of Ship-Unloading Time Using Neural Networks. Applied Sciences, 14(18), 8213. https://doi.org/10.3390/app14188213

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop