Next Article in Journal
Sustainable Provision of School Buildings in The Netherlands: An Empirical Productivity Analysis of Local Government School Building Operations
Next Article in Special Issue
Determining an Improved Traffic Conflict Indicator for Highway Safety Estimation Based on Vehicle Trajectory Data
Previous Article in Journal
Assessing Multi-Micronutrients Deficiency in Agricultural Soils of India
Previous Article in Special Issue
Crash Prediction Models for Horizontal Curve Segments on Two-Lane Rural Roads in Thailand
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning-Based Models for Accident Prediction at a Korean Container Port

Department of Transportation & Logistics Engineering, Hanyang University, Ansan 15588, Korea
*
Author to whom correspondence should be addressed.
Sustainability 2021, 13(16), 9137; https://doi.org/10.3390/su13169137
Submission received: 20 July 2021 / Revised: 10 August 2021 / Accepted: 12 August 2021 / Published: 16 August 2021

Abstract

:
The occurrence of accidents at container ports results in damages and economic losses in the terminal operation. Therefore, it is necessary to accurately predict accidents at container ports. Several machine learning models have been applied to predict accidents at a container port under various time intervals, and the optimal model was selected by comparing the results of different models in terms of their accuracy, precision, recall, and F1 score. The results show that a deep neural network model and gradient boosting model with an interval of 6 h exhibits the highest performance in terms of all the performance metrics. The applied methods can be used in the predicting of accidents at container ports in the future.

1. Introduction

Most of the existing studies on the risk assessment of maritime ports have focused on port security [1,2,3] and port safety [4,5,6]. Notably, most studies focusing on the security considered only unusual events such as hostile attacks [7] and the smuggling of weapons [8], and most studies pertaining to port safety focused on accidents that occurred during usual port activities such as loading, discharging, importing, and exporting. In this regard, research on port safety must be emphasized over that of port security.
In the maritime field, several researchers have examined safety by predicting vessel accidents on the waterway [9], forecasting coastal waves [10], and examining ship collisions [11,12], among other aspects. Moreover, although several researchers have implemented risk assessment methods to identify the risk factors associated with container ports [4,5,6], research regarding the prediction of container port accidents remains limited.
In a container port, several activities, including the loading, discharging, importing, and exporting of containers are performed by port workers using equipment such as yard tractors and container cranes. Owing to these extensive activities, container ports are prone to accidents, such as equipment–equipment collisions, equipment–container collisions, injuries, and container damages during discharging, loading, and moving. These accidents may result in damages to workers, equipment, and containers, as well as in economic loss. In particular, according to the statistics of occupational accidents in Korea [13], as shown in Table 1, the total number of accidents occurring at all container ports in Korea led to 4 fatalities and 91 injuries in 2015 and 3 fatalities and 96 injuries in 2019; the damages have gradually increased every year, except for in 2016. The estimated economic losses resulting from the accidents were approximately KRW 18.8 and 20.5 billion in 2015 and 2018, respectively. Moreover, the number of accidents involving minor injuries is considerably larger. Therefore, predicting accidents at a container port is essential for minimizing the economic loss in port operation and enhancing the port safety.
To predict uncommon events such as accidents, machine learning methods including neural network models, random forest, and gradient boosting have been used. In the transportation field, these methods have been widely applied to predict traffic accidents on roads. Specifically, neural networks have been used to predict vehicle crash accidents on roads [14] and the duration of traffic accidents [15]. Random forest models have been applied to detect traffic accidents [16] and identify taxi drivers with a high risk of accidents [17]. Gradient boosting models have been used to predict traffic accidents on roads [18,19] and railways [20]. Moreover, several comparative studies have been performed regarding machine learning methods, including the use of neural networks to predict the severity associated with traffic accidents [21,22].
In the field of maritime and port logistics, machine learning methods have been applied in certain studies to predict the risk associated with accidents. Several researchers have applied neural networks to predict vessel accidents on the waterway [9] and risk in maritime safety [23] and forecast coastal waves [10]. A neural network model was applied to develop a ship collision risk model [11]. Moreover, a random forest model was applied to predict the severity of ship collision accidents [12]. However, relatively few studies have applied machine learning methods to predict accidents in a container port.
In Korea, most studies associated with analyzing and predicting accidents have been conducted in road transportation fields, considering the characteristics of vehicles and pedestrians [24,25,26,27]. Most studies on container ports in Korea were focused on analyzing the safety factors related to loading and unloading activities [28], developing risk assessment methods based on the type of accidents occurring in the port [29,30], and examining the influence of education on the risk factors of accident occurrences in the ports [31]. Research on the prediction of accidents that occur in activities associated with the port is relatively limited.
Therefore, in this study, the accidents that can occur in a container port are predicted by applying machine learning methods, including a neural network model, random forest model, and gradient boosting model to a historical dataset of accidents that occurred at a container port in Busan, Korea.

2. Methods and Dataset

2.1. Methods

2.1.1. Neural Network Model

A neural network is a machine learning model that has been widely applied in prediction applications. The basic element of a neural network model is the processing node [32]. In the model, each processing node performs two functions: (1) To sum the input values of the node; (2) To pass this information through an activation function to generate an output. All the processing nodes in a neural network are arranged in layers, and each layer is interconnected to the following layer. There is no interconnection among the nodes in the same layer. In general, a neural network model has an input layer that functions as a distribution structure for the data being used as the input, and this layer is not involved in any type of processing. The input layer is followed by the hidden layer, which consists of one or more processing layers. The final processing layer is the output layer.
The intersections between the nodes have a certain weight. When a value is passed through the input layer, the value is multiplied by the weight and summed to derive the total input n j to the unit, as shown in Equation (1).
n j = i w ji o i
where w ji is the weight of the interconnection from the input unit i to another unit j, and o i is the output of i. The total input calculated using Equation (1) is transformed by the activation function to produce an output   o j of j.
For a neural network, the parameters must be set by the users. These parameters include the number and type of hidden layers, number of nodes in each hidden layer, activation function for the output, weight initialization method, optimization algorithm, learning rate of the optimization algorithm, batch size (i.e., number of training samples used in one iteration), and number of epochs (one epoch is defined as the period in which an entire training dataset passes once through the neural network).
Neural networks are trained by searching for the optimal set of weights for the mapping function from the inputs to the outputs with the given dataset by initializing and updating the weights. In this study, a neural network with the adaptive moment estimation (Adam) optimizer is applied through the Keras Python package [33].

2.1.2. Random Forest Model

Random forests represent an ensemble machine learning technique. A random forest model employs an advanced decision tree analysis method to overcome overfitting issues, which is a drawback of decision tree analyses [34]. In the learning process, a random forest model generates classification trees by selecting subsets of the given dataset and randomly selecting subsets of variables for prediction. The number of trees is set in advance, and the average results for each tree are derived as the final outputs, based on the results generated in each tree. The learning process of random forests using bootstrap sampling consists of the following steps: (i) generate trees and datasets from the training dataset by sampling the bootstrap, (ii) train a basic sorter for the trees, (iii) combine the basic sorter (i.e., tree) into one sorter (i.e., random forest), and (iv) derive the final results of prediction by the majority voting rule. The observed values in the random forest that are not included in the learning process are considered out-of-bag (OOB) values, and they are used in the model validation. OOB values are used to estimate the predicted values and classify variables that cause anomalies. The number of times OOB values are selected in all trees varies for each tree, and the expected values are different for each tree. The probability of correctly predicting the OOB values for each observation in the original category, i.e., category k, can be calculated using Equation (2).
Prob k ( x i ) = j OOB i ¯ I [ y ( x i , T j ) = k ] | OOB i | ,   for   k
where i is an indicator that is set as 1 and 0 when the value in the parenthesis is true and false, respectively. y ( x i , T j ) is the predicted category, and T j is the jth decision tree among the generated trees T, in the forest. OOB i represents a group of decision trees that are not used in the learning process and are bagged as an observed variable. If a set of decision trees does not include x i , the ratio of the number of decision trees predicting x i to category k is Prob k ( x i ) . For a random forest, the Gini importance is computed and used to indicate the importance of the independent variables. At each node τ within the binary trees t of the random forest, the optimal split is found using the Gini impurity i( τ ), which indicates how well a potential split separates the samples of the two categories in a particular node.
Let p k = n k n represent the fraction of n k samples from category k = {0, 1} among the total n samples at node τ . The Gini impurity i( τ ) can be calculated using Equation (3).
i ( τ ) = 1 p 1 2 p 0 2 .
The change in i( τ ), Δi, which can be attributed to the splitting and transmission of the samples to two subnodes τ l and τ r (with sample fractions p l = n l n and p r = n r n , respectively) based on a threshold t θ for variable θ , can be calculated using Equation (4).
Δ i ( τ ) = i ( τ ) p l i ( τ l ) p r i ( τ r )
In the search for all variables θ available at the node and all possible thresholds t θ , the pair { θ , t θ } leading to the maximum Δ i is determined. The change in the Gini impurity resulting from the optimal split, Δ i θ ( τ , T ) , is recorded and accumulated for all nodes τ in all trees T in the forest for all θ values, as shown in Equation (5).
I G ( θ ) = T τ Δ i θ ( τ , T )
The Gini importance I G indicates how often a particular variable θ is selected for a split and the contribution of this value to the classification problem.
This study adopts the scikit-learn package in Python, an open-source programming language software that provides a user-customizable random forest model [35].

2.1.3. Gradient Boosting Decision trees

Gradient boosting decision trees are decision tree models that can prevent overfitting and demonstrate an enhanced prediction accuracy [36]. In gradient boosting decision trees, F(x) is assumed to be an approximation function of the output y based on a set of input variables x. The squared error function is applied as the loss function L to estimate the approximation function, as indicated in Equation (6).
L ( y , F ( x ) ) = [ y F ( x ) ] 2
Assuming that the number of splits is J for each regression tree, each tree partitions the input space into J disjoint regions R 1 m , , R jm and predicts a constant value b jm for region R jm . In this case, each decision tree exhibits the additive form, as indicated in Equation (7).
h m ( x ) = j = 1 J b jm I ( x R jm )
I = { 1         if   x R jm 0         otherwise
Using the training data, the gradient boosting model iteratively constructs M decision trees h 1 ( x ) , , h M ( x ) . The updating approximation function F m ( x ) and gradient descent step size ρ m can be defined using Equation (8) and (9).
F m ( x ) = F m 1 ( x ) + ρ m j = 1 J b jm I ( x R jm )
ρ m = argmin ρ i = 1 N L ( y i , F m 1 ( x i ) + ρ j = 1 J b jm I ( x R jm ) )
With a separate optimal γ jm for each region R jm , b jm can be discarded. Equation (8) can be expressed as Equation (10):
F m ( x ) = F m 1 ( x ) + j = 1 J γ jm I ( x R jm )
and the optimal γ jm can be calculated using Equation (11).
γ jm = argmin γ x R jm L ( y i , F m 1 ( x i ) + γ )
= argmin γ x R jm ( y ~ i γ ) 2
where   y ~ i = [ L ( y i ,F ( x i ) ) F ( x i ) ] F m ( x ) = F m 1 ( x )
Gradient boosting decision trees build the model sequentially and update it by minimizing the expected value of the loss function. To avoid overfitting and increase the prediction accuracy, a learning rate strategy is applied. The learning rate is used to scale the contribution of each tree model by introducing a factor ξ   ( 0   <   ξ     1 ) , as indicated in Equation (12).
F m ( x ) = F m 1 ( x ) + ξ · j = 1 J γ jm I ( x R jm ) ,   0   <   ξ     1
In Equation (12), a smaller ξ corresponds to a higher learning rate. Through the learning rate strategy, the overfitting issue can be avoided by reducing the impact of an additional tree. A smaller learning rate leads to a higher reduction in the loss function value. However, a larger number of trees may be added to the model. In this case, another parameter C, which refers to the number of splits, can be used for fitting each decision tree. This parameter represents the depth of variable interaction in a tree. Increasing C can help capture more complex interactions among variables and exploit the strength of gradient boosting decision trees. Depending on the value of the learning rate and C, the optimal number of trees can be identified by examining how well the model fits the test dataset. The performance of gradient boosting decision trees depends on the combination of the learning rate and tree complexity. In this study, the gradient boosting model is applied through the scikit-learn package in Python, which provides a user-customizable model [35].

2.1.4. Synthetic Minority Oversampling Technique (SMOTE)

Despite their importance in analyzing the risk to safety, accident data are usually a minority class owing to their relative unavailability. Therefore, if oversampling to the minority class is not performed, the results from machine learning models may be skewed to the majority class (i.e., non-accident data), leading to the inferior performance of the models. SMOTE is an oversampling method that can be used to overcome this imbalanced data issue [37]. In this method, a minority class is oversampled by creating synthetic samples. SMOTE generates synthetic samples in the following order: 1) Consider the difference between the feature vector (sample) and its nearest neighbor; 2) Multiply this difference by a random number between 0 and 1; 3) Add this value to the feature vector under consideration. Through this process, SMOTE effectively enables the enhanced generalization of the minority class. Because SMOTE can address the imbalanced data issue, several studies on the predicting of uncommon events such as accidents have applied this method to enhance the model performance [38,39,40]. This study uses time-series datasets, in which the amount of historical data of accidents is small. Consequently, the datasets contain accident data as a minority class, which leads to an imbalanced data issue. To address this problem, SMOTE is applied to the dataset in the model training phase.

2.2. Dataset

Three datasets were aggregated and used in the analysis: (1) An operation dataset for 2017, 2018, and 2020, formulated at container port A in Busan, Korea; (2) A historical dataset of the accidents that occurred at container port A in 2017, 2018, and 2020, formulated at container port A; (3) Weather observation dataset for Busan for 2017, 2018, and 2020, collected by the Korea Meteorological Administration [41]. Notably, because the data for 2019 are missing in the second dataset, the analysis was performed using the data for 2017, 2018, and 2020.
The first dataset contained information regarding the movements of containers (i.e., loading, discharging, importing, and exporting) and equipment (e.g., yard trucks and container cranes). The second dataset contained information of the accidents, including the time at which an accident occurred and type of accident (i.e., injury, collision, etc.). The weather dataset included data of the temperature, humidity, wind speed, and precipitation.
Table 2, Table A1 and Table A2 present the results of the basic statistical analysis of the datasets. As shown in Table 2, at container port A, 26, 39, and 78 accidents occurred in 2017, 2018, and 2020, respectively. Therefore, the number of accidents has increased from 2017 to 2020. Table A1 and Table A2 present the results for the third and first datasets, respectively. These datasets are integrated into a single time-series dataset. Five datasets are generated according to the intervals of 1 h, 3 h, 6 h, 12 h, and 24 h.

3. Results

The data of 2017 and 2018 were used for the model training, and the data of 2020 were used for testing the models. As described in Section 2.1.4, SMOTE was applied to the dataset for training to overcome an imbalanced data issue and enhance the model performances. From the training and testing datasets, hourly weather data (i.e., temperature, precipitation, wind speed, and humidity) from the third dataset and hourly operation data for terminal A (i.e., number of ships in berth, number of containers loaded/unloaded from the ships, number of containers imported/exported from the port, number of trucks entering/exiting the port, and number of container cranes/yard equipment/yard trucks in operation) were used as the input variables. The occurrence of accidents (or lack thereof) in the time intervals was used as the output variable. Moreover, as described in Section 2, the model hyperparameters were the learning rate, max depth, max features, min samples leaf, min samples split, n-estimator, and subsample. Table 3, Table 4 and Table 5 list the values of hyperparameters for each model that can enhance the model performance.
To determine the model accuracy in the training process, a 10-fold cross-validation method was applied to the training dataset. Specifically, every model in this study was trained 10 times with 10 datasets for each time interval. The results of the cross validations for the models are presented in Table 6, Table 7 and Table 8.
As shown in Table 6, Table 7 and Table 8, the average accuracies of the models in the training phase were higher than 90%, except for the accuracy of the deep neural network with a 24 h interval. This finding shows that the model performance is acceptable in terms of its accuracy. In the testing phase, the model performance was evaluated using the test data. The model performance indicators, specifically, the accuracy, precision, recall, and F1 score, were calculated using Equations (13)–(16), respectively.
Accuracy = True   Accidents + True   NoAccidents True   Accidents + True   NoAccidents + False   Accidents + False   NoAccidents
Precision = True   Accidents True   Accidents + False   Accidents
Recall = True   Accidents True   Accidents + False   NoAccidents
F 1   score = 2 × Precision × Recall Precision + Recall
Table 9 summarizes the results of implementing the models over the testing dataset. All the models exhibit the highest accuracy for the 1 h interval. As the time intervals increase, the accuracies decrease, although the precision, recall, and F1 score increase. For the deep neural network model, random forest model, and gradient boosting model, the accuracy for the 1 h interval is approximately 98.2%, 90.4%, and 98.3%, respectively, which decreases to approximately 76.5%, 76.8%, and 62.3% for the 24 h interval, respectively. The gradient boosting model exhibits the highest increase in the precision, recall, and F1 score as the time interval increases (i.e., from 1.4%, 1.3%, and 1.4% for the 1 h interval to 20.2%, 39.1%, and 26.6% for the 24 h interval, respectively). The second-largest increase pertains to the deep neural network model (i.e., from 2.3%, 2.6%, and 2.4% for the 1 h interval to 17.7%, 21.9%, and 19.6% for the 24 h interval, respectively), and the lowest increase pertains to the random forest model (i.e., from 1.1%, 11.5%, and 2.1% for the 1 h interval to 18.2%, 9.4%, and 12.4% for the 24 h interval, respectively).

4. Discussion

The results presented in Section 3 show that all the considered models exhibit different performances in predicting accidents in terms of their accuracy, precision, recall, and F1 score under various time intervals. As shown in Table 9, the precision, recall, and F1 score of the models increase as the time intervals increase, whereas the accuracy decreases. In addition to the accuracy, the precision, recall, and F1 score are important measures of the model performance. Higher precision, recall, and F1 score values correspond to a higher model performance. In comparison, a model with an accuracy of at least 85% is considered to have a high performance. Several studies that have applied machine learning methods to predict the accidents indicated that the highest accuracy of the existing models in predicting accidents was 85% [42,43,44]. Therefore, in terms of the precision, recall, and F1 score, as well as the accuracy, the models using the input data with a time interval of 6 h exhibit a reasonable performance. For the 6 h interval, the deep neural network exhibits an accuracy, precision, recall, and F1 score of 90.9%, 7.4%, 6.7%, and 7.0%, respectively. The corresponding values for the random forest model are 86.9%, 4.7%, 8.0%, and 5.9%. The corresponding values for the gradient boosting model are 85.1%, 8.7%, 20.0%, and 12.1%. The gradient boosting model exhibits the best F1 score of 12.1%, followed by the deep neural network (11.2%). Therefore, the gradient boosting model and deep neural network model are preferable for predicting the accident occurrence at a container port.
In this study, accident data for container port A for three years was used in the analysis. In Korea, the accident data for each container port is collected and managed by the container terminal operator of that port, and there is no legal organization that manages the overall accident data for all container ports nationwide. The accident data for each container port is classified as confidential, and so it is not fully available to the public, except for occupational accidents that are compensated by insurance, whereas the terminal operating data for the port is partially available. Moreover, it is difficult to obtain data because accidents are considered to be a sensitive issue in port operations, and so the terminal operator rarely provides it. This makes it difficult to collect a sufficient amount of accident data at container ports nationwide. As a result, this study used an accident dataset of limited size that was available for analyzing whether accidents occurred or not, rather than focusing on the types of accident that occurred.

5. Conclusions

This study adopts machine learning methods to predict the accidents that can occur in a container port. Time-series datasets with various time intervals are applied, and the model performance is evaluated based on these intervals. According to the results, as the time interval increases, the accuracy in predicting accidents decreases and the precision, recall, and F1 score increases. In terms of all the indicators, the models using the dataset with a 6 h interval exhibit the highest performance. Under the same time interval, the gradient boosting model and deep neural network model are the best in predicting accidents at the container port. These results demonstrate that machine learning methods can be applied to predict accidents at container ports.
Nevertheless, this study involves certain limitations that must be addressed in future work. The operation dataset of the container port and weather dataset are considered as independent variables affecting the occurrence of accidents. However, other variables, such as the accident type, cause, and time of incidence, can also directly affect the occurrence of accidents. Accident data for container ports, especially in Korea, are of significance, and can affect the port operation. However, although the annually aggregated statistics of accidents are available [13], it is difficult to collect an adequately large dataset including the raw data from the port. The accident dataset from container port A contains accident types, including vehicle collision, container damage, injury, and death. However, considering the total number of accidents (i.e., 26 accidents in 2017, 39 accidents in 2018, and 78 accidents in 2020), it is difficult to specify the accidents by type. Therefore, in this study, the number of accidents is considered as the output variable. In future work, by categorizing adequate accident data according to the accident type (i.e., injury, collision, and container damage, among other events), it may be possible to predict the accident type, analyze the factors affecting the accidents, and assess the risks in a container port. Moreover, accidents at a container port happen during work in any environment. Operational data for a container terminal is stored as an hourly dataset, and shows the changes in working status at the port well. However, these data and accident data together cannot provide enough information about the situation in which an accident occurred, because the accident data contains reasons for the occurrence (i.e., carelessness during working, rough driving, and so on) and the hourly aggregated operations dataset cannot describe the dynamic situation in which the accident happened. Therefore, to analyze accidents with their causes, the dynamic operating situation at the container port when the accident happened should be considered, rather than the hourly aggregated data. In a future study, with disaggregated operational data and accident data for a container port, it may be possible to not only predict accident occurrence but also to analyze accident risk.

Author Contributions

Conceptualization, J.H.K. and G.L.; methodology, J.H.K. and G.L.; software, J.K.; validation, J.H.K. and J.K.; formal analysis, J.H.K.; writing—original draft preparation, J.H.K., J.K., G.L., and J.P.; writing—review and editing, J.H.K., G.L. and J.P.; supervision, G.L.; project administration, J.P.; funding acquisition, G.L. and J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was a part of the project titled ‘Smart Port IoT convergence and operation technology development [20190399]’, funded by the Ministry of Oceans and Fisheries, Korea.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data was obtained from container port A in Korea and are available with the permission of container port A.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Basic Statistical Analysis of the Dataset Associated with Container Port A in Korea

Table A1. Basic statistical analysis of the weather dataset pertaining to Busan, Korea.
Table A1. Basic statistical analysis of the weather dataset pertaining to Busan, Korea.
YearMonthAverage Temperature
(℃)
Average Precipitation
(mm)
Average Wind Speed
(m/s)
Average Humidity
(%)
201714.10.03.346.1
25.50.03.544.7
38.90.13.152.8
414.90.13.863.4
519.00.13.363.1
621.20.12.970.3
726.00.23.582.2
827.00.12.973.6
922.60.52.868.9
1018.10.22.967.6
1111.30.02.949.4
123.50.03.040.0
201811.90.13.144.7
23.80.13.441.8
39.90.33.963.1
414.50.24.162.4
517.90.23.972.5
621.50.43.175.2
726.70.23.077.8
827.90.23.674.8
922.00.43.076.2
1016.50.22.061.8
1112.40.12.357.1
125.70.12.948.5
202016.40.23.056.3
27.10.13.055.2
310.40.13.557.0
412.60.13.753.0
517.90.13.172.3
622.40.43.073.5
722.11.12.787.2
827.00.54.082.4
922.00.33.673.9
1017.30.02.758.0
1112.40.12.852.0
124.40.02.942.7
Source: Average values are calculated based on data collected from the Korea Meteorological Administration Weather Data Service (data.kma.go.kr).
Table A2. Basic statistical analysis of the operation dataset pertaining to container port A.
Table A2. Basic statistical analysis of the operation dataset pertaining to container port A.
YearMonthAverage Number of Ships in BerthAverage Number of Containers UnloadedAverage Number of Containers LoadedAverage Number of Containers ImportedAverage Number of
Containers
Exported
Average Number of Trucks EnteringAverage Number of Trucks
Exiting
Average Number of Operation Container Cranes
Average Number of Operating Yard
Equipment
Average Number of Operating Yard Trucks
20171583807679666873239
2584828187727673438
3591838792768083440
4594898891767883442
5591888588747683441
6594898692757983442
7592908387727583443
8588848385737483340
9591908990797983342
10587817681667083341
1159810096968484103550
125102939091788093651
20181590948691767993548
2594908687757693548
3692918889787993648
4510210090917981103650
551009689937881103750
6610310092958183113753
761009593978386103652
86938986927782103647
96959486897678103650
1069994941008388103750
11510296961018588103951
12599898994788393749
20201580828388727583646
2585858083707283748
3592898490747993651
4592858487747693848
5587838184717383746
6591828190727883846
7589847888697793847
8588857481667193648
96898690887978103651
105948785937682103851
1161009489947983103954
125978893958284103953
Source: Busan Container Port A, Republic of Korea.

References

  1. Abdelfattah, M.; Elsayeh, M.; Abdelkader, S. A proposed port security risk assessment approach, with application to a hypothetical port. Aust. J. Marit. Ocean. Aff. 2021, 1–18. [Google Scholar] [CrossRef]
  2. Andritsos, F.; Mosconi, M. Port Security in EU: A systemic approach. In Proceedings of the 2010 International WaterSide Security Conference, Carrara, Italy, 3–5 November 2010; p. 11863527. [Google Scholar] [CrossRef]
  3. Orosz, M.D.; Southwell, C.; Barrett, A.; Chen, J.; Ioannou, P.; Abadi, A.; Maya, I. PortSec: A port security risk analysis and resource allocation system. In Proceedings of the 2010 IEEE International Conference on Technologies for Homeland Security, Boston, MA, USA, 8–10 November 2010; p. 11679258. [Google Scholar] [CrossRef]
  4. Budiyanto, M.A.; Fernanda, H. Risk Assessment of Work Accident in Container Terminals Using the Fault Tree Analysis Method. J. Mar. Sci. Eng. 2020, 8, 466. [Google Scholar] [CrossRef]
  5. Chlomoudis, C.I.; Pallis, P.L.; Tzannatos, E.S. A Risk Assessment Methodology in Container Terminals: The Case Study of the Port Container Terminal of Thessalonica, Greece. J. Traffic Transp. Eng. 2016, 4, 251–258. [Google Scholar]
  6. Hamka, M.A. Safety Risks Assessment on Container Terminal Using Hazard Identification and Risk Assessment and Fault Tree Analysis Methods. Procedia Eng. 2017, 194, 307–314. [Google Scholar] [CrossRef]
  7. Rosoff, H.; von Winterfeldt, D. A risk and economic analysis of dirty bomb attacks on the ports of Los Angeles and Long Beach. Risk Anal. 2007, 27, 533–546. [Google Scholar] [CrossRef]
  8. Concho, A.L.; Ramirez-Marquez, J.E. An evolutionary algorithm for port-of-entry security optimization considering sensor thresholds. Reliab. Eng. Syst. Saf. 2010, 95, 255–266. [Google Scholar] [CrossRef]
  9. Hashemi, R.R.; Le Blanc, L.A.; Rucks, C.T.; Shearry, A. A Neural Network for Transportation Safety Modeling. Expert Syst. Appl. 1995, 9, 247–256. [Google Scholar] [CrossRef]
  10. Doong, D.; Chen, S.; Chen, Y.; Tsai, C. Operational Probabilistic Forecasting of Coastal Freak Waves by Using an Artificial Neural Network. J. Mar. Sci. Eng. 2020, 8, 165. [Google Scholar] [CrossRef] [Green Version]
  11. Wang, Y.; Wang, L.; Jiang, J.; Wang, J.; Yang, Z. Modelling ship collision risk based on the statistical analysis of historical data: A case study in Hong Kong waters. Ocean Eng. 2020, 197, 106869. [Google Scholar] [CrossRef]
  12. Tang, L.; Tang, Y.; Zhang, K. Prediction of Grades of Ship Collision Accidents Based on Random Forests and Bayesian Networks. In Proceedings of the the 5th International Conference on Transportation Information and Safety, Liverpool, UK, 14–17 July 2019; pp. 1377–1381. [Google Scholar]
  13. Korea Port Logistics Association. Statistics and Cases of Cargo Handling and Accident Situation; Korea Port Logistics Association: Seoul, Korea, 2018. [Google Scholar]
  14. Dong, C.; Shao, C.; Li, J.; Xiong, Z. An Improved Deep Learning Model for Traffic Crash Prediction. J. Adv. Transp. 2018, 2018, 1–13. [Google Scholar] [CrossRef]
  15. Yu, B.; Wang, Y.T.; Yao, J.B.; Wang, J.Y. A comparison of the performance of ANN and SVM for the prediction of traffic accident duration. Neural Netw. World 2016, 26, 271–287. [Google Scholar] [CrossRef] [Green Version]
  16. Dogru, N.; Subasi, A. Traffic Accident Detection Using Random Forest Classifier. In Proceedings of the 2018 15th Learning and Technology Conference, Jeddah, Saudi Arabia, 25–26 February 2018; pp. 40–45. [Google Scholar]
  17. Lee, S.; Kim, J.H.; Park, J.; Oh, C.; Lee, G. Deep-Learning-Based Prediction of High-Risk Taxi Drivers Using Wellness Data. Int. J. Environ. Res. Public Health. 2020, 17, 9505. [Google Scholar] [CrossRef] [PubMed]
  18. Parsa, A.B.; Movahedi, A.; Taghipour, H.; Derrible, S.; Mohammadian, A. Toward safer highways, application of XGBoost and SHAP for real-time accident detection and feature analysis. Accid. Anal. Prev. 2020, 136, 105405. [Google Scholar] [CrossRef]
  19. Zhang, Z.; Yang, W.; Wushour, S. Traffic Accident Prediction Based on LSTM-GBRT Model. J. Contr. Sci. Eng. 2020, 2020, 1–10. [Google Scholar] [CrossRef]
  20. Lu, P.; Zheng, Z.; Ren, Y.; Zhou, X.; Keramati, A.; Tolliver, D.; Huang, Y. A Gradient Boosting Crash Prediction Approach for Highway-Rail Grade Crossing Crash Analysis. J. Adv. Transp. 2020, 2020, 1–10. [Google Scholar] [CrossRef]
  21. Cuenca, L.G.; Puertas, E.; Aliane, N.; Andres, J.F. Traffic Accidents Classification and Injury Severity Prediction. In Proceedings of the 2018 3rd International Conference on Intelligent Transportation Engineering, Singapore, 3–5 September 2018; pp. 52–57. [Google Scholar]
  22. AlMamlook, R.E.; Kwayu, K.M.; Alkasisbeh, M.R.; Frefer, A.A. Comparison of Machine Learning Algorithms for Predicting Traffic Accident Severity. In Proceedings of the 2019 IEEE Jordan International Joint Conference on Engineering and Information Technology (JEEIT), Amman, Jordan, 9–11 April 2019; pp. 272–276. [Google Scholar]
  23. Ung, S.T.; Williams, V.; Bonsall, S.; Wang, J. Test case based risk predictions using artificial neural network. J. Safety Res. 2006, 37, 245–260. [Google Scholar] [CrossRef] [PubMed]
  24. Park, Y. An analysis of Traffic Accidents in Terms of Human Factors: The Case of Bus-Drivers. Korean J. Ind. Organ. Psychol. 2000, 13, 75–90. [Google Scholar]
  25. Choi, J.; Kim, S.; Hwang, K.; Baik, S. Severity Analysis of the Pedestrian Crash Patterns Based on the Ordered Logit Model. Int. J. Highw. Eng. 2009, 11, 153–164. [Google Scholar]
  26. Choi, J.; Kim, S.; Kim, S.; Yeon, J.; Kim, C. A Study on Pedestrian Crashes Contributing Factors During Jaywalking—Focused on the case of Seoul. J. Korea Inst. Intell. Transp. Syst. 2015, 14, 38–49. [Google Scholar]
  27. Lee, H.; Kum, K.; Son, S. A Study on the factor analysis by grade for highway traffic accident. Int. J. Highw. Eng. 2011, 13, 157–165. [Google Scholar] [CrossRef]
  28. Kim, B.; Park, S.; Gong, J.; Yeo, G. A Study on the Safety Factor Analysis of Bulk Cargo Handling Using Fuzzy-AHP: Focused on steel cargo. J. Digit. Converg. 2018, 16, 179–188. [Google Scholar]
  29. Kim, D. A risk analysis of accidents for improving port logistics productivity—A case study of a container operator of a port. Product. Rev. 2016, 30, 53–79. [Google Scholar]
  30. Yeun, D.; Choi, Y.; Kim, S. An Assessment & Analysis of Risk Based on Accident Category for Container Terminals. J. Shipp. Logist. 2014, 30, 843–858. [Google Scholar]
  31. Cha, S.; Noh, C. The Accidents Analysis for Safety Training in the Container Terminal. J. Korean Navig. Port Res. 2016, 40, 197–205. [Google Scholar] [CrossRef] [Green Version]
  32. Singh, G.; Pal, M.; Yadav, Y.; Singla, T. Deep neural network-based predictive modeling of road accidents. Neural Comput. Appl. 2020, 32, 12417–12426. [Google Scholar] [CrossRef]
  33. Keras: The Python Deep Learning API. Available online: http://keras.io (accessed on 3 March 2021).
  34. Breiman, L. Random Forests. Mach Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  35. Scikit-learn: Machine Learning in Python. Available online: http://scikit-learn.org (accessed on 3 March 2021).
  36. Ma, X.; Ding, C.; Luan, S.; Wang, Y.; Wang, Y. Prioritizing Influential Factors for Freeway Incident Clearance Time Prediction Using the Gradient Boosting Decision Trees Method. IEEE Trans. Intell. Transp. Syst. 2017, 18, 2303–2310. [Google Scholar] [CrossRef]
  37. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  38. Parsa, A.B.; Taghipour, H.; Derrible, S.; Mohammadian, A. Real-time accident detection: Coping with imbalanced data. Accid. Anal. Prev. 2019, 129, 202–210. [Google Scholar] [CrossRef]
  39. Li, P.; Abdel-Aty, M.; Yuan, J. Real-time crash risk prediction on arterials based on LSTM-CNN. Accid. Anal. Prev. 2020, 135. [Google Scholar] [CrossRef] [PubMed]
  40. Bobbili, N.P.; Cretu, A. Adaptive Weighting with SMOTE for Learning from Imbalanced Datasets: A Case Study for Traffic Offense Prediction. In Proceedings of the 2018 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA), Ottawa, ON, Canada, 12–13 June 2018. [Google Scholar] [CrossRef]
  41. Open MET Data Portal. Korea Meteorological Administration (KMA) Weather Data Service. Available online: data.kma.go.kr (accessed on 15 January 2021).
  42. Reichenbach, A.; Navaro-Barrientos, J.E. A model for traffic incident prediction using emergency braking data. arXiv 2021, arXiv:2102.06674. [Google Scholar]
  43. Varghese, V.; Chikaraishi, M.; Urata, J. Deep Learning in Transport Studies: A Meta-analysis on the Prediction Accuracy. J. Big Data Anal. Transp. 2020, 2, 199–220. [Google Scholar] [CrossRef]
  44. Zhang, Z.; He, Q.; Gao, J.; Ni, M. A deep learning approach for detecting traffic accidents from social media data. Transp. Res. Part C Emerg. Technol. 2018, 86, 580–596. [Google Scholar] [CrossRef] [Green Version]
Table 1. Statistics of occupational accidents and economic loss associated with container ports in Korea (2015–2019).
Table 1. Statistics of occupational accidents and economic loss associated with container ports in Korea (2015–2019).
20152016201720182019
AccidentsFatality41153
Injury
Total
91
94
79
80
96
97
94
99
96
99
Economic loss
(billion KRW)
18.817.220.820.5NA
Source: Korea Port Logistics Association.
Table 2. Basic statistical analysis of the accidents that occurred in container port A.
Table 2. Basic statistical analysis of the accidents that occurred in container port A.
YearMonthNumber of Accidents YearMonthNumber of Accidents YearMonthNumber of Accidents
201711201814202017
212122
313431
424247
535454
626164
727477
838488
9293910
103102107
113115119
1231251212
Total26 Total39 Total78
Source: Busan Container Port A, Republic of Korea.
Table 3. Values of hyperparameters for the deep neural network model.
Table 3. Values of hyperparameters for the deep neural network model.
HyperparameterValue or Method
OptimizerAdam
LossBinary cross-entropy
CallbacksEarly stopping
Batch size128
Epochs100
MonitorVal loss
Table 4. Values of hyperparameters for the random forest model.
Table 4. Values of hyperparameters for the random forest model.
HyperparameterDefinitionValues
Max depthMaximum depth of decision trees5
Min samples leafMinimum samples in a leaf node18
Min samples splitMinimum samples in a node to be considered for splitting12
N-estimatorNumber of decision trees in the random forest100
Table 5. Values of hyperparameters for the gradient boosting model.
Table 5. Values of hyperparameters for the gradient boosting model.
HyperparameterDefinitionValues
Learning rateImpact of each tree on the final outcome1.0
Max depthMaximum depth of decision trees5
Max featuresNumber of features to consider while searching for the best split0.25
Min samples leafMinimum samples required in a leaf node18
Min samples splitMinimum samples to be considered for splitting12
N-estimatorNumber of sequential trees to be modeled100
SubsampleFraction of observations to be selected for each tree0.90
Table 6. Results of cross-validation for deep neural network model (accuracy).
Table 6. Results of cross-validation for deep neural network model (accuracy).
1 h3 h6 h12 h24 h
10.980.940.940.930.84
20.990.940.930.910.72
30.980.920.920.860.90
40.970.960.910.860.69
50.980.940.880.830.69
61.001.000.990.970.91
71.001.000.940.960.95
81.000.990.990.990.99
91.001.000.990.980.94
101.001.000.980.980.97
Average0.990.970.950.930.86
Table 7. Results of cross-validation for the random forest model (accuracy).
Table 7. Results of cross-validation for the random forest model (accuracy).
1 h3 h6 h12 h24 h
10.960.970.940.900.90
20.920.930.930.930.93
30.930.950.920.920.93
40.920.940.940.980.96
50.930.930.950.940.96
60.970.960.960.940.93
70.910.890.910.920.92
80.930.910.930.940.95
90.930.940.920.930.95
100.930.960.950.940.93
Average0.930.940.940.930.93
Table 8. Results of cross-validation for the gradient boosting model (accuracy).
Table 8. Results of cross-validation for the gradient boosting model (accuracy).
1 h3 h6 h12 h24 h
10.990.970.960.940.86
20.911.001.000.980.98
31.001.000.990.970.97
40.891.001.000.990.96
51.001.000.990.970.96
60.961.000.990.980.95
71.001.000.990.970.89
80.960.990.990.950.92
90.841.000.980.980.92
101.000.990.980.940.83
Average0.960.990.990.970.92
Table 9. Testing results for all models.
Table 9. Testing results for all models.
ModelTime IntervalsAccuracy (%)Precision (%)Recall (%)F1 Score (%)
Deep neural
network
1 h98.22.32.62.4
3 h93.64.26.55.1
6 h90.97.46.77.0
12 h81.410.011.110.5
24 h76.517.721.919.6
Random
forest
1 h90.41.111.52.1
3 h91.32.76.53.8
6 h86.94.78.05.9
12 h83.611.39.710.4
24 h76.818.29.412.4
Gradient
boosting
1 h98.31.41.31.4
3 h90.03.911.75.8
6 h85.18.720.012.1
12 h76.613.325.017.4
24 h62.320.239.126.6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, J.H.; Kim, J.; Lee, G.; Park, J. Machine Learning-Based Models for Accident Prediction at a Korean Container Port. Sustainability 2021, 13, 9137. https://doi.org/10.3390/su13169137

AMA Style

Kim JH, Kim J, Lee G, Park J. Machine Learning-Based Models for Accident Prediction at a Korean Container Port. Sustainability. 2021; 13(16):9137. https://doi.org/10.3390/su13169137

Chicago/Turabian Style

Kim, Jae Hun, Juyeon Kim, Gunwoo Lee, and Juneyoung Park. 2021. "Machine Learning-Based Models for Accident Prediction at a Korean Container Port" Sustainability 13, no. 16: 9137. https://doi.org/10.3390/su13169137

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop