Next Article in Journal
Special Issue on the Advances in Fluid Mechanics
Next Article in Special Issue
Characteristics and Causes of Coastal Water Chemistry in Qionghai City, China
Previous Article in Journal
The Effect of Cold Plasma on Selected Parameters of Bovine Colostrum
Previous Article in Special Issue
Study on Sedimentary Evolution of the Hanjiang River Delta during the Late Quaternary
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Pilot Study of Stacked Autoencoders for Ship Mode Classification

1
Romantique, Contents AI Research Center, 27 Daeyeong-ro, Busan 49227, Republic of Korea
2
Division of Marine System Engineering, Korea Maritime and Ocean University, 727 Taejong-ro, Yeongdo-gu, Busan 49112, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(9), 5491; https://doi.org/10.3390/app13095491
Submission received: 20 February 2023 / Revised: 17 April 2023 / Accepted: 26 April 2023 / Published: 28 April 2023
(This article belongs to the Special Issue Advances in Applied Marine Sciences and Engineering)

Abstract

:
With the evolution of the shipping market, artificial intelligence research using ship data is being actively conducted. Smart ships and reducing ship greenhouse gas emissions are among the most actively researched topics in the maritime transport industry. Owing to the massive advances in information and communications technology, the internet of things, and big data technologies, smart ships have emerged as a very promising proposition. Numerous methodologies and network architectures can smoothly collect data from ships that are currently in operation, as is currently done in research on reducing ship fuel consumption by deep learning or conventional methods. Many extensive studies of stacked autoencoders have been carried out in the past few years. However, prior studies have not addressed the development of algorithms or deep learning-based models to classify the operating states of ships. In this paper, we propose for the first time a deep learning-based stacked autoencoder model that can classify the operating state of a ship broadly into the categories of At Sea, Stand By, and In Port, using actual ship power load data. In order to maximize the model’s performance, the stacked autoencoder architecture, number of hidden layers, and number of neurons contained in each layer were measured by performance metrics such as true positive rate, false positive rate, Matthews correlation coefficient, and accuracy. It was found that the model’s performance was not always improved by increasing its complexity, so the feasibility of developing and utilizing an efficient model was verified by comparing it to real data. The best-performing model had a (5–128) structure with latent layer size 9. It achieved a true positive rate of 0.9035, a false positive rate of 0.0541, a Matthews correlation coefficient of 0.9054, and an accuracy of 0.9612, clearly demonstrating that deep learning can be used to analyze ship operating modes.

1. Introduction

In the maritime transport industry, smart ships and reducing ship greenhouse gas emissions have been actively researched [1]. Smart ships have emerged as information and communications technology, the internet of things, and big data technologies have advanced. Different from a conventional ship, a smart ship is characterized by its ability to use data collected by sensors installed within the ship to self-navigate or to provide appropriate information to assist in the decisions of crew members operating the ship [2]. Studies on reducing ship greenhouse gas emissions have mainly focused on reducing ship fuel consumption and eco-friendly ships that do not use oil.
Research on smart ships has been actively pursued, and various methodologies and network architectures that can smoothly collect data from ships in operation have been proposed. This research has included areas such as big data collection systems [3], cyber security considerations [4], data management to reduce learning costs [5], framework structures for index systems [6], surveys of architectures and applications [7], and priority items for smart shipping [8]. Furthermore, various data on actual ships is being collected, which is used in research on reducing ship fuel consumption by deep learning or conventional methods to perform knowledge-free path planning with reinforcement learning [9], energy-saving auto path planning algorithms [10], energy-saving management systems for smart ships [11], power scheduling for saving energy with reinforcement learning [12], and forecasting ship fuel consumption [13]. However, the application of ship operation mode classification remains unresearched.
Due to the characteristics of ship operations, the operating state of a ship can be broadly classified as At Sea, Stand By, or In Port. In the At Sea state, all of the devices on the ship are connected and powered, and the load changes in each device are small. Stand By refers to the state in which the ship is entering or exiting a port, and it is characterized by large fluctuations in the total power consumption due to changes in the ship speed and the use of auxiliary devices. Lastly, In Port refers to the operating state in which a ship has entered a port and cargo is being loaded or unloaded from the ship. In this state, fewer auxiliary devices are being powered on the ship, and total power consumption and power fluctuations are low. The power fluctuation characteristics of each ship type are as follows:
  • Container Ship: During the In Port state, the total power consumption and power fluctuations are large if many reefer containers are being carried.
  • LNG Ship: Cargo pumps are used when loading or unloading crude oil. Therefore, the ship is under the highest power load during the In Port state.
  • Bulk Ship: Bulk ships with cargo cranes installed are under a very large power load during the In Port state.
The operating states of ships, which are presently being acquired in large quantities, have not, however, been adequately researched. As such, researchers are faced with the problem of needing to label data manually based on ship voyage information to use the collected ship data. Research on ship state classification models that can classify the operating states of ships is required to overcome this issue.
Prior studies have not addressed the development of algorithms or deep learning-based models to classify the operating states of ships. Autoencoders are used in handwriting recognition [14], anomaly detection [15], fault diagnosis [16], and fraud diagnosis [17] and produce better results in comparison with existing algorithms. Thus, we selected the autoencoder model for classifying ship data.
However, the performance of the stacked autoencoder model depends on proper control of the components. When utilizing an autoencoder-based classification model, model design considerations include:
  • Which structure has better performance?
  • What are the appropriate values for the number of hidden layers and the number of neurons in each layer to achieve better performance?
  • What size of latent layer is suitable? The size of the latent layer has a significant effect on the performance of the classification model.
In order to address these issues, we conducted comparative experiments on the structure of the classification model, the appropriate values for the number of hidden layers and the number of neurons in each layer, and changes in the size of latent layers. We find the best model to classify the ship’s operating state as either At Sea, Stand By, or In Port using actual ship power load data.
This paper makes the following contributions: First, the structure of the first stacked autoencoder model using actual ship data is presented. Second, the performance change of the model according to the components of the stacked autoencoder was investigated. In particular, since there is no previous study that uses actual ship data, we design and perform performance comparison experiments of classification models according to structural changes of the stacked autoencoder. Third, the experimental results are analyzed.

2. Related Works

In the past several years, many studies have applied autoencoder models to solve practical problems. An autoencoder [18] is a learning neural network model that approximates input and output values to reconstruct them the same, and the main purpose of the autoencoder is to learn informative representations of data using an unsupervised method. Types of autoencoders include the stacked autoencoder, sparse autoencoder, denoising autoencoder, contractive autoencoder, and variational autoencoder [19].
The stacked autoencoder can learn efficiently to create robust features from training data. Research has been conducted on the benefits of stacked autoencoders to solve their problems. Ghosh et al. [20] used a stacked autoencoder model to classify human emotional data and achieved good results in categorizing human emotions with a spectrogram dataset. Ghosh’s approach was able to produce better results compared to traditional methods, which could not distinguish between happy and angry people.
Ambaw et al. [21] compared different conventional neural networks, support vector machines, and stacked autoencoders on the recognition of continuous phase frequency-shift keying under carrier frequency offset, noisy, and fast-fading channel conditions with the based model. In this study, the three features selected for recognition were the approximate entropy of the received signal, the approximate entropy of the received signal’s phase, and the approximate entropy of the instantaneous frequency of the received signal. It was found that the stacked autoencoder performed better than support vector machines and traditional neural networks; a stacked autoencoder model can give a better accuracy result for most signal to noise ratios.
Singh et al. [22] proposed a stacked autoencoder model to reduce complexity and processing time for detecting epilepsy. The model classified epileptic data into normal, ictal, and preictal. He selected machine learning algorithms such as Bayes Net, Naïve Bayes, multilayer perceptron, radial basis function neural networks, and the C4.5 decision tree classifier as comparison models. He proved that the stacked autoencoder model had the best performance score with the least processing time.
Law et al. [23] suggested using a cascade of two types of networks, stacked autoencoders and extreme learning machines, for multi-label classification to enhance a stacked autoencoder’s performance. The proposed model was compared with eleven other algorithms with seven datasets. However, she claimed that the model had promising performance.
Aouedi et al. [24] introduced a stacked sparse autoencoder model to integrate feature extraction and classification processes. The model uses denoising and a dropout technique to enhance feature extraction performance and prevent overfitting. It was proven that the model produced a better output than conventional models.
Deperlioglu [25] built a stacked autoencoder classification model for heart sound analysis. Traditional methods use data transforms to get the S1 and S2 segments of heart sounds. Deperlioglu’s novel approach was to utilize only a stacked autoencoder model to get segments of heart sounds for direct classification. This model was compared with conventional algorithms. The proposed model’s performance was similar to prior models. According to this study, a stacked autoencoder can be used in the medical field with efficient and effective classification results.
Gokhale et al. [26] compared a proposed stacked autoencoder model with seven previously established algorithms using ten datasets to find the key genes for cancer. Traditional gene selection approaches using statistical or feature selection methods have accuracy problems. However, the proposed stacked autoencoder-based framework outperformed conventional methods in this study.
Arafa et al. [27] introduced a reduced noise-autoencoder for solving the problem of imbalanced data in genomic datasets. Arafa’s approach was able to solve the dimensionality problem with the stacked autoencoder with feature reduction and create new low-dimensional data. In addition, the accuracy score was improved by more than eight percent.

3. Theoretical Background

3.1. Autoencoder

An autoencoder consists of an encoder, a latent layer, and a decoder, as shown in Figure 1. The encoder is called a recognition network and extracts the features of the original data entered as input. The layer that stores the extracted features during this task is called the latent layer. The decoder is known as a generative network, and it converts the features into output. Through this process, the autoencoder can reorganize the core information in the input data.
The encoder of the autoencoder can be defined as
h = σ ( W e x + b e )
Here, x is the input data, and W e and b e are the e th weight value matrix and bias vectors, respectively. σ is the activation function. h is the encoder output. The decoder of the autoencoder can be expressed as
x ^ = σ ( W d h + b d )
In the decoder, the encoder output h is used as an input variable. W d and b d are the d th weight value matrix and bias vectors, respectively. σ is the activation function. x ^ is the decoder output.
The autoencoder learns in order to make the decoder’s output value as similar to the input value as possible. Therefore, minimizing the difference between the input value and the decoder output value by adjusting the parameters during the training of the autoencoder model is important. As such, selecting a loss function that is suitable for the goals of the appropriate model is also critical. If the mean square error is used as the loss function, it can be expressed as follows:
L ( x , x ^ ) = 1 N i = 1 N ( x ^ x ) 2
Here, x is the input value, and x ^ is the decoder output value. N is the N th term.

3.2. SAE

An SAE [28], also known as a deep autoencoder, is organized as shown in Figure 2. It has a structure in which multiple hidden layers are contained in the encoder and decoder, and its structure is symmetrical in relation to the latent layer. The latent layer is located between the encoder and the decoder, as in an autoencoder, and it stores the feature data that are acquired from the encoder.
The SAE model is trained using the greedy layer-wise training methodology [29]. This methodology was proposed to determine the optimal parameters of an SAE, and it has been proven to be effective in learning an SAE with multiple hidden layers. This methodology can also reduce the network size of the SAE model and increase the training speed. Furthermore, it has the advantage of potentially reducing the risk of overfitting.

3.3. Dataset

Real ship data are not open-access data. Thus, we collected data from a real ship. The data in this study were those of a 13,000 TEU container ship used in actual operations. It is propelled by one MAN-Burmeister and Wain diesel engine and has four 3480 kW generators. Table 1 lists the specifications of the target ship.
The power load of the ship continuously fluctuates according to the operating state of the ship [30]. Furthermore, the steering equipment installed on the ship is powered by an electric motor, and the power load consumed by the electric motor is affected by the external resistance of the ship’s hull [31]. Therefore, this study collected data on the electric load, which is the total electric load of the ship, as well as data that indicate the external resistance of the ship, such as its heading, rudder angle, water depth, water speed, wind angle, wind speed, and ship speed. Table 2 shows the types of data measured in this study.
The data were measured in 10 min intervals, and a total of 30,340 data values were collected. Figure 3 shows the collected ship power load data. Changes in the power load occurred as the ship was operated.
Table 3 shows the number of data points according to the ship’s state. The ship is most often operated At Sea, and the least amount of time is spent in Stand By. In addition, by the number of data points measured, there are twice as many instances of the In Port state compared to the Stand By state.

4. Approach

This section describes the design approach for the SAE that classifies ship operating modes. In particular, we hoped to determine whether the structure of the SAE model and the size of its latent layer affected the operating mode classification performance.

4.1. Overview

An overview of the process of this research is given below.
  • Training: The stacked autoencoder is sufficient to extract features and reconstruct. However, there is a need to find the structure of the model that is the most efficient. The stacked autoencoder models are trained to classify the features of input data, while the Adam optimizer is utilized as the loss function. In the training process, five structures, which are composed of depth and size, were trained by latent layer values (3, 6, 9, and 12).
  • Selection of models for comparison: After completing the training process, the true positive rate (TPR), false positive rate (FPR), accuracy, and Matthews correlation coefficient (MCC) were used to select the best performance model for each structure. Thus, only five models were kept for a final comparison. Trained models classified ships as being in one of three states: In Port, Stand By, and At Sea. In this process, the MCC was considered to be the most important of all the evaluation metrics.
  • Performance comparison: Five models from the previous step of the process were compared. For comparison, we used the same evaluation metrics and a confusion matrix. The confusion matrix is a tabular summary for users to check precisely the correct and incorrect results made by classification models. By using this technique, the performance of the classification models was analyzed in detail.

4.2. Model Design

The SAE described in Section 2 was used in the operating mode classification model. To improve the performance of the SAE model, this study considered two aspects: the structure of the model and the size of the latent layer. Here, the model structure refers to the number of hidden layers within the encoder and decoder and the number of neurons used in each hidden layer. The size of the latent layer refers to the number of neurons in the latent layer. Figure 4 shows the basic structure of the model. The encoder and decoder were arranged in a symmetric form, and a softmax layer [32] was added to the output part of the decoder for classification.
The softmax layer uses a softmax activation function, which is employed in a deep learning-based model, to produce a classification model that classifies data into three or more classes [33]. If the input vector is z , the softmax activation function can be expressed as follows:
s o f t m a x ( z ) = exp ( a i ) j = i k exp ( a j )
Here, k is the number of classes that should be output by the multi-class classifier, and exp ( a i ) is the standard exponential function of the i th input vector. Lastly, exp ( a j ) is the standard exponential function of the j th output vector.

4.3. Evaluation Metrics

This study selected accuracy, TPR, FPR, and MCC as the evaluation metrics to compare the performance of the models. The accuracy is the ratio of the number of data points that the multi-class classification model correctly classifies to the overall number of data points. TPR is the level at which the actual correct answer is clearly predicted. FPR is used to evaluate multi-class classification performance. Lastly, MCC has the advantage of being able to express the confusion matrix of a multi-class classification model in a balanced way; it is also a balanced evaluation metric that is good for representing the performance of such models [34,35]. MCC was considered to be the most important of all the evaluation metrics because this study presents research on a model that performs multi-class classification. The evaluation metrics can be expressed as shown in the following equations:
True   Positive   Rate   ( TPR ) = T P T P + F N
False   Positive   Rate   ( FPR ) = F P F P + T N
A c c u r a c y = T P + T N T P + F P + T N + F N
MCC = c × s k K p k × t k ( s 2 k K p k 2 ) × ( s 2 k K t k 2 )
Here, T P is the number of true positives, T N is the number of true negatives, F P is the number of false positives, and F N is the number of false negatives. c is the number of samples correctly predicted out of the number of samples, and s is the number of samples.   K is the total number of classes, and k is an individual class (a class from 1 to K). p k is the number of times class k was predicted, and t k is the number of times class   k truly occurred.

4.4. Composition of Models for Comparison Experiments

We performed experiments on SAE models that had various structures and latent layer sizes to find the best SAE model. The hidden layers of the model were all fully connected, and a rectified linear unit (ReLU) activation function was used. A total of five model structures were employed in the comparison experiments. We used (depth, size) to express the model structures and depict the composition of the models. Here, depth refers to the number of hidden layers in the encoder and decoder, including the latent layer. Size refers to the number of neurons in the first hidden layer of the encoder. The encoder had a structure in which the number of hidden layer neurons decreased by half compared to that in the previous layer. Additionally, the decoder structure was symmetrical with the encoder structure. Table 4 shows the structures of the models used in the comparison experiments.
Here, to find the exact size of the latent layer, we performed experiments that changed the latent layer values to 3, 6, 9, and 12. In addition, the collected dataset was divided into training, testing, and validation datasets to prevent the overfitting problem when training the classification model. Further, 60% of the entire dataset was used as the training dataset, and 20% was used as the testing dataset. The last 20% of the dataset was utilized as the validation dataset. Furthermore, an Adam optimizer [36] was employed as the optimization function in the training of all models, and a value of 1 × 10 4 was used as the learning rate. The evaluation experiments for comparison were performed with the Python, Scikit-Learn, and TensorFlow libraries.

5. Experimental Results and Discussion

5.1. Experimental Results

This section assesses the performance of the models by comparing the model structures used in the comparison experiments. For this analysis, 200 epochs of training were performed on all models, and the batch size was 64. Additionally, the evaluation metrics were used to evaluate the performance of each model structure, and the latent layer size suitable for each model structure was found. Comparative evaluations were also performed on the models that showed the best performance for each model structure.
Table 5 lists the evaluation results for the (5–32) structure. The accuracy performance is stable regardless of the latent layer size. However, the FPR and MCC values are affected by the latent layer size. A latent layer size of 9 produces the best performance in the (5–32) structure.
Table 6 shows the evaluation results for the (5–64) structure. Different from the evaluation results for the (5–32) structure, there is a significant difference in accuracy. The performance rapidly worsens when the latent layer size is 6. However, the FPR’s performance is exceptional. The performance comparison results obtained using MCC indicate that the best performance occurs when the latent layer size is 12.
Table 7 illustrates the evaluation results for the (5–128) structure. The TPR performance is best when the latent layer size is 12, and the FPR performance is best when the latent layer size is 6. However, the MCC and accuracy performance are best when the latent layer size is 9.
Table 8 lists the evaluation results for the (7–64) structure. The TPR performance is the best when the latent layer size is 3, and the FPR performance is the best when the latent layer size is 9. However, the MCC and accuracy performance are at their best when the latent layer size is 12.
Table 9 shows the evaluation results for the (7–128) structure. The TPR performance is the best when the latent layer size is 9, and the FPR, MCC, and accuracy performance are the best when the latent layer size is 6.
Table 10 illustrates the best model structures based on MCC scores. In the comparison results, the MCC evaluation metric confirms that the (5–128) structure is the best at classifying the ship operating mode when the latent layer size is 9. Furthermore, Figure 5 shows the MCC value according to the latent layer size of the models used in the comparison experiments.
Table 11, Table 12, Table 13, Table 14 and Table 15 exhibit the confusion matrices of the best model structures based on MCC scores. The rows of the tables present the classification results, and the columns show the actual classification classes. The diagonal cells of the matrices show the numbers of successful classification results, and the off-diagonal cells show the numbers of misclassified results.
Table 11 presents the confusion matrix of the (5–32) structure with a latent layer size of 9. It can be seen that the classification performance for the Stand By state is the best among the confusion matrix entries compared.
Table 12 shows the confusion matrix of the (5–64) structure with a latent layer size of 12. The classification performance for the At Sea state is the best among the confusion matrix entries compared.
Table 13 provides the confusion matrix of the (5–128) structure with a latent layer size of 9. The classification performance for the In Port and Stand By states is relatively high among the confusion matrix entries compared, whereas the classification performance for the At Sea state is the lowest. However, it can also be seen that the classification model performance for the At Sea state is not especially low, with only 20 misclassified data points.
Table 14 shows the confusion matrix of the (7–64) structure with a latent layer size of 12. The classification performance for the In Port state is the best among the confusion matrix entries compared. Conversely, the performance for the Stand By state is the lowest.
Table 15 presents the confusion matrix of the (7–128) structure with a latent layer size of 6. This model has the lowest average performance out of the confusion matrix entries compared.

5.2. Discussion

Performance by Model Structure and Latent Size

This section discusses the importance of the findings of this study. First, we focused on model structure, which was found to affect the classification performance: the (5–128) structure gave the best results. This less complex model structure (5–128) had a better performance result than more complex model structures (7–64, 7–128).
Second, we investigated the effect of latent size on model performance. We found that the performance of the model increased with latent size up to a certain point and declined after that point. Hence, finding the appropriate latent size can improve performance even when a model’s structure is fixed.
Through this study, several limitations of the variational autoencoder (VAE) model were identified. First, the proposed VAE model utilizes only container vessel data. These data may differ from those obtained from other types of ships, such as LNG ships and bulk ships. This pilot study focused on performing comparative experiments using confusion matrices and evaluation metrics to evaluate the ship state classification performance according to changes in the parameters of VAE models using actual ship data. Therefore, the direct discussion that is possible based on the current results is limited. In order to improve the performance of the ship state classification model in the future, a comparative study with deep learning-based classification models using various structures that are currently used in other fields is needed.
Second, we found that the classification performance for the Stand By state was low. The model is affected by the balance of data, and the quantity of data for the Stand By state was very small in this study. This class imbalance problem can be solved in three ways: data-level, algorithm-level, and hybrid methods [37]. Data-level methods apply various data sampling methods that try to create balanced, distributed data for the training dataset. The cost-sensitive approaches focus on diminishing the bias toward major groups [38]. Hybrid methods combine the benefits of the previous two types of methods and minimize their weaknesses to improve classification model performance [39]. By adopting these methods, the problem of imbalanced data can be solved.
In the commercial realm, no algorithm has been presented that can automatically classify current ship states. However, as research on smart ships, reducing ship fuel consumption, and eco-friendly algorithms is conducted, the VAE model proposed in this pilot study can provide ship state-applied ship data to smart ship researchers. In addition, research can be conducted on reducing ship fuel consumption and improving eco-friendliness using the characteristics of the ship’s state.

6. Conclusions

Artificial intelligence research using ship data is being actively conducted as the shipping market evolves. However, studies have not been performed on classifying ship operating modes, despite previous studies on using ship data to predict power loads. An SAE has the advantage of being able to perform effective classification by analyzing the features of the input data. Therefore, we conducted a pilot study on deep learning models that can classify ship operating modes using an SAE. Furthermore, experiments were performed to compare the performance according to changes in the structure of the SAE and changes in the size of its latent layer. The key points to be verified through research are as follows:
  • TPR, FPR, accuracy, and MCC were selected as evaluation metrics to perform experiments that compared the performance according to changes in the structure of the SAE and the size of its latent layer. Even though previous studies have not been conducted in this area, performance comparison among the models was possible according to changes in the structure of the SAE and the size of its latent layer.
  • In the results of the SAE model comparison experiments, the (5–128) structure with a latent layer size of 9 showed the best operating mode classification performance.
  • The classification performance for the In Port and At Sea modes used in the experiments was generally excellent; however, the classification performance for the Stand By mode was low. Therefore, more data are needed to improve the prediction performance of the Stand By mode.
Through this pilot study, we found that our VAE-based deep learning model can be used to analyze ship operating modes. Furthermore, a model that can be used as a comparison target group for the classification of the actual operating state of the ship in the future was found. Based on this model structure, it will be possible to develop enhanced models. Although the classification performance for the Stand By mode was limited because of the imbalanced data, it was possible to propose a VAE model structure that can maximize the data classification performance of the In Port and At Sea modes. However, further research is required to address some limitations of this study. First, the issue of handling imbalanced data needs to be studied using real ship data; data-level techniques, algorithm-level methods, and hybrid methodologies can be utilized to find the most appropriate method to improve the classification model. Second, research to compare denoising, sparse, and stacking autoencoder models could be carried out to improve classification performance and establish which autoencoder-based model best interprets the features of real ship data.

Author Contributions

Conceptualization, methodology, and software, J.-Y.K.; project administration, funding acquisition, J.-S.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Korea Institute of Marine Science and Technology Promotion (KIMST), funded by the Korea Coast Guard (20190460).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Tang, Y.-l.; Shao, N.-n. Design and Research of Integrated Information Platform for Smart Ship. In Proceedings of the 2017 4th International Conference on Transportation Information and Safety (ICTIS), Banff, AB, Canada, 8–10 August 2017; pp. 37–41. Available online: https://ieeexplore.ieee.org/document/8047739 (accessed on 12 January 2023).
  2. Smart Shipping: Comprehensive Automation in the Maritime Sector. Available online: https://www.government.nl/topics/maritime-transport-and-seaports/smart-shipping-comprehensive-automation-in-the-maritime-sector (accessed on 12 January 2023).
  3. Zeng, X.; Chen, M. A novel big data collection system for ship energy efficiency monitoring and analysis based on BeiDou system. J. Adv. Transp. 2021, 2021, 9914720. [Google Scholar] [CrossRef]
  4. Reilly, G.; Jorgensen, J. Classification considerations for cyber safety and security in the smart ship era. In Proceedings of the International Smart Ships Technology Conference, London, UK, 26–27 January 2016; pp. 26–27. Available online: https://ww2.eagle.org/content/dam/eagle/Archived-Assets/leadership/articles-archives/ABS-RINA-Cyber-Safety-Security-Ship-Tech.pdf (accessed on 27 January 2023).
  5. Pérez Fernández, R.; Benayas Ayuso, A.; Pérez Arribas, F.L. Data Management for Smart Ship or How to Reduce Machine Learning Cost in IoS Applications. 2018. Available online: https://www.researchgate.net/publication/322635826_DATA_MANAGEMENT_FOR_SMART_SHIP_OR_HOW_TO_REDUCE_MACHINE_LEARNING_COST_IN_IoS_APPLICATIONS (accessed on 10 January 2023).
  6. Xiao, Y.; Chen, Z.; McNeil, L. Digital empowerment for shipping development: A framework for establishing a smart shipping index system. Marit. Pol. Manag. 2022, 49, 850–863. [Google Scholar] [CrossRef]
  7. Aslam, S.; Michaelides, M.P.; Herodotou, H. Internet of ships: A survey on architectures, emerging applications, and challenges. IEEE Internet Things J. 2020, 7, 9714–9727. [Google Scholar] [CrossRef]
  8. Ahn, Y.G.; Kim, T.; Kim, B.R.; Lee, M.K. A study on the development priority of smart shipping items—Focusing on the expert survey. Sustainability 2022, 14, 6892. [Google Scholar] [CrossRef]
  9. Chen, C.; Chen, X.Q.; Ma, F.; Zeng, X.J.; Wang, J. A knowledge-free path planning approach for smart ships based on reinforcement learning. Ocean Eng. 2019, 189, 106299. [Google Scholar] [CrossRef]
  10. Ding, Y.; Li, R.; Shen, H.; Li, J.; Cao, L. A novel energy-saving route planning algorithm for marine vehicles. Appl. Sci. 2022, 12, 5971. [Google Scholar] [CrossRef]
  11. Accetta, A.; Pucci, M. A First Approach for the Energy Management System in DC Micro-Grids with Integrated RES of Smart Ships. In Proceedings of the 2017 IEEE Energy Conversion Congress and Exposition (ECCE), Cincinnati, OH, USA, 1–5 October 2017; pp. 550–557. Available online: https://ieeexplore.ieee.org/document/8095831 (accessed on 19 February 2023).
  12. Hasanvand, S.; Rafiei, M.; Gheisarnejad, M.; Khooban, M.H. Reliable power scheduling of an emission-free ship: Multiobjective deep reinforcement learning. IEEE Trans. Transp. Electr. 2020, 6, 832–843. [Google Scholar] [CrossRef]
  13. Kim, J.Y.; Lee, J.H.; Oh, J.H.; Oh, J.S. A comparative study on energy consumption forecast methods for electric propulsion ship. J. Mar. Sci. Eng. 2022, 10, 32. [Google Scholar] [CrossRef]
  14. Almotiri, J.; Elleithy, K.; Elleithy, A. Comparison of Autoencoder and Principal Component Analysis Followed by Neural Network for E-learning Using Handwritten Recognition. In Proceedings of the IEEE Long Island Systems, Applications and Technology Conference, Farmingdale, NY, USA, 5 May 2017; Available online: https://ieeexplore.ieee.org/document/8001963 (accessed on 19 February 2023).
  15. Yao, R.; Liu, C.; Zhang, L.; Peng, P. Unsupervised Anomaly Detection Using Variational Auto-Encoder based Feature Extraction. In Proceedings of the International Conference on Prognostics and Health Management, San Francisco, CA, USA, 17 June 2019; Available online: https://ieeexplore.ieee.org/abstract/document/8819434 (accessed on 10 February 2023).
  16. Ma, S.; Chen, M.; Wu, J.; Wang, Y.; Jia, B.; Jiang, Y. High-voltage circuit breaker fault diagnosis using a hybrid feature transformation approach based on random forest and stacked autoencoder. IEEE Trans. Ind. Electron. 2018, 66, 9777–9788. [Google Scholar] [CrossRef]
  17. Thirukovalluru, R.; Dixit, S.; Sevakula, R.K.; Verma, N.K.; Salour, A. Generating Feature Sets for Fault Diagnosis using Denoising Stacked Auto-encoder. In Proceedings of the International Conference on Prognostics and Health Management, Ottawa, ON, Canada, 20 June 2016; Available online: https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=7542865 (accessed on 12 February 2023).
  18. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef]
  19. Bank, D.; Koenigstein, N.; Giryes, R. APR, 2021. Autoencoders. Available online: https://arxiv.org/pdf/2003.05991.pdf (accessed on 12 February 2023).
  20. Ghosh, S.; Laksana, E.; Morency, L.P.; Scherer, S. Representation Learning for Speech Emotion Recognition. In Proceedings of the INTERSPEECH 2016, San Francisco, CA, USA, 8 September 2016; Available online: https://doi.org/10.21437/Interspeech.2016-692 (accessed on 19 February 2023).
  21. Ambaw, A.B.; Bari, M.; Doroslovački, M. A case for stacked autoencoder based order recognition of continuous-phase FSK. In Proceedings of the Annual Conference on Information Sciences and Systems (CISS), Baltimore, MD, USA, 22 March 2017; Available online: https://ieeexplore.ieee.org/xpl/conhome/7917217/proceeding (accessed on 19 February 2023).
  22. Singh, K.; Malhotra, J. Stacked Autoencoders Based Deep Learning Approach for Automatic Epileptic Seizure Detection. In Proceedings of the (ICSCCC), 2018 First International Conference on Secure Cyber Computing and Communication, Jalandhar, India, 15 December 2018; Available online: https://ieeexplore.ieee.org/document/8703357 (accessed on 2 February 2023).
  23. Law, A.; Ghosh, A. Multi-label classification using a cascade of stacked autoencoder and extreme learning machines. Neurocomputing 2019, 358, 222–234. [Google Scholar] [CrossRef]
  24. Aouedi, O.; Piamrat, K.; Bagadthey, D. A Semi-supervised Stacked Autoencoder Approach for Network Traffic Classification. In Proceedings of the International Conference on Network Protocols, Madrid, Spain, 13 October 2020; Available online: https://ieeexplore.ieee.org/xpl/conhome/9259328/proceeding (accessed on 30 January 2023).
  25. Deperlioglu, O. Heart sound classification with signal instant energy and stacked autoencoder network. Biomed. Signal Process. Control 2021, 64, 102211. [Google Scholar] [CrossRef]
  26. Gokhale, M.; Mohanty, S.; Ojh, A. A stacked autoencoder based gene selection and cancer classification framework. Biomed. Signal Process. Control 2022, 78, 103999. [Google Scholar] [CrossRef]
  27. Arafa, A.; El-Fishawy, N.; Badawy, M.; Radad, M. RN-Autoencoder: Reduced Noise Autoencoder for classifying imbalanced cancer genomic data. J. Biol. Eng. 2023, 17, 7. [Google Scholar] [CrossRef] [PubMed]
  28. Baldi, P. Autoencoders, unsupervised learning, and deep architectures. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, Irvine, CA, USA, June 2012; pp. 37–49. Available online: http://proceedings.mlr.press/v27/baldi12a/baldi12a.pdf (accessed on 6 February 2023).
  29. Bengio, Y.; Lamblin, P.; Popovici, D.; Larochelle, H. Greedy layer-wise training of deep networks. Adv. Neural. Inf. Process. Syst. 2006, 19. [Google Scholar]
  30. Kim, J.H.; Choi, J.E.; Choi, B.J.; Chung, S.H. Twisted rudder for reducing fuel-oil consumption. Int. J. Naval Archit. Ocean Eng. 2014, 6, 715–722. [Google Scholar] [CrossRef]
  31. Nguyen, D.H.; Le, M.D.; Ohtsu, K. Ship’s optimal autopilot with a multivariate auto-regressive exogenous model. IFAC Proc. Vol. 2000, 33, 277–282. [Google Scholar] [CrossRef]
  32. Nwankpa, C.; Ijomah, W.; Gachagan, A.; Marshall, S. Activation functions: Comparison of trends in practice and research for deep learning. arXiv 2018, arXiv:1811.03378. Available online: https://arxiv.org/pdf/1811.03378.pdf (accessed on 6 February 2023).
  33. Kagalkar, A.; Raghuram, S. Activation Functions: CORDIC Based Implementation of the Softmax Activation Function. In Proceedings of the International Symposium on VLSI Design and Test, Bhubaneswar, India, 23 July 2020; Available online: https://ieeexplore.ieee.org/abstract/document/9190498 (accessed on 30 January 2023).
  34. Gorodkin, J. Comparing two K-category assignments by a K-category correlation coefficient. Comp. Biol. Chem. 2004, 28, 367–374. [Google Scholar] [CrossRef]
  35. Jurman, G.; Riccadonna, S.; Furlanello, C. A Comparison of MCC and CEN Error Measures in Multi-Class Prediction. 2012. Available online: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0041882 (accessed on 20 January 2023).
  36. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. Available online: https://arxiv.org/abs/1412.6980 (accessed on 15 January 2023).
  37. Krawczyk, B. Learning from imbalanced data: Open challenges and future directions. Prog. Artif. Intell. 2016, 5, 221–232. [Google Scholar] [CrossRef]
  38. Zięba, M.; Tomczak, J.M. Boosted SVM with active learning strategy for imbalanced data. Soft Comput. 2015, 19, 3357–3368. [Google Scholar] [CrossRef]
  39. Woźniak, M. Hybrid classifiers. In Methods of Data, Knowledge, and Classifier Fusion; Springer: Berlin/Heidelberg, Germany, 2013; p. 519. [Google Scholar] [CrossRef]
Figure 1. Autoencoder.
Figure 1. Autoencoder.
Applsci 13 05491 g001
Figure 2. SAE.
Figure 2. SAE.
Applsci 13 05491 g002
Figure 3. Ship electric load.
Figure 3. Ship electric load.
Applsci 13 05491 g003
Figure 4. SAE model structure.
Figure 4. SAE model structure.
Applsci 13 05491 g004
Figure 5. MCC results of structures with latent layer sizes.
Figure 5. MCC results of structures with latent layer sizes.
Applsci 13 05491 g005
Table 1. Specifications of a container ship.
Table 1. Specifications of a container ship.
Ship TypeContainer Ship
Length365 m
Width48 m
Draft10.8 m
Engine Output79,106 BHP
Generator Output3480 kW × 4
Maximum Speed23.0 knots
TEU13,154 TEU
Table 2. Various types of data measured.
Table 2. Various types of data measured.
Kind of DataUnitRange
Electric LoadkW585–4377
Heading°0–360
Rudder Angle°−35–35
Water Depthm0–839
Water Speedm/s−4–6
Wind Angle°0–360
Wind Speedm/s0–47
Ship Speedknots0–19
Table 3. Number of data collected for each state of the ship.
Table 3. Number of data collected for each state of the ship.
State of the ShipNumber of Data
In port5324
Stand by2582
At sea22,434
Table 4. Structures of the models for comparison.
Table 4. Structures of the models for comparison.
Structure (Depth, Size)Neurons in Each Layer
(5, 32)32–16-size of the latent layer-16–32
(5, 64)64–32-size of the latent layer-32–64
(5, 128)128–64-size of the latent layer-64–128
(7, 64)64–32–16-size of the latent layer-16–32–64
(7, 128)128–64–32-size of the latent layer-32–64–128
Table 5. Evaluation results of the (5–32) structure.
Table 5. Evaluation results of the (5–32) structure.
Latent Layer SizeTPRFPRMCCAccuracy
30.86560.07420.87980.9510
60.78480.04990.88310.9522
90.95410.09740.88770.9541
120.83920.05970.88360.9525
Table 6. Evaluation results of the (5–64) structure.
Table 6. Evaluation results of the (5–64) structure.
Latent Layer SizeTPRFPRMCCAccuracy
30.91410.08670.86920.9469
60.49240.03550.84700.9372
90.95560.10940.87820.9503
120.90770.07410.88670.9538
Table 7. Evaluation results of the (5–128) structure.
Table 7. Evaluation results of the (5–128) structure.
Latent Layer SizeTPRFPRMCCAccuracy
30.85240.07060.89060.9551
60.87760.10890.88080.9512
90.90380.05410.90540.9612
120.90850.08590.88640.9536
Table 8. Evaluation results of the (7–64) structure.
Table 8. Evaluation results of the (7–64) structure.
Latent Layer SizeTPRFPRMCCAccuracy
30.94400.13560.85810.9423
60.84360.06760.89260.9559
90.82530.02190.89810.9581
120.84700.03190.90360.9604
Table 9. Evaluation results of the (7–128) structure.
Table 9. Evaluation results of the (7–128) structure.
Latent Layer SizeTPRFPRMCCAccuracy
30.93670.07920.88670.9538
60.87500.05180.89800.9583
90.93190.05850.89790.9583
120.89470.05560.89550.9573
Table 10. Best models from structures based on MCC score.
Table 10. Best models from structures based on MCC score.
StructureLatent Layer SizeTPRFPRMCCAccuracy
(5–32)90.95410.09740.88770.9541
(5–64)120.90770.07410.88670.9538
(5–128)90.90380.05410.90540.9612
(7–64)120.84700.03190.90360.9604
(7–128)60.87500.05180.89800.9583
Table 11. Confusion matrix of (5–32) structure with a latent layer size of 9.
Table 11. Confusion matrix of (5–32) structure with a latent layer size of 9.
In PortStand byAt Sea
In port991160
Stand by9733310
At sea01554466
Correct0.91080.66070.9977
Incorrect0.08910.33920.0022
Table 12. Confusion matrix of (5–64) structure with a latent layer size of 12.
Table 12. Confusion matrix of (5–64) structure with a latent layer size of 12.
In PortStand byAt Sea
In port1012310
Stand by763055
At sea01684471
Correct0.93010.60510.9988
Incorrect0.06980.39480.0011
Table 13. Confusion matrix of (5–128) structure with a latent layer size of 9.
Table 13. Confusion matrix of (5–128) structure with a latent layer size of 9.
In PortStand byAt Sea
In port1048350
Stand by4032920
At sea01404456
Correct0.96320.65270.9955
Incorrect0.03670.34720.0044
Table 14. Confusion matrix of (7–64) structure with a latent layer size of 12.
Table 14. Confusion matrix of (7–64) structure with a latent layer size of 12.
In PortStand byAt Sea
In port1060540
Stand by282997
At sea01514469
Correct0.97420.59320.9984
Incorrect0.02570.40670.0015
Table 15. Confusion matrix of (7–128) structure with a latent layer size of 6.
Table 15. Confusion matrix of (7–128) structure with a latent layer size of 6.
In PortStand byAt Sea
In port1043440
Stand by4530812
At sea01524464
Correct0.95860.61110.9973
Incorrect0.04130.38880.0026
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J.-Y.; Oh, J.-S. A Pilot Study of Stacked Autoencoders for Ship Mode Classification. Appl. Sci. 2023, 13, 5491. https://doi.org/10.3390/app13095491

AMA Style

Kim J-Y, Oh J-S. A Pilot Study of Stacked Autoencoders for Ship Mode Classification. Applied Sciences. 2023; 13(9):5491. https://doi.org/10.3390/app13095491

Chicago/Turabian Style

Kim, Ji-Yoon, and Jin-Seok Oh. 2023. "A Pilot Study of Stacked Autoencoders for Ship Mode Classification" Applied Sciences 13, no. 9: 5491. https://doi.org/10.3390/app13095491

APA Style

Kim, J. -Y., & Oh, J. -S. (2023). A Pilot Study of Stacked Autoencoders for Ship Mode Classification. Applied Sciences, 13(9), 5491. https://doi.org/10.3390/app13095491

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop