Next Article in Journal
Performance Measurement in R&D Projects: Relevance of Indicators Based on US and German Experts
Previous Article in Journal
Population Decline through Tourism Gentrification Caused by Accommodation in Kyoto City
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tab2vox: CNN-Based Multivariate Multilevel Demand Forecasting Framework by Tabular-To-Voxel Image Conversion

1
School of Industrial and Management Engineering, Korea University, Seoul 02841, Korea
2
Center for Defense Resource Management, Korea Institute for Defense Analyses, Seoul 02455, Korea
*
Authors to whom correspondence should be addressed.
Sustainability 2022, 14(18), 11745; https://doi.org/10.3390/su141811745
Submission received: 6 August 2022 / Revised: 13 September 2022 / Accepted: 15 September 2022 / Published: 19 September 2022

Abstract

:
Since demand is influenced by a wide variety of causes, it is necessary to decompose the explanatory variables into different levels, extract their relationships effectively, and reflect them in the forecast. In particular, this contextual information can be very useful in demand forecasting with large demand volatility or intermittent demand patterns. Convolutional neural networks (CNNs) have been successfully used in many fields where important information in data is represented by images. CNNs are powerful because they accept samples as images and use adjacent voxel sets to integrate multi-dimensional important information and learn important features. On the other hand, although the demand-forecasting model has been improved, the input data is still limited in its tabular form and is not suitable for CNN modeling. In this study, we propose a Tab2vox neural architecture search (NAS) model as a method to convert a high-dimensional tabular sample into a well-formed 3D voxel image and use it in a 3D CNN network. For each image representation, the 3D CNN forecasting model proposed from the Tab2vox framework showed superior performance, compared to the existing time series and machine learning techniques using tabular data, and the latest image transformation studies.

1. Introduction

1.1. Backgroud and Purpose

Understanding the process by which demand is created, identifying the cause of demand generation, and considering these when demand forecasting can help develop demand forecasting technology. In particular, this contextual information can be very useful in demand forecasting for consumer goods with high demand volatility or spare parts with intermittent demand patterns [1].
Most causes of demand are unclear because seasonal factors, regional factors, and user group factors are all integrated. Therefore, it is necessary to determine the explanatory variables by decomposing them into multi-levels, such as year, region, and user group, as shown in Figure 1. That is, for each sample, the following is generated: k (number of explanatory variables) × l (number of clusters in level 1) × m (number of clusters in level 2) × n (number of clusters in level 3) dimensional features to be considered for each sample. More information in the forecasting process requires more parameter estimation, so more complex predictive models and methods are needed [2].
Convolutional neural networks (CNNs) demonstrate high performance in classification and prediction using image data, and they have been successfully used in many applications, such as image and video recognition [3,4,5,6], medical image analysis [7,8], natural language processing [9], and speech recognition [10]. However, the tabular data used for demand forecasting is not suitable for CNN modeling, so the application of CNNs in the demand forecasting domain is still limited.
In order to overcome this limitation, it is necessary to convert a tabular feature set into an image, and in this process, it is necessary to develop a method that can effectively perform feature arrangement. Therefore, in this study, we propose a Tab2vox framework that reads spatial context information through 3D CNNs, reflects the multivariate multilevel feature set in demand forecasting, and evaluates its effect (See Figure 2).
Tab2vox is proposed based on DARTS [11], one of the neural architecture search (NAS) models; DARTS enables the learning of optimized voxel images and 3D CNN structures for demand forecasting based on gradient descent. By extending the DARTS design to solve the classification problem by searching the 2D CNN architecture, we also searched the input-embedding architecture mapping the tabular input data to the optimal voxel image. Furthermore, the deep learning network was modified from the 2D to 3D convolution search space and from the classifier to regressor. These modifications were carried out so that demand forecasting could be performed in consideration of the spatial information of the multivariate multilevel feature.

1.2. Related Work

1.2.1. Previous Studies on Demand Forecasting

Demand forecasting research has become much more developed over the past few decades. Traditionally, statistical techniques such as simple exponential smoothing (SES), simple moving average (SMA), ARIMA, and Croston have been mainly used to predict future demand only with past demand, without explanatory variables [12]. Statistical techniques have been widely used for their simplicity and ease of implementation, but they do not reflect demand-influencing factors in forecasting, and their purpose is not to improve forecasting accuracy, but to find and infer the line that best explains the current situation. Therefore, it is difficult to generalize when new patterns of demand appear [13,14,15].
Data mining is a method to help decision-making analytically by extracting influencers, relationships, patterns, and rules from large-scale data [16]. In the field of demand forecasting, machine learning (ML) approaches such as xGBoost [17], Adaboost [18], and random forests [19] showed excellent performance. Recently, RNN-based [20] algorithms such as LSTM [21] and DeepAR [22], which are specialized for time series data, and techniques using deep learning, such as Transformer-based Tabnet [23] and TabTransformer [24], have been utilized a lot [25].
However, the RNN-based algorithm is effective for data with a sufficiently long time series, and the transformer-based model has many hyperparameters to be tuned with a high level of expertise among deep learning models, so it is not yet widely applied.
Various prediction methodologies were applied by industries, such as the medical [26], resource [27,28,29,30], economic [31], and military industries [32], but there have been no studies of CNN deep learning network applications using image data.

1.2.2. Studies on Converting Tabular Data into Images

The case of converting tabular format data into images for CNN deep learning-based prediction has been proposed for gene classification in several recent studies in the biotechnology field, but none in the field of demand forecasting.
The first proposed OmicsMapNet [33] developed a method for converting genetic data into 2D images for tumor grade prediction in cancer patients. However, ontology knowledge extracted from gene and genome encyclopedias was used for feature hierarchical rearrangement, and there is a limitation in that it is difficult to utilize if there is no domain knowledge about the corresponding dataset.
Several studies have proposed methods for projecting n feature vectors into 2D space by reducing them to p(<n) dimensions using dimension reduction methodology [34,35,36]. When dimensionality reduction is performed, high-dimensional features are displayed in a space that can be visually recognized by humans, so it has the advantage of intuitively helping to understand data and reducing the number of features. However, since the dimension-reduced features have completely different values from the existing features, information loss occurs from the original data. Therefore, it is difficult to interpret the extracted variables [37].
IGTD [38] took a slightly different approach of converting genetic data into a two-dimensional image and predicting an anticancer drug reaction with CNNs. It generates an image by creating an R matrix representing the correlation between features and a Q matrix representing the distance between image coordinates. By exchanging the feature positions of the R matrix so that the R matrix and Q matrix are equal, the highly correlated features are optimized to be located nearby.

1.2.3. Research Gaps

As a result of reviewing the previous studies presented in Section 1.2.1 and Section 1.2.2, a research gap was found. It can be summarized as follows:
  • Although demand forecasting methodology has become increasingly sophisticated, studies applying the CNN architecture using image data have not yet been conducted.
  • Recently, there have been attempts to express tabular data as images in some studies in the field of biotechnology, but image transformation and an evaluation of predictive models were performed independently. That is, it was not a method in which optimal image conversion was performed by linking with the loss function (predictive measurement accuracy).
  • In addition, previous studies converting tabular features into images have extracted features using 2D CNNs in the form of integrated demand sources; however, a high-dimensional CNN that integrates and reflects multi-dimensional information between subdivided levels is required.
This study differs from existing demand forecasting studies because it uses image data. It differs from the existing tabular-to-image conversion studies because it applies image transformation directly linked to loss, and uses 3D CNNs to reflect the aspects of multilevel features.

1.3. Organization of This Paper

The rest of this paper contains five sections.
Chapter 2 describes Tab2vox, a 3D CNN-based demand forecasting model that converts tabular data into 3D voxel images. Chapter 3 introduces the research framework applied to the military vehicle spare parts case. Chapter 4 compares different methods related to forecast accuracy and examines the usefulness of using 3D image data and 3D CNN models for demand forecasting. Finally, Chapters 5 and 6 conclude the paper by summarizing the study and explaining the limitations of this paper to be explored in future research.

2. Materials and Methods

2.1. Neural Architecture Search (NAS)

Automated machine learning (AutoML) is a research field that automates the entire machine learning pipeline, from data preparation to model evaluation. Recently, neural architecture search (NAS), which automatically finds network architectures such as layer type and number of layers, has attracted attention [39]. NAS uses the objective function to minimize validation loss or forecast accuracy, and it is trained by defining which architecture to use as a search space.
There are four representative approaches to NAS: reinforcement learning, evolutionary, Bayesian, and gradient-based algorithms, and which algorithm to choose remains a manual problem. All other NAS methodologies before DARTS [11] performed architecture searches in such a way that the model was evaluated after sampling the architecture candidates from the discrete search space, and then the procedure of sampling and evaluating other architecture candidates was repeated. In this process, evaluating thousands or tens of thousands of structures results in a huge search cost. To overcome this problem, DARTS applies continuous relaxation to the discrete search space to make it differentiable and to perform a fast and efficient architecture search through gradient descent-based optimization.
In many cases, model run time has a particular importance in demand forecasting, such as in the case of consumer goods that rely on checking the sale price, weather, season, and event status in real time, or the defense domain that requires combat-readiness. In this study, based on the DARTS algorithm, which has the fastest run time among NAS methodologies, we propose a demand forecasting framework that (1) searches for an optimal method for converting tabular data into 3D voxel images, and (2) searches for an optimal 3D CNN network.

2.2. Proposed Method: Tab2vox

Tab2vox extends the existing DARTS, that ordinarily searches only CNN network architectures, to search architectures that map tabular input data to images. This section explains how Tab2vox performs NAS. There are two architectures that Tab2vox searches: an input representation architecture that maps tabular format data to a 3D voxel image, and a 3D CNN architecture.
The first stage searches the input-embedding architecture, which includes a four-step workflow.
(i) As an example, suppose that we do not know in what order to map the features of each level to the 3D voxel image, as shown in Table 1(a). First, the image is initialized by arranging the feature set for each level in the tabular form on each arbitrary axis of the voxel image. In tabular format data, the order in which the feature set is listed is not important, but in the voxel image, adjacent cells are meaningful.
(ii) Next, clustering is performed on each feature set for each level, as shown in Table 1(b). The point of this study is to determine how to map features divided into k explanatory variables (level 0), l region (level 1), and m users (level 2) to each axis of the 3D voxel image. For this, the number of cases k! × l! × m! should be defined as the search space and searched. Therefore, in this study, first, features by level were grouped to reduce the search space. In this case, grouping may be performed using k-means [40], knn [41], hierarchical clustering [42], etc., which are commonly used cluster analysis techniques.
Assume that, after performing clustering, groups for each level are classified into, for example, 5, 3, and 2 clusters, respectively. Even with clustering, the number of voxel images that can be created is 5! × 3! × 2! = 1440, and if you train the model weight parameters for all these architectures and repeat this process, a large search cost will occur.
(iii) Therefore, in Tab2vox, as shown in Table 1(c), the output value of the input-embedding network is defined as a weighted sum by candidate mapping images to make it differentiable. This applies the idea of DARTS. For the operation o that connects tabular data to image i, α o i is considered a weight, multiplied by o(x) by the softmax weight, and all the values are then added to calculate the weight of candidate image operations. That is, the larger α o i is, the more the operation is affected, and the closer that α o i is to 0, the more the operation does not play any role. The weights defined in this way become continuous variables, meaning that these weights can be trained to minimize validation loss using gradient descent.
Secondly, CNN architecture is also searched by using the network the architecture search concept of DARTS as it is. For the purpose of this study, only the operation set was extended from two dimensions to three dimensions by referring to existing studies to read multi-dimensional information between levels [43,44,45]. DARTS also constructs a CNN model by creating cells like VGG [46] and RestNet [47] and stacking cells repeatedly. At this time, the number of cells, the number of nodes in the cell, and the normal/reduction cell configuration are all based on the assumptions of DARTS.
Finally, for the regressor, the network at the end is flattened, connected, and passed to the fully connected network for prediction. DARTS was originally designed to solve the classification problem, but in this study, the output of the last linear layer was converted into one node so that the regression task could be performed.
By training the Tab2vox network designed in this way, the optimal unique image in the form of a 3D voxel for each sample and the optimal 3D CNN network are obtained. Table 1 summarizes the differences between Tab2vox and the existing DARTS. The Supplementary Materials includes detailed information on the Tab2vox framework.

3. Case Study of MND (Ministry of National Defense, Republic of Korea) Spare Parts Dataset by Tab2vox

3.1. Datasets

To verify the performance of Tab2vox, we proposed a research framework, as shown in Figure 3. The following sections describe the details of the proposed research framework.

3.1.1. Data Collection

For the analysis, the ROK K-511 vehicle spare parts dataset was used. Military vehicles have the largest number of weapon systems in operation in the ROK and have a diverse operating area, so relatively sufficient samples and explanatory variables can be used. In addition, the analysis was conducted on 2290 types of spare parts that were in demand between 2010 and 2019.
The K-511 Military Vehicle dataset used in the analysis was collected from the Defense Logistics Integrated Information System (DELIIS). This system stores information related to the ROK military’s weapon systems and spare parts, such as supply transaction history, maintenance history, and equipment operation record. All data available since 2010 when the system was first developed were used in this study.
Furthermore, topographical and climate data were collected from the National Statistics Portal of the Korea National Statistical Office (https://kosis.kr/, accessed on 1 January 2022) and the Data Open Portal of the Korea Meteorological Administration (https://data.kma.go.kr/, accessed on 1 January 2022), respectively, and they were used as explanatory features for units in operation K-511.

3.1.2. Feature Generation

The collected data, such as the daily spare parts replacement history, supply/maintenance history, etc., were directly processed and used for analysis by year, base, and equipment level. As mentioned at the beginning of this paper, the cause of most demand is unclear because seasonal factors, regional factors, and user group factors are all integrated. It is necessary to consider decomposing features into multi-level decomposition.
(a) equipment-level decomposition: In addition to the basic model used to transport military supplies and troops, K-511 vehicles are operated as various derivatives for maintenance, oil transport, command and control, water supply, decryption, and corrosion transport. Vehicles operating with the same mission have a high failure probability of similar spare parts, and conversely, vehicles with different characteristics and operating concepts will show different demand patterns, even for the same item.
(b) base-level decomposition: Even with the same spare part, the spare parts of vehicles operating in coastal areas can be more corroded due to sea winds, and the spare parts of equipment operating in mountainous areas—such as track spare parts—can be frequently in demand due to the rugged terrain.
(c) year-level decomposition: Moreover, yearly data can reflect time-series characteristics of spare parts failure.
Table 2 lists and explains 52 variables created for military vehicle demand forecasting. Variables are divided into 13 spare part specification variables, 23 demand history variables, and 16 equipment operating environment variables. The explanatory feature set includes all variables used in previous studies. In previous studies, features were divided up to the year level and the base level at the maximum, and in this study, features were extended to the equipment level and used for prediction.
The explanatory variable consists of data from 2010–2018, and demand in 2019 was defined as the target variable.

3.1.3. Data Preparation

For the explanatory features divided by the level created in this way, outliers were removed by applying the outer fence of the box-plot, and the missing value was corrected using similar items and similar base information. In addition, except for the time series models that only uses annual demand data, the data distribution of the existing variables was converted into a normal distribution with a mean of 0 and a variance of 1, and was used as input data after normalization.
For fair comparison, all predictive models were trained and tested with 5-fold cross validation using the same data partition (i.e., training, validation, and testing sets). For the collected data, 60% of the item data was used for model training, 20% was used for validation to avoid overfitting, and the remaining 20% was used to test the prediction performance.

3.1.4. Data Representation & Prediction

As shown in Figure 4, the collected data were initially refined into a matrix consisting of m = 1,008,549 rows (= item × base × equipment × year) and n = 56 columns (=4 level distinguisher + 52 explanatory features). After that, according to the input data format for each model, it was converted into tabular univariate or multivariate data, or two-dimensional or three-dimensional images and used for prediction.

3.2. Evaluation with Comparisons

3.2.1. Overview

The main goal of this study is to show that CNN architecture can be applied using non-image data through Tab2vox implementation. To prove the effectiveness of the Tab2vox demand forecasting framework, the performance was evaluated in comparison with the time series, regression, and ensemble methods, which have typically been adopted as benchmarking methods in previous studies. For the hyperparameters of comparative models, random grid search [48] optimization was performed to ensure a fair comparison with the best performance for each model. Next, 5-fold cross validation was performed to compare the average prediction accuracy for five test sets.
In addition, the performance was compared with IGTD, which is achieving state-of-the-art (SOTA) results among existing studies converting tabular data into image form.

3.2.2. Tab2vox Architecture for the MND Dataset

By applying Tab2vox NAS to the MND dataset, the optimal image-embedding architecture and CNN architecture—as shown in Figure 5—were derived and used for comparison with other prediction techniques.

3.2.3. Machine Learning Methods

Table 3 shows the prediction models analyzed together for this study’s Tab2vox performance evaluation; all models use tabular input data.
To implement each technique, the machine learning technique used the PyCaret AutoML module and IGTD used the code from GitHub.
  • Machine Learning: the PyCaret Python package was used with the tuned parameters.
  • IGTD: the public IGTD source code was used, which is available at https://github.com/zhuyitan/IGTD (accessed on 1 January 2022).

4. Results

4.1. Evaluation Metrics

To evaluate the demand forecast performance between the predicted value F i and the actual value A i for item i derived by the models, three measures commonly used to evaluate the forecast model—the root mean square error (RMSE), the mean absolute error (MAE), and the min–max accuracy method—were used. The calculation formula for each measure is as follows (1)–(3).
The RMSE is calculated by rooting the mean of the squares of the difference between the predicted and actual values. The RMSE measure can be very useful for the dataset used in this study because it reduces the scale when the error has a very large value. The MAE is measured by averaging the absolute value of the difference between the predicted value and the actual value, and it is effective when used in an evenly distributed error. For both the RMSE and the MAE, the smaller the value, the smaller the prediction error, which means a better performance model.
The min–max accuracy method measures how close the predicted value is to the actual value by taking the average of the minimum and maximum values between the actual and predicted values. The min–max accuracy ranges between 0 and 1, and if the min–max accuracy is 1, the actual and predicted values are equal. Unlike RMSE and MAE, min-max accuracy has the advantage of easy interpretation because it measures the ratio of the observed value to the predicted value regardless of the scale.
Formula of the RMSE:
RMSE = i = 1 n F i A i 2 n
Formula of the MAE:
MAE = 1 n i = 1 n F i A i
Formula of the min–max accuracy:
M i n M a x   A c c u r a c y = m e a n min A i ,   F i max A i ,   F i
In the above equation, i denotes each sample, F i denotes a predicted value by each model, and A i denotes an actual value.

4.2. Prediction Accuracy by a Single Model

Table 4 and Figure 6 show the demand forecasting accuracy derived using various data representations and forecasting models. To comprehensively evaluate the proposed Tab2vox model, we compared the mean of performance metrics for five test sets using time series and ML techniques.
Among all methods, Tab2vox showed the best performance, showing 63.1% in terms of the average min–max prediction accuracy of all items. Next, decision tree, a regression method, showed good performance with 62.9%; random forest, an ensemble method, with 54.5%; XGBoost with 50.6%; and Catboost with 49.3%. For reference, since RMSE and MAE have scale-dependent error characteristics, they are presented as auxiliary indicators of min-max prediction accuracy, meaning that Tab2vox showed better performance in items with relatively small average demand size.
On the other hand, other time series techniques, regression models, and classical ensemble techniques did not show good performance in predicting the demand for spare parts. IGTD, a state-of-the-art model applying CNN architecture by converting tabular-format data into images, also did not perform well. It is thought that this may be due to the fundamental difference in data characteristics between the spare parts dataset of the intermittent demand pattern and the dense genetic dataset. Further analysis is required.

4.3. Ensemble Approach

4.3.1. Model Selection Problem

“All models are wrong.” This is a quote from the British statistician George EP Box [49,50].
This means that there is no single model that can completely explain all cases. In this chapter, we tried to determine whether better performance could be secured with the most effective model combination for each item group, rather than applying a single model to all items, as shown in Figure 7.
In this analysis, 458 randomly selected items were analyzed. The time series method that does not consider demand explanatory features, and IGTD—which had the lowest prediction accuracy—was excluded. Only the Tab2vox model and multivariate machine learning method were included. The forecast values for each item of the individual model were created as a new dataset, and the average forecasts of the item group were used as metadata.
First, cluster analysis was performed to classify item groups. Table 5 shows the results of cluster analysis. Four item groups were classified through hierarchical clustering based on the item unit cost and average annual demand, the features most correlated with demand generation. The number of clusters was set in consideration of the silhouette score, a metric that can quantitatively evaluate clustering.
Next, an integer programming (IP) model—which searches for the optimal combination of predictive models for each item group—was proposed based on the assumptions and notation in Table 6. The IP model was coded in the General Algebraic Modeling System (GAMS) 39.2.1 with CPLEX 22.1.0 on a Core i7 processor and 16GB RAM.
M a x i m i z e   i     I j     J ( c i / c )   a c c i j   x i j
j x i j = 1 ,         i   I
i     I j     J ( c i / c )   t j   x i j   T
x i j 0 , 1           f o r       i ,   j
Equation (4) searches for a combination of predictive models that maximize the accuracy of each item group as an objective function. In constraint (5), one item group selects one predictive model, and constraint (6) sets the upper limit of the model run time of all item groups. Finally, Equation (7) is a constraint for setting a decision variable.
As a result of exploring the optimal predictive model combination for the four item groups, it was found that 65.58% accuracy could be achieved if the proposed Tab2vox model was used with dt, lasso, and xgboost, as shown in Table 7. This result shows an improvement of 2.58%, compared to when Tab2vox was applied as a single technique to all items.
This time, we analyzed the feasible region of Tab2vox according to the number of item groups, run time constraints, and robustness scenarios.

4.3.2. Optimal Model Combination Considering the Number of Item Group Clusters

In the previous section, the optimal cluster size of the test item group was analyzed as four. However, when using the model in an actual industrial field, it is also possible to use a model for a smaller or larger item group according to the capacity. Figure 8 and Table 8 show the results of deriving the optimal model combination and prediction accuracy according to item group size using the previously defined IP model. The item group classification result and its criteria can be found in the Supplementary Materials.
When the Tab2vox model was used together with the dt, xgboost, lasso, and br models, it was found that a prediction accuracy of 66.85% could be achieved when there were 10 item groups. Just because an item group is currently supplemented by other machine learning models does not mean that Tab2vox cannot work well with that item group. This is because the Tab2vox model applied in this study has an image embedding- and CNN network architecture that has been searched for good performance on average for all items. If the optimal architecture is searched by applying Tab2vox to each item group, Tab2vox can be applied to more item groups.
On the other hand, as the item groups are subdivided and the number of models used increases, management requirements increase accordingly, and there is a possibility that the model combination may be judged as overfitting. Therefore, it is necessary to apply an appropriate item group x model combination, which is beyond the scope of this study.

4.3.3. Optimal Model Combination According to Run Time Constraints

We looked for the optimal model combination according to the run-time constraint and the feasible region that can use Tab2vox. As shown in Figure 9, if the run-time is given for more than four hours based on the 1832 items of the train dataset used in this analysis, Tab2vox can be useful as it has a high performance of 65.58%.
In this study, due to the difficulty of measuring quantitative indicators, only the pure training time of the model was considered as the run time. However, considering the time spent on feature engineering or hyperparameter tuning in machine learning techniques, Tab2vox can be much more useful.

4.3.4. Optimal Model Combination Considering Robustness

By changing the objective function of the IP model proposed in Section 4.3.2, we examined the optimal model combination considering the robustness. The model robustness was defined as the standard deviation of five test sets set to 5-fold, and the analysis was conducted based on the optimal number of clusters: four item groups. As shown in Figure 10, even considering the ratio of robustness up to half, Tab2vox was found to work effectively in most item groups.
M a x i m i z e   i     I j     J W 1 × ( c i / c ) × a c c i j   x i j W 2 × ( c i / c ) × s t d i j   x i j

5. Discussion

A CNN takes a sample as an image and uses a set of adjacent pixels to learn various features. However, the application of CNNs in the field of demand forecasting has been limited. This is because tabular data, which has been traditionally used for demand forecasting, assumes that the relationship among features is independent. To apply high-dimensional tabular data to CNNs, it is necessary to link non-image information to image-based learning. In this work, we propose an optimal image conversion search framework for using high-dimensional multilevel features as inputs to CNN architecture.
As expected, the mapping of tabular data found by the Tab2vox NAS model to the voxel image space appears to be useful for feature extraction using a 3D CNN architecture. Compared to traditional time series and machine learning, it showed better performance, and it even surpassed the performance of the latest image conversion model proposed in the field of biotechnology. As a result of applying it to the ROK military vehicle spare parts dataset, Tab2vox showed 63.10% performance when applied with a single model to all items; this is the best performance among other baseline models. Furthermore, when Tab2vox was used together with other machine learning techniques, the prediction accuracy in the optimal item group x model combination was 65.58%.

6. Conclusions

6.1. Implications

The significance of the proposed Tab2vox methodology can be summarized as follows.
First, the proposed demand forecasting framework does not solve all demand forecasting problems. However, it does offer enormous potential because it provides a cornerstone for applying the demand forecasting problem to CNNs, an advanced model with many advantages. Converting similar features into images by inserting them into neighbors is more meaningful and powerful than ignoring and processing neighbor information in machine learning. The 3D CNNs extract features that are subdivided into several levels in multiple dimensions, so they are particularly useful for identifying and predicting the cause of demand for items with large demand fluctuations or intermittent demand.
Second, previous studies have proposed image conversion and CNN architectures fixed to specific data and domains. However, in this study, we propose a method to train the architecture itself according to the dataset of the domain using the NAS. This is meaningful because Tab2vox is a general-purpose model that can be applied not only to the forecasting demand of items but also to various applications using tabular data. Furthermore, it can be extended to various tasks, such as classification and regression.
Third, since human intervention is minimized using the concept of NAS, which is an AutoML model, it can be a much more powerful model in real-world industrial settings.

6.2. Limitations and Further Research

Further studies can be conducted to overcome the limitations of this study:
1.
It is only recently that researchers have begun to gather big data. We wanted to verify the Tab2vox framework in datasets from various industries, but it was difficult to acquire benchmark datasets segmented to a multi-level format. Additional verification by datasets from various domains is required.
2.
After DARTS was first introduced, many NAS models with differentiable concepts appeared. With a more recent and advanced NAS model, the Tab2vox input embedding method could be further advanced.
3.
Unlike previous studies using dimensionality reduction methodology, Tab2vox maps the features of each level to each dimension of the voxel image while preserving the dimension of the feature set. By applying the explainable artificial intelligence (XAI) technique, it is possible to add explainability to the model by performing actions such as displaying a heat map showing features on which the model focuses, and generating demand. Intuitive interpretation is possible because a single voxel itself has an explicit meaning.
4.
Because of the intermittent demand characteristics, we had no choice but to use yearly data. In addition, despite using all datasets after 2010, when the Defense Logistics Integrated Information System (DELIIS) was developed, the time series period was not sufficient. Therefore, the target year was set to 1 year for analysis. The effect on mid- to long-term forecasting can also be reviewed with industrial data, such as production line data and electricity demand data, where demand occurs continuously rather than intermittently and the time series is sufficient.
5.
In this work, the average effective image-embedding architecture and CNN architecture were searched for entire items by applying the Tab2vox framework to entire items. However, if the Tab2vox framework is applied to each item group, it will be possible to find a network that fits the characteristics of the item group. This is possible when sufficient samples for each item group are secured.
Key Points
  • Tab2vox is a NAS model that embeds tabular data into optimal voxel images and finds the optimal CNN architecture for the images.
  • We demonstrated better performance, compared to previous time series and machine learning techniques, using tabular data and the latest tabular to image conversion studies.
  • Tab2vox is particularly meaningful because it is a general-purpose model that provides a framework that can be applied to demand forecasting in all sectors, rather than proposing a fixed architecture verified only in a specific dataset and domain.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/su141811745/s1, Figure S1: Architecture of the tab2vox demand forecasting framework; Figure S2: Result of item group analysis considering unit price of item and annual demand and its criteria.

Author Contributions

Conceptualization, E.L. and H.L.; methodology, E.L.; validation, E.L. and M.N.; formal analysis, E.L. and M.N.; investigation, E.L. and M.N.; resources, H.L.; data curation, E.L.; writing—original draft preparation, E.L.; writing—review and editing, E.L. and H.L.; visualization, E.L.; supervision, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The implementation of Tab2vox is available at https://github.com/kellyeunalee/tab2vox.

Acknowledgments

The authors would like to thank the anonymous reviewers for their valuable comments that helped revise the original version of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Syntetos, A.A.; Boylan, J.E. Intermittent Demand Forecasting: Context, Methods and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
  2. Gonçalves, J.N.C.; Cortez, P.; Carvalho, M.S.; Frazão, N.M. A multivariate approach for multi-step demand forecasting in assembly industries: Empirical evidence from an automotive supply chain. Decis. Support Syst. 2021, 142, 113452. [Google Scholar] [CrossRef]
  3. Garcia, C.; Delakis, M. Convolutional face finder: A neural architecture for fast and robust face detection. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1408–1423. [Google Scholar] [CrossRef] [PubMed]
  4. Hadsell, R.; Sermanet, P.; Ben, J.; Erkan, A.; Scoffier, M.; Kavukcuoglu, K.; Muller, U.; LeCun, Y. Learning Long-Range Vision for Autonomous Off-Road Driving. J. Field. Robot. 2009, 26, 120–144. [Google Scholar] [CrossRef]
  5. Tompson, J.; Goroshin, R.; Jain, A.; LeCun, Y.; Bregler, C. Efficient object localization using Convolutional Networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 648–656. [Google Scholar] [CrossRef]
  6. Sermanet, P.; Kavukcuoglu, K.; Chintala, S.; Lecun, Y. Pedestrian detection with unsupervised multi-stage feature learning. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3626–3633. [Google Scholar] [CrossRef]
  7. Kather, J.N.; Pearson, A.T.; Halama, N.; Jäger, D.; Krause, J.; Loosen, S.H.; Marx, A.; Boor, P.; Tacke, F.; Neumann, U.P.; et al. Deep learning can predict microsatellite instability directly from histology in gastrointestinal cancer. Nat. Med. 2019, 25, 1054–1056. [Google Scholar] [CrossRef]
  8. Schmauch, B.; Romagnoni, A.; Pronier, E.; Saillard, C.; Maillé, P.; Calderaro, J.; Kamoun, A.; Sefta, M.; Toldo, S.; Zaslavskiy, M.; et al. A deep learning model to predict RNA-Seq expression of tumours from whole slide images. Nat Commun. 2020, 11, 3877. [Google Scholar] [CrossRef]
  9. Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural Language Processing (almost) from Scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. Available online: http://arxiv.org/abs/1103.0398 (accessed on 1 January 2022).
  10. Sainath, T.N.; Kingsbury, B.; Mohamed, A.-R.; Dahl, G.E.; Saon, G.; Soltau, H.; Beran, T.; Aravkin, A.Y.; Ramabhadran, B. Improvements to deep convolutional neural networks for LVCSR. arXiv 2013, arXiv:1309.1501. Available online: http://arxiv.org/abs/1309.1501 (accessed on 1 January 2022).
  11. Liu, H.; Deepmind, K.S.; Yang, Y. Darts: Differentiable architecture search. arXiv Prepr. 2018, arXiv:1806.09055. Available online: https://github.com/quark0/darts (accessed on 1 January 2022).
  12. Syntetos, A.A.; Boylan, J.E. The accuracy of intermittent demand estimates. Int. J. Forecast. 2005, 21, 303–314. [Google Scholar] [CrossRef]
  13. Bzdok, D.; Altman, N.; Krzywinski, M. Statistics versus machine learning. Nat. Methods 2018, 15, 233. [Google Scholar] [CrossRef]
  14. Romeijnders, W.; Teunter, R.; van Jaarsveld, W. A two-step method for forecasting spare parts demand using information on component repairs. Eur. J. Oper. Res. 2012, 220, 386–393. [Google Scholar] [CrossRef]
  15. Dekker, R.; Pinçe, Ç.; Zuidwijk, R.; Jalil, M.N. On the use of installed base information for spare parts logistics: A review of ideas and industry practice. Int. J. Prod. Econ. 2013, 143, 536–545. [Google Scholar] [CrossRef]
  16. Suh, J.H. Generating future-oriented energy policies and technologies from the multidisciplinary group discussions by text-mining-based identification of topics and experts. Sustainability 2018, 10, 3709. [Google Scholar] [CrossRef]
  17. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar] [CrossRef]
  18. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 1995; Volume 904, pp. 23–37. [Google Scholar] [CrossRef]
  19. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  20. Sherstinsky, A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network. Phys. D Nonlinear Phenom. 2018, 404, 132306. [Google Scholar] [CrossRef]
  21. Sak, H.H.; Senior, A.; Google, B. Long Short-Term Memory Recurrent Neural Network Architectures for Large Scale Acoustic Modeling. In Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, Singapore, 14–18 September 2014; pp. 338–342. [Google Scholar]
  22. Salinas, D.; Flunkert, V.; Gasthaus, J. DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks. arXiv 2017, arXiv:1704.04110. Available online: http://arxiv.org/abs/1704.04110 (accessed on 1 January 2022). [CrossRef]
  23. Arik, S.O.; Pfister, T. TabNet: Attentive Interpretable Tabular Learning. arXiv 2019, arXiv:1908.07442. Available online: http://arxiv.org/abs/1908.07442 (accessed on 1 January 2022). [CrossRef]
  24. Huang, X.; Khetan, A.; Cvitkovic, M.; Karnin, Z. TabTransformer: Tabular Data Modeling Using Contextual Embeddings. arXiv 2020, arXiv:2012.06678. Available online: http://arxiv.org/abs/2012.06678 (accessed on 1 January 2022).
  25. Rosienkiewicz, M.; Chlebus, E.; Detyna, J. A hybrid spares demand forecasting method dedicated to mining industry. Appl. Math. Model. 2017, 49, 87–107. [Google Scholar] [CrossRef]
  26. Soltani, M.; Farahmand, M.; Pourghaderi, A.R. Machine learning-based demand forecasting in cancer palliative care home hospitalization. J. Biomed. Inform. 2022, 130, 104075. [Google Scholar] [CrossRef]
  27. Hussain, A.; Alam Memon, J.; Murshed, M.; Alam, S.; Mehmood, U.; Alam, M.N.; Rahman, M.; Hayat, U. A time series forecasting analysis of overall and sector-based natural gas demand: A developing South Asian economy case. Environ. Sci. Pollut. Res. 2022. [Google Scholar] [CrossRef]
  28. Shorfuzzaman, M.; Hossain, M.S. Predictive Analytics of Energy Usage by IoT-Based Smart Home Appliances for Green Urban Development. ACM Trans. Internet Technol. 2022, 22, 1–26. [Google Scholar] [CrossRef]
  29. Sheikhoushaghi, A.; Gharaei, N.Y.; Nikoofard, A. Application of Rough Neural Network to forecast oil production rate of an oil field in a comparative study. J. Pet. Sci. Eng. 2022, 209, 109935. [Google Scholar] [CrossRef]
  30. Fan, G.-F.; Yu, M.; Dong, S.-Q.; Yeh, Y.-H.; Hong, W.-C. Forecasting short-term electricity load using hybrid support vector regression with grey catastrophe and random forest modeling. Util. Policy 2021, 73, 101294. [Google Scholar] [CrossRef]
  31. Tschora, L.; Pierre, E.; Plantevit, M.; Robardet, C. Electricity price forecasting on the day-ahead market using machine learning. Appl. Energy 2022, 313, 118752. [Google Scholar] [CrossRef]
  32. Choi, B.; Suh, J.H. Forecasting spare parts demand of military aircraft: Comparisons of data mining techniques and managerial features from the case of South Korea. Sustainability 2020, 12, 6045. [Google Scholar] [CrossRef]
  33. Ma, S.; Zhang, Z. OmicsMapNet: Transforming omics data to take advantage of Deep Convolutional Neural Network for discovery. Available online: http://www.kegg.jp/ (accessed on 1 January 2022).
  34. Sharma, A.; Vans, E.; Shigemizu, D.; Boroevich, K.A.; Tsunoda, T. DeepInsight: A methodology to transform a non-image data to an image for convolution neural network architecture. Sci. Rep. 2019, 9, 11399. [Google Scholar] [CrossRef]
  35. Bazgir, O.; Zhang, R.; Dhruba, S.R.; Rahman, R.; Ghosh, S.; Pal, R. Representation of features as images with neighborhood dependencies for compatibility with convolutional neural networks. Nat. Commun. 2020, 11, 4391. [Google Scholar] [CrossRef]
  36. Tang, H.; Yu, X.; Liu, R.; Zeng, T. Vec2image: An explainable artificial intelligence model for the feature representation and classification of high-dimensional biological data by vector-to-image conversion. Brief Bioinform. 2022, 23, bbab584. [Google Scholar] [CrossRef]
  37. Box, P.O.; van der Maaten, L.; Postma, E.; van den Herik, J. Tilburg Centre for Creative Computing Dimensionality Reduction: A Comparative Review Dimensionality Reduction: A Comparative Review. 2009. Available online: http://www.uvt.nl/ticc (accessed on 1 January 2022).
  38. Zhu, Y.; Brettin, T.; Xia, F.; Partin, A.; Shukla, M.; Yoo, H.; Evrard, Y.A.; Doroshow, J.H.; Stevens, R.L. Converting tabular data into images for deep learning with convolutional neural networks. Sci. Rep. 2021, 11, 11325. [Google Scholar] [CrossRef]
  39. He, X.; Zhao, K.; Chu, X. AutoML: A Survey of the State-of-the-Art. Available online: https://www.datasearch.elsevier.com/ (accessed on 1 January 2022).
  40. Mannor, S.; Jin, X.; Han, J.; Zhang, X. K-Means Clustering. In Encyclopedia of Machine Learning; Springer: Boston, MA, USA, 2011; pp. 563–564. [Google Scholar] [CrossRef]
  41. Altman, N.S. An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression. Am. Stat. 1992, 46, 175–185. [Google Scholar] [CrossRef]
  42. Nielsen, F. Hierarchical clustering. In Introduction to HPC with MPI for Data Science; Springer: Cham, Switzerland, 2016; pp. 195–211. [Google Scholar] [CrossRef]
  43. Yu, Q.; Yang, D.; Roth, H.; Bai, Y.; Zhang, Y.; Yuille, A.L.; Xu, D. C2FNAS: Coarse-to-Fine Neural Architecture Search for 3D Medical Image Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  44. Zhou, B.; Li, Y.; Wan, J. Regional Attention with Architecture-Rebuilt 3D Network for RGB-D Gesture Recognition. arXiv 2021, arXiv:2102.05348. Available online: http://arxiv.org/abs/2102.05348 (accessed on 1 January 2022). [CrossRef]
  45. Carreira, J.; Zisserman, A. Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. arXiv 2017, arXiv:1705.07750. Available online: http://arxiv.org/abs/1705.07750 (accessed on 1 January 2022).
  46. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. Available online: http://arxiv.org/abs/1409.1556 (accessed on 1 January 2022).
  47. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. Available online: http://image-net.org/challenges/LSVRC/2015/ (accessed on 1 January 2022).
  48. Bergstra, J.; Ca, J.B.; Ca, Y.B. Random Search for Hyper-Parameter Optimization Yoshua Bengio. 2012. Available online: http://scikit-learn.sourceforge.net (accessed on 1 January 2022).
  49. Carrera, B.; Kim, K. Comparison analysis of machine learning techniques for photovoltaic prediction using weather sensor data. Sensors 2020, 20, 3129. [Google Scholar] [CrossRef]
  50. Box, G.E.P. Science and Statistics. J. Am. Stat. Assoc. 1976, 71, 791–799. [Google Scholar] [CrossRef]
Figure 1. Decomposition of demand explanatory multivariate variables by level.
Figure 1. Decomposition of demand explanatory multivariate variables by level.
Sustainability 14 11745 g001
Figure 2. An example figure where the 3D CNN spatially reads multivariate multilevel features and extracts a feature map. Demand can be decomposed into levels by year, region, and vehicle model, and the demand explanatory variable is also divided and reflected for each level. The 3D CNN spatially reads the demand cause information across variables between neighboring levels and reflects it in the prediction.
Figure 2. An example figure where the 3D CNN spatially reads multivariate multilevel features and extracts a feature map. Demand can be decomposed into levels by year, region, and vehicle model, and the demand explanatory variable is also divided and reflected for each level. The 3D CNN spatially reads the demand cause information across variables between neighboring levels and reflects it in the prediction.
Sustainability 14 11745 g002
Figure 3. The research framework proposed by this paper.
Figure 3. The research framework proposed by this paper.
Sustainability 14 11745 g003
Figure 4. Initial ROK military vehicle spare parts demand dataset in tabular format. Demand by item is divided into base level, equipment level, and year level, and then explanatory features are considered.
Figure 4. Initial ROK military vehicle spare parts demand dataset in tabular format. Demand by item is divided into base level, equipment level, and year level, and then explanatory features are considered.
Sustainability 14 11745 g004
Figure 5. Architecture of the tabular-to-image conversion and the convolutional neural network (CNN) used for predicting the spare parts demand for military vehicles based on Tab2vox.
Figure 5. Architecture of the tabular-to-image conversion and the convolutional neural network (CNN) used for predicting the spare parts demand for military vehicles based on Tab2vox.
Sustainability 14 11745 g005
Figure 6. A boxplot showing the accuracy deviation of five test sets for each model.
Figure 6. A boxplot showing the accuracy deviation of five test sets for each model.
Sustainability 14 11745 g006
Figure 7. An ensemble approach to explore the combination of predictive models by item group.
Figure 7. An ensemble approach to explore the combination of predictive models by item group.
Sustainability 14 11745 g007
Figure 8. Optimal prediction model combination and prediction accuracy graph considering the number of item groups.
Figure 8. Optimal prediction model combination and prediction accuracy graph considering the number of item groups.
Sustainability 14 11745 g008
Figure 9. Optimal prediction model combination and prediction accuracy considering run time.
Figure 9. Optimal prediction model combination and prediction accuracy considering run time.
Sustainability 14 11745 g009
Figure 10. Optimal prediction model combination and prediction accuracy considering robustness.
Figure 10. Optimal prediction model combination and prediction accuracy considering robustness.
Sustainability 14 11745 g010
Table 1. DARTS vs Tab2vox.
Table 1. DARTS vs Tab2vox.
DARTSTab2vox
Input
Embedding
Architecture
Search
none Sustainability 14 11745 i001 Sustainability 14 11745 i002
Sustainability 14 11745 i003
CNN
Architecture
Search
Sustainability 14 11745 i004 Sustainability 14 11745 i005
Sustainability 14 11745 i006 Sustainability 14 11745 i007
TaskClassificationRegression
Table 2. Definition of explanatory features used for forecasting the spare parts demand of the military vehicle system in year t (t = 2010–2018).
Table 2. Definition of explanatory features used for forecasting the spare parts demand of the military vehicle system in year t (t = 2010–2018).
CategoryFeatureDefinitionTypeRangeLevel 1
123
Spare parts
specifications
X1Unit price of spare parts (KRW)numerical[3, 13,519,541]
X2Length of spare parts (cm)numerical[0, 350]
X3Width of spare parts (cm)numerical[0, 520]
X4Height of spare parts (cm)numerical[0, 138]
X5Weight of spare parts (kg)numerical[0, 1800]
X6Volume of spare parts (m3)numerical[0, 3]
X7Quantity Per Application (QPA): Number of spare parts installed per weapon systemnumerical[1, 168]
X8Base Repair Percent (BRP): Average percentage of spare parts repaired at bases out of maintenance requestsnumerical[0, 1]
X9Depot Repair Percent (DRP): Average percentage of spare parts repaired at depot out of maintenance requests from basesnumerical[0, 1]
X10Condemnation Percentage (ConPct): Average percentage of spare parts discarded as non-repairable by depotnumerical[0, 1]
X11Average repair cost of spare parts at depot base levelnumerical[2, 13,699,405]
X12Average repair cost of spare parts at depot maintenance levelnumerical[0.9, 7,229,964]
X13Procurement Lead Time (PROLT): Lead time, which is the period time (in days) from the purchase request to the receipt completion for spare partsnumerical[18, 999]
Consumption
history
X14Monthly spare parts demand average in year tnumerical[0, 243]
X15Monthly spare parts demand median in year tnumerical[0, 115]
X16Monthly spare parts demand deviation in year tnumerical[0, 182,992]
X17Daily spare parts demand average in year tnumerical[0, 8]
X18Daily spare parts demand median in year tnumerical[0, 1]
X19Daily spare parts demand deviation in year tnumerical[0, 4749]
X20The amount of spare parts consumption in January in year tnumerical[0, 1]
X21The amount of spare parts consumption in February from 2010 to year tnumerical[0, 1]
X22The amount of spare parts consumption in March from 2010 to year tnumerical[0, 1]
X23The amount of spare parts consumption in April from 2010 to year tnumerical[0, 1]
X24The amount of spare parts consumption in May from 2010 to year tnumerical[0, 1]
X25The amount of spare parts consumption in June from 2010 to year tnumerical[0, 1]
X26The amount of spare parts consumption in July from 2010 to year tnumerical[0, 1]
X27The amount of spare parts consumption in August from 2010 to year tnumerical[0, 1]
X28The amount of spare parts consumption in September from 2010 to year tnumerical[0, 1]
X29The amount of spare parts consumption in October from 2010 to year tnumerical[0, 1]
X30The amount of spare parts consumption in November from 2010 to year tnumerical[0, 1]
X31The amount of spare parts consumption in December from 2010 to year tnumerical[0, 1]
X32Average monthly spare parts inventory quantity in year tnumerical[0, 9150]
X33Spare parts consumption in year tnumerical[0, 2919]
X34Squared coefficient of variation (CV2): Standard deviation of the demand divided by the average demand for non-zero demand periods.numerical[0, 63]
X35Average inter-demand interval (ADI): Average interval time between two demand occurrences.numerical[0, 9]
X36Mean time between failures (MTBF):numerical[1, 9]
Operating
environment
X37Operating distance (in hours) of the weapon system, related to a spare part, in year tnumerical[0, 3,229,201]
X38Operating time (in hours) of the weapon system, related to a spare part, in year tnumerical[0, 7,854,334]
X39Oil consumption in year tnumerical[0, 406,682]
X40Average equipment age in year tnumerical[2.85, 35]
X41Number of weapon systems operated in year tnumerical[0, 3584]
X42Average temperature up to year t in the weapon system operating basesnumerical[21.2, 25]
X43Maximum temperature up to year t in the weapon system operating basesnumerical[25.3, 30]
X44Minimum temperature up to year t in the weapon system operating basesnumerical[15.9, 22]
X45Average amount of precipitation up to year t in the weapon system operating basesnumerical[1.2, 17]
X46Average humidity up to year t in the weapon system operating basesnumerical[61.4, 86]
X47Weapon system utilization rate in year tnumerical[0.98, 1]
X48Forest ratio to total area (%)numerical[23.76, 97]
X49Road pavement ratio to total area (%)numerical[3.72, 100]
X50Land ratio to total area (%)numerical[0.79, 1]
X51Sea ratio to total area (%)numerical[0, 0]
X52Urban ratio to total area (%)numerical[0.53, 100]
1 level 1: bases, level 2: equipment, level 3: years.
Table 3. Machine learning methods used with their abbreviations.
Table 3. Machine learning methods used with their abbreviations.
Time SeriesRegressionEnsemble
Arithmetic Mean (am)Linear Regression (lr)Random Forest (rf)
Simple Exponential Smoothing (ses)Decision Tree (dt)AdaBoost (ada)
Weighted Moving Average (wma)Ridge Regression (ridge)Gradient Boosting Regressor (gbr)
Bayesian Ridge (br)Extreme Gradient Boosting (xgboost)
Lasso Regression (lasso)Light Gradient Boosting Machine (lightgbm)
ElasticNet (en)CatBoost Regressor (catboost)
Table 4. Comparison of the military vehicle demand prediction performance of different data representations and prediction models. The number is the average and standard deviation of accuracy across five cross-validation trials. Bold indicates the best model with the highest min–max accuracy.
Table 4. Comparison of the military vehicle demand prediction performance of different data representations and prediction models. The number is the average and standard deviation of accuracy across five cross-validation trials. Bold indicates the best model with the highest min–max accuracy.
Data
Representation
Prediction ModelMin-MaxRMSEMAE
MeanStdMeanStdMeanStd
Tabular dataTime SeriesArithmetic Mean0.343 0.015 486.92 84.67 110.98 10.05
Simple Exponential Smoothing0.486 0.010 249.15 63.51 56.47 9.36
Weighted Moving Average0.449 0.012 389.38 77.37 83.57 9.39
RegressionLinearRegression0.333 0.043 140.57 24,799.10 41.16 1371.85
DecisionTree0.629 0.024 166.41 40.27 36.40 5.97
Ridge0.332 0.044 140.62 52.31 41.25 6.76
BayesianRidge0.334 0.044 140.62 26.69 41.09 4.70
Lasso0.230 0.022 179.27 28.70 47.59 4.37
ElasticNet0.295 0.024 213.86 39.06 52.70 6.17
EnsembleRandomForest0.545 0.021 148.61 37.62 34.58 6.09
AdaBoost0.094 0.031 305.53 1622.54 182.08 898.37
GradientBoosting0.388 0.025 154.79 29.27 34.17 4.10
XGBoost0.506 0.016 133.17 33.40 31.42 5.36
LGBM0.468 0.019 150.36 30.80 32.42 4.89
CatBoost0.493 0.006 156.24 37.63 33.51 5.47
ImagesIGTD (2D images)2D-CNN0.171 0.006 485.45 157.85112.58 28.71
Tab2Vox (3D images)3D-CNN0.6310.014 518.81 160.65 92.28 22.86
Table 5. Clustering result on four item groups and their criteria.
Table 5. Clustering result on four item groups and their criteria.
Item GroupClustering RulesNumber of Items
Item Cost (IC)Annual Demand (AD)
Item group 1<8M KRW<1.3433
Item group 2 1.319
Item group 3 44
Item group 4 8M KRW-2
Table 6. Notation for model selection.
Table 6. Notation for model selection.
SymbolDescription
Sets
Set of item groups = {0, 1, …, n}
Set of prediction models = {0, 1, …, 12}
Parameters
c i Number of items in item group i ,   i     I
a c c i j Prediction accuracy of model j in item group i ,   i     I ,     j     J
s t d i j Standard deviation between the prediction accuracies of 5-fold test sets in model j in item group i ,   i     I ,     j     J
t j Run time of model j ,   j     J
c Total number of test items
T Total run time
Decision variables
x i j Binary decision variable, equal to 1 if the item group i   I selects the prediction model j   J .
Table 7. Min–max accuracy for optimal item group and optimal model combination.
Table 7. Min–max accuracy for optimal item group and optimal model combination.
ClusterNumber of
Items
adabrcatboostdtengbrlassolightgbmlrrfridgexgboostTab2vox
14330.0740.3220.4880.6300.2830.3770.2120.4600.3220.5410.3190.4980.655
2190.5750.6310.6410.6060.5640.6280.5930.6380.6310.6340.6300.6590.189
340.0280.2750.3660.7670.2180.4850.3850.5560.2220.6440.2590.6710.435
420.0370.2710.4820.3860.2990.3360.4970.4150.2250.3330.4250.4220.041
Total average min–max accuracy: 65.58%.
Table 8. Optimal prediction model combination and prediction accuracy table considering the number of item groups.
Table 8. Optimal prediction model combination and prediction accuracy table considering the number of item groups.
Model CombinationsNumber of Item Groups
(A)
Number of Models
(B)
Workload
( A × B )
Min–Max Accuracy
tab2vox + dt2240.6358
tab2vox + dt + xgboost3390.6553
tab2vox + dt + xgboost + lasso4 (optimal)4160.6558
tab2vox + dt + xgboost + lasso + br55250.6560
tab2vox + dt + xgboost + lasso + br65300.6680
tab2vox + dt + xgboost + lasso + br75350.6683
tab2vox + dt + xgboost + lasso + br85400.6685
tab2vox + dt + xgboost + lasso + br95450.6685
tab2vox + dt + xgboost + lasso + br105500.6685
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, E.; Nam, M.; Lee, H. Tab2vox: CNN-Based Multivariate Multilevel Demand Forecasting Framework by Tabular-To-Voxel Image Conversion. Sustainability 2022, 14, 11745. https://doi.org/10.3390/su141811745

AMA Style

Lee E, Nam M, Lee H. Tab2vox: CNN-Based Multivariate Multilevel Demand Forecasting Framework by Tabular-To-Voxel Image Conversion. Sustainability. 2022; 14(18):11745. https://doi.org/10.3390/su141811745

Chicago/Turabian Style

Lee, Euna, Myungwoo Nam, and Hongchul Lee. 2022. "Tab2vox: CNN-Based Multivariate Multilevel Demand Forecasting Framework by Tabular-To-Voxel Image Conversion" Sustainability 14, no. 18: 11745. https://doi.org/10.3390/su141811745

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop