Next Article in Journal
Unveiling the Dynamics of Residential Energy Consumption: A Quantitative Study of Demographic and Personality Influences in Singapore Using Machine Learning Approaches
Previous Article in Journal
The Application of Machine Learning and Deep Learning in Intelligent Transportation: A Scientometric Analysis and Qualitative Review of Research Trends
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Proposing Machine Learning Models Suitable for Predicting Open Data Utilization

1
Graduate School of Management of Technology, Sungkyunkwan University, 2066 Seobu-ro, Jangan-gu, Suwon 16419, Republic of Korea
2
Department of Systems Management Engineering, Sungkyunkwan University, 2066 Seobu-ro, Jangan-gu, Suwon 16419, Republic of Korea
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(14), 5880; https://doi.org/10.3390/su16145880
Submission received: 14 June 2024 / Revised: 5 July 2024 / Accepted: 7 July 2024 / Published: 10 July 2024
(This article belongs to the Section Economic and Business Aspects of Sustainability)

Abstract

:
As the digital transformation accelerates in our society, open data are being increasingly recognized as a key resource for digital innovation in the public sector. This study explores the following two research questions: (1) Can a machine learning approach be appropriately used for measuring and evaluating open data utilization? (2) Should different machine learning models be applied for measuring open data utilization depending on open data attributes (field and usage type)? This study used single-model (random forest, XGBoost, LightGBM, CatBoost) and multi-model (stacking ensemble) machine learning methods. A key finding is that the best-performing models differed depending on open data attributes (field and type of use). The applicability of the machine learning approach for measuring and evaluating open data utilization in advance was also confirmed. This study contributes to open data utilization and to the application of its intrinsic value to society.

1. Introduction

Since the onset of the COVID-19 pandemic, digital transformation has accelerated in society, and open data has garnered attention as a key resource in digital conversion within the public sector [1]. Concurring with this trend, discussions regarding the development of information and communications technology have accelerated [2], and the potential of open data, as a digital asset and resource, to contribute to social and economic growth through sustainable value creation has been mentioned [3]. In the context of sustainable environmental, social, and governance management, Helbig et al. [4] noted that open data use can provide significant strategic value to organizations and increase business efficiency.
In response, governments—as key data providers—worldwide have pursued policies for open government data. Since the initiation in the United Kingdom and the United States of America of the development of open government policies aimed at open data openness and transparency enhancements in 2009, many countries, including South Korea, have endeavored to utilize data-based innovations to increase transparency [5,6,7]. According to evaluations by the Organization for Economic Co-operation and Development on the Open-Useful-Reusable DATA index, South Korea achieved the top position for four years (i.e., 2015, 2017, 2019, and 2023), scoring 0.91 points (out of 1) in 2023. As of March 2024, the South Korean government has facilitated the opening of over 87,000 diverse datasets from 1031 organizations through the Open Data Portal (data.go.kr).
However, there are limitations to integrating the quantitative outcomes of open government data with qualitative utilization outcomes [8]. The intrinsic value of open data lies not merely in the openness of the data itself but in the added value that can be derived from its application to businesses. This means that open data can generate substantial value when utilized by end users [9], making efforts to assess and improve open data value and to render users aware of open data utility critical endeavors [10,11,12].
Various studies have been conducted on open data utilization, but most have focused on indirect and complementary factors (e.g., legal frameworks, governance, policy, and technology) rather than on the intrinsic value of open data utilization. For example, there are various studies on open data legal frameworks and governance, which have emphasized the role and structure of relevant legislation in our rapidly changing society, highlighting the need for well-established legal frameworks [13,14,15]. In another study, trust in open data governance was emphasized as a key factor in open data utilization [16]. In the context of South Korea’s open data utilization policies, a study compared the open data policies of different administrations, finding an initial focus on the openness and scalability of open data utilization in diverse areas [17]. Research on open data technology has also provided technical recommendations for the enhancement of platform infrastructure, along with insights for analytical tools and technical standards [10,18,19,20,21,22]. Thus, previous studies have primarily offered results relevant for post-prescriptive alternatives regarding open data utilization, whereas systematic approaches to measure and evaluate the intangible value of open data utilization have not yet been established, and our understanding of how to proactively tackle the matter is limited.

2. Related Research

2.1. Research on Open Data Utilization

Schumpeter et al. [23] argued that innovations such as new products, production methods, markets, and economic organizations are both directly and indirectly related to open data utilization [24,25]. Meanwhile, open data has been recognized as a resource for innovation that can create additional value based on its intrinsic characteristics [26,27]. Various studies have confirmed that active open data use is beneficial for researchers, companies, and other stakeholders, including for creating new businesses and supporting alternative decision-making [28,29]. There are also many explorations of the potential of open data utilization to foster innovation in various sectors of private enterprise products and services [25,30]. Regarding the economic impact of open data, the estimates point to tens of billions of euros annually [31].
According to the South Korean government, among the companies that utilized open data and were surveyed (n = 1003), a considerable number (63.7%) reported that open data plays a vital role in their business [32]. Janssen et al. [33] further demonstrated that open data utilization can activate data–business linkages, support decision-making, and enhance business quality. This points to the indispensability of open data utilization as a trend for leading nations and companies in the era of artificial intelligence. In addition, research on legal frameworks and governance regarding open data utilization notes that while open data should contribute to a range of social and political goals, the release of government data containing personal information can threaten people’s privacy and related rights and interests; accordingly, there have been proposals for frameworks that consider privacy risks in open data utilization in the public sector [34]. Thompson et al. [35] suggested organizational governance adjustments tailored to unique situations and emphasized the need for strong partnerships between information technology professionals and data specialists when actively using open data.
In a policy study related to open data utilization, Zuiderwijk et al. [36] noted that governments aim to promote and induce data release and gain benefits from data utilization when establishing open data policies. These cited authors also developed a framework for comparing individual policies based on the open data policies of Dutch government agencies. Meanwhile, Bertot et al. [37] noted the need for a robust data sharing and interoperability framework as big data and open data are being increasingly exchanged in real time.
Research on open data utilization-related technology is also underway. Regarding technological infrastructure (e.g., data platforms), Máchová et al. [38] evaluated national open data portal usability based on data explorability, accessibility, and reusability. Osagie et al. [19] noted the extensibility of open data platforms and suggested improvements to enable citizens and civil society organizations to effectively utilize open data. Furthermore, research has been conducted on other technical aspects related to open data utilization, such as data quality, with Vetrò et al. [39] pointing out that opening data without proper quality management could diminish data reuse and influence negative usability.
In summary, the literature presents trends toward the advancement of open data utilization from diverse perspectives. The major limitation of the related studies is that they only deliver indirect alternatives for measuring and evaluating open data utilization value. Therefore, this study proposes a discriminative alternative that directly measures and evaluates the intangible value of open data utilization based on the characteristics of the open data managed by the South Korean government.

2.2. Research on Machine Learning Application

In this study, we focused on machine learning as a quantitative tool for evaluating open data utilization. Machine learning is already being applied in various fields related to everyday life, such as the medical sector [40,41,42], environmental sector [43,44,45,46], and construction industry [47]. The introduction of machine learning into these fields has greatly expanded the scope and number of insights as to the application of theoretical knowledge in the real world. Machine learning has also customarily been used as a tool to predict intangible values (e.g., through risk assessment) in the manufacturing industry [48,49].
Regarding the use of machine learning as a tool, several studies have used machine learning to, for instance, predict citation counts as a direct indicator of patent utilization (i.e., high citation counts indicate highly utilized patents). Indeed, patents in the top 1% of citation counts are defined as impactful breakthrough inventions relevant to commercialization and future technology development [50]. Researchers have also used machine learning to classify utility through classification algorithms such as SOM, KPCA, and SVM [51]; predict patent citation counts using boosting algorithms (e.g., the XGBoost classifier) [52]; and predict utilizable patents for research and development investment decisions in companies [53]. Therefore, machine learning tools can be considered appropriate to measure and evaluate the utilization value of intangible assets such as patents.
There has also been considerable research on training data for machine learning. Concerning the characteristics of training data, related studies have mentioned that tree-based algorithm utilization is appropriate when including a large amount of categorical data in the training dataset [54,55]. For example, a study related to construction waste generation suggested that the development of machine learning prediction models for small-scale datasets composed of categorical variables could be improved by applying random forest and GBM algorithms [54,55]. Corroborating these assertions, another research [56] mentioned the superiority of random forest over GBM in terms of stability and accuracy for small-scale datasets composed of categorical variables, demonstrating performance differences that reflect data characteristics.
Regarding training data size, it is generally known from previous research that large-scale (vs. small-scale) training datasets provide better performance. However, Ramezan et al. [57] demonstrated performance differences depending on the algorithm as the training data size was adjusted, highlighting the importance of specific situational considerations. Based on prior research on training data, this study attempted to compare model performance by distinguishing between training based on an integrated dataset without field distinctions and on 16 field-specific datasets. The goal was to confirm the potential performance differences according to the training data size across the 16 fields.
There is a plethora of research on machine learning algorithm performance. A comparison of the SVM, random forest, and ELM algorithms when predicting intrusion detection showcased the superiority of the ELM model in terms of performance [58]. In research related to early diabetes diagnosis [59], after analyzing 66,325 patient records based on 18 risk factors, logistic regression was deemed the superior model. Moreover, when predicting the bending strength of steel structures and the bonding strength of surfaces [60], the ANN algorithm seemed to have superior performance based on a comparison with other algorithms (i.e., random forest, XGBoost, and LightGBM). Research [61] predicting the residual value of construction machinery compared the coefficients of determination for LightGBM, XGBoost, and MDT, reporting that the MDT algorithm was the best prediction model (accuracy of 0.9284).
Concerning machine learning stacking ensembles (i.e., a method of combining individual prediction results to enhance final performance), a study predicting corporate bankruptcy used basic models (e.g., KNN, Decision Tree, SVM, and random forest) based on the financial data of companies to show that the stacking ensemble model with LightGBM achieved an accuracy of over 97% [62]. In a study predicting harmful algal blooms (HABs), the base models of XGBoost, LightGBM, and CatBoost and linear regression were used to construct a metamodel, which in turn confirmed the applicability of the stacking ensemble technique. Based on these previous studies, the current research conducts comparative analyses of single models and stacking ensemble techniques to enhance model performance.

3. Materials and Methods

3.1. Data

In this study, we delimited our scope to the structured data provided by the South Korean government from 2012 to 2022, which is openly available at the Open Data Portal (https://www.data.go.kr). We collected and analyzed metadata through the download method (File Data) and the API method (OpenAPI Data), excluding unusable variables such as contact information (e.g., the responsible person’s name and phone number). The metadata includes detailed information, including related fields and data descriptions. This study considered that the amount of data varied across the 16 fields depending on the scope of the South Korean government’s administrative work; thus, the experiments were conducted while dividing the training dataset into an integrated dataset without field division and 16 field datasets with field division. We evaluated the performance of the models separately for each case. The utilized data consisted of metadata for 44,648 File Data and 6677 OpenAPI Data after removing duplicates and missing values. The modeling was performed using K-fold cross validation (k = 5, train set:test set = 8:2) and confirmed the overfitting of the model. Also, Python 3 based package was employed to identify duplicate data using the list key of open data and applied mode imputation, which fills missing values with the mode (most frequent value). This method was chosen due to the abundance of categorical data in our training dataset.

3.1.1. Input Variables

This study separated the training datasets of File Data and OpenAPI Data depending on the utilization method. The File Data metadata comprised 37 variables, among which 23 were continuous and 14 were categorical variables. The OpenAPI Data metadata were accessible in real-time via API calls, and comprised 42 variables, of which 23 were continuous and 19 were categorical variables.

3.1.2. Target Variables

To establish the target variables that quantitatively represent open data utilization [63], we constructed indicators with a normal distribution, which served to consider model performance in the analyses [64,65,66]. We then verified the normal distribution of each target variable. For File Data, we adjusted the number of downloads based on the provision period and considered the number of attachments, as shown in Equation (1). The provision period incorporates the concept of patent citation half-life, reflecting an adjustment for the period of utilization in the patent field. For OpenAPI Data, we utilized the number of API calls and API utilization requests as the target variables. Similar to procedures for the File Data dataset, we established target variables considering the service period and confirmed whether they followed a normal distribution, as illustrated in Equation (2) [66].
T a r g e t   v a r i a b l e   o f   F i l e   D a t a = l o g ( D o w n l o a d   c o u n t s p r o v i s i o n   p e r i o d   ×   N u m b e r   o f   a t t a c h m e n t s )
T a r g e t   v a r i a b l e   o f   O p e n A P I   D a t a = l o g ( A P I   c a l l   c o u n t s   ×   A P I   r e q u e s t   c o u n t s s e r v i c e   p e r i o d )

3.2. Proposed Methods

For the machine learning analysis in this study, Intel Xeon® Silver 4116 (CPU) (Intel, Santa Clara, CA, USA), 64 GB (memory), and 1.8 TB (disk) were used, and the average training time was less than 4 h for File Data and OpenAPI Data.

3.2.1. Single-Model Methods

As the training dataset contained many categorical variables, we selected tree-based algorithms (random forest, XGBoost, LightGBM, and CatBoost) for use in this study [54,67,68].
  • Random Forest: Random forest algorithms are primarily composed of decision trees, and the results of these trees are summed to produce a final result that maximizes algorithm performance [69]. The advantage of decision tree analysis is the easy and intuitive understanding it provides as the results are presented in a single tree structure; the disadvantage is its lower predictiveness owing to the consideration of only one predictor when dividing the tree branches. Small data changes can also transform tree composition [69,70]. Therefore, while decision tree analysis has a relatively low bias, it has a high variance error, rendering model generalization more difficult. A machine learning algorithm used to overcome this weakness is random forest, which analyzes and aggregates multiple decision trees to form a forest of randomly sampled decision trees, which is then used to create a final prediction model. Random forest algorithms iteratively create independent decision trees to maximize sample and variable selection randomness, thereby reducing prediction error by lowering variance and sustaining a low bias in the decision tree [69,70]. When using data with multiple explanatory variables, random forest algorithms also provide stability by considering interactions and nonlinearities between the explanatory variables. A visual representation of a random forest model is presented in Figure 1a. The hyperparameters of random forest utilized in this study are n_estimators 500, max_depth 30, min_samples_split 4, min_samples_leaf 2, and max_features sqrt.
  • eXtreme Gradient Boosting (XGBoost): XGBoost is an algorithm proposed by Chen et al. [71] for use with large-scale datasets and is meant to compensate for overfitting issues while improving stability and training speed. XGBoost is known for its performance and effectiveness owing to the implementation of the gradient boost learning technique, which is a well-known technique in machine learning. Specifically, it uses a greedy algorithm to construct the most optimal model and improve weak classifiers, and this occurs while controlling complexity using distributed processing to compute optimal weights; this all serves to minimize learning loss and overfitting. This algorithm can be trained on categorical and continuous data, and each leaf contributes to the final score of the model. Its analysis procedure is as follows: (1) measure the accuracy of the generated tree classifiers; (2) randomly generate strong-to-weak classifiers in each order; and (3) sequentially improve the classifiers to generate a strong tree classifier. XGBoost proceeds to the max_depth parameterized during training and then prunes in reverse if the improvement in the loss function does not reach a certain level [71]. During this process, the model can be pruned to remove unnecessary parts of the tree classifier and prevent overfitting. A visual representation of the XGBoost algorithm is presented in Figure 1b. The hyperparameters of XGBoost utilized in this study are n_estimators 300, max_depth 5, learning_rate 0.2, subsample 0.8, and colsample_bytree 1.0.
  • Light Gradient Boosting Machine (LightGBM): developed by Microsoft, this model uses the leaf-wise partitioning method to create highly accurate models [72]. It is based on the gradient-boosting decision tree ensemble learning technique, which has the advantage of dividing the branches at each node based on the best-fit nodes. It uses this learning technique with various algorithms to reduce the number of dimensions of individual data. This technique uses level-wise training for horizontal growth and the traverse of the nodes of the decision tree, preferentially from the root node. For vertical growth, it splits at the node with the largest maximum delta loss, assuming that the loss can be further reduced by growing the same leaf. Furthermore, the two methods used in LightGBM to reduce the number of samples and features are gradient-based one-side sampling and exclusive feature bundling. Gradient-based one-side sampling is an under-sampling technique guided by the training set’s skewness, considering that samples with a larger skewness in absolute value contribute more to learning; accordingly, those with a smaller gradient are randomly removed. A visual representation of the LightGBM algorithm is shown in Figure 1c. The hyperparameters of LightGBM utilized in this study are n_estimators 300, max_depth 5, learning_rate 0.2, subsample 0.8, and colsample_bytree 1.0.
  • Categorical Boosting (CatBoost): This is a library based on gradient boosting [73]. CatBoost performs well with categorical data [74] and processes them using the statistics of the target values while converting each categorical variable to a number. Thus, it is a great performer for most machine learning tasks that require categorical data processing [75]. In a past study, CatBoost performed better than other gradient-boosting libraries because of its ability to handle categorical variables and its optimized algorithm [74]. As aforementioned, it converts categorical variables into numbers using various methods, implying the non-need for the preprocessing of categorical data and the possibility of directly processing it using this algorithm [73]. CatBoost also uses multiple strategies to avoid overfitting, which is a common problem in gradient boosting [75]. This algorithm is hence primarily used as a tool for solving classification and regression problems, performing particularly well on categorical data-related problems [76]. A visual representation of the CatBoost algorithm is presented in Figure 1d. The hyperparameters of CatBoost utilized in this study are iterations 300, max_depth 10, learning_rate 0.2, subsample 0.8, and colsample_bytree 1.0.

3.2.2. Multi-Model Method (Stacking Ensemble)

In this study, we utilized the multi-model stacking ensemble method in addition to the single-model method. Stacking ensembles are constructed by combining various different single models to achieve better performance, enabling the use of the strengths of each algorithm and compensation for their corresponding weaknesses. That is, it aims to create a better-performing model by combining different models [77]. In this study, single models (random forest, XGBoost, LightGBM, and CatBoost) were used to form a stacking ensemble, and GBM was used as the metamodel for the final prediction [66,78,79]. Figure 2 shows a diagram of the stacking ensemble and multi-model method utilized in this study.

3.3. Model Performance Evaluation

To evaluate model performance, we utilized the mean squared error (MSE) and root mean squared error (RMSE) metrics to measure error size (i.e., the difference between the predicted and actual values). Because MSE squares the difference between the predicted and actual values, it is sensitive to outliers, meaning that if the predicted value differs from the actual value by a large amount, the difference that is yielded will be relatively large. We hence decided to also implement RMSE to compare the tendency of the effect on outliers. The formula used is shown in Equation (3).
RMSE is an indicator of scale-dependent errors and is organized as shown in Equation (4). It tends to increase when the magnitude of the value to be predicted is large and decrease when the magnitude of the value to be predicted is small. This metric is used to evaluate the prediction performance of a model by calculating the square root of the squared error [80], and in so doing, it has the same units as the error value [81]—unlike MSE. Thus, the metric can give researchers an intuitive idea of the average size of the error between the predicted and true values [80]. Both MSE and RMSE are widely used in regression models in machine learning to evaluate model predictive performance [81]. Specifically, the smaller the MSE and RMSE, the better the model’s prediction performance. We conducted a comparative analysis of MSE and RMSE between the single-model and multi-model methods. The workflow for model performance evaluation is illustrated in Figure 3.
M S E = 1 n i = 1 n ( y i y ^ i ) 2 ( y i = p r e d i c t e d v a l u e , y ^ i = a c t u a l v a l u e )
R M S E = 1 n i = 1 n ( y i y ^ i ) 2 ( y i = p r e d i c t e d v a l u e , y ^ i = a c t u a l v a l u e )

4. Results

4.1. File Data

In the File Data, similar trends were observed for MSE and RMSE (Table 1, Table 2 and Table 3). The range of MSE was 0.256–0.793 for the single-model and 0.314–0.763 for the multi-model method; for RMSE, the ranges were 0.506–0.891 and 0.560–0.873, respectively. Regarding the single-model method, the following ranges were identified for the algorithms: for random forest, MSE from 0.530 to 0.793 and RMSE from 0.592 to 0.891; for XGBoost, 0.313–0.754 and 0.560–0.868, respectively; for LightGBM, 0.313–0.736 and 0.559–0.858, respectively; for CatBoost, 0.256–0.678 and 0.506–0.824, respectively.
Regarding the performance of the single-model and multi-model methods, the single-model method excelled in thirteen fields, whereas the multi-model method showed superior performance in three fields. For the single-model method, CatBoost exhibited superior performance across all 13 fields. When the single-model method included CatBoost trained on the integrated data, the model was the best performing in 11 fields (i.e., public administration, education, transportation and logistics, land management, agriculture and fisheries, law, social welfare, industry and employment, food and health, unification and diplomacy, environment and meteorology); when trained on field-specific data, it showed the best performance in the science and technology and finance fields. For the multi-model method, it showed superior performance in three fields (culture and tourism, healthcare, disaster and safety) when trained on integrated data.
Comparing the performance of each field based on MSE and RMSE, the field with the best performance was law (MSE, 2.533; RMSE, 1.592), while the field with the poorest performance was transportation and logistics (MSE, 14.204; RMSE, 3.769). In the law field, the best performance was observed with the multi-model method trained on integrated data. In the transportation and logistics field, the best performance was observed with the single-model method with LightGBM trained on integrated data.
When comparing the performance of models within the same field, the law field showed the largest performance difference (among the 16 fields) across models. Specifically, based on RMSE, the performance of the best model was approximately 4.378 times superior to that of the poorest model; based on MSE, the performance of the best model was approximately 19.166 times superior. Meanwhile, the public administration field showed the smallest performance difference across models. When considering RMSE, the performance of the best model was approximately 1.089 times superior to that of the poorest model; according to MSE, it was approximately 1.185 times superior. A summary of the superior algorithms, training data, and performance metrics (MSE and RMSE) for each field is presented in Table 1. The scatter plots of the File Data are shown in Figure A1.
In analyzing the impact of File Data features, the data size feature had the most significant effect on the predicted value (target). The narrow distribution of the SHAP values for the data size feature in the positive region and the wide distribution in the negative region indicate that the data size feature does not significantly increase the predicted value. In other words, the utilization of File Data did not increase with the size of the File Data, but there was a clear relationship between the size of the File Data and the decrease in data utilization. Additionally, the distribution of blue points in the negative region of the SHAP values suggests that the predicted value decreases as the data size feature decreases.
For the open data provider feature, the color of the dots (increasing/decreasing) as a nominal variable is not meaningful. The wide distribution of SHAP values in both directions indicates that the predictive value can either increase or decrease depending on the institution’s name. Regarding the data core keyword count feature, it can be interpreted that the predicted value tends to increase. The distribution of red dots in the positive region suggests a positive relationship between the increase in the data core keyword count feature and the increase in the predicted value. The summary plot related to the SHAP analysis results is presented in Figure A3.
Also, 5-fold cross-validation was conducted to evaluate the performance of the single-model. The results indicate that single models achieve high consistency. Based on RMSE, the values of random forest (0.702 ± 0.00618), XGBoost (0.684 ± 0.00273), LightGBM (0.694 ± 0.00520), and CatBoost (0.688 ± 0.00476) were confirmed. Also, the value of the multi-model means high consistency as 0.679 ± 0.00464.

4.2. OpenAPI Data

For OpenAPI Data, similar trends were observed for the MSE and RMSE metrics (Table 4, Table 5 and Table 6), as occurred for the File Data. The MSE ranges were 2.906–48.547 for the single-model and 2.533–35.837 for the multi-model method, whereas the RMSE ranges were 1.705–6.968 and 1.592–5.986, respectively. Regarding the single-model method, the following ranges were identified for the algorithms: for random forest, MSE from 2.906 to 29.163 and RMSE from 1.705 to 5.400; for XGBoost, 4.239–48.547 and 2.059–6.968, respectively; for LightGBM, 4.259–26.606 and 2.064–5.158, respectively; for CatBoost, 4.332–28.268 and 2.081–5.317, respectively. The difference between the results from OpenAPI Data and File Data was that there was a wider range of superior model algorithms in OpenAPI Data. Regarding the performance of the single-model and multi-model methods, the single-model method excelled in thirteen fields, while the multi-model method showed superior performance in three fields. For the single-model method, the best performance was achieved in six fields (public administration, science and technology, culture and tourism, disaster safety, finance, unification and diplomacy) with the random forest algorithm trained on field-specific data.
For the education and industry and employment fields, the best performance was achieved with the XGBoost model trained on integrated data. For the transportation and logistics field, the best performance was achieved with the LightGBM model trained on integrated data. For the land management and food and health fields, the best performance was achieved with the LightGBM model trained on field-specific data. In the social welfare field, the best performance was achieved with the CatBoost model trained on field-specific data. In the environment and meteorology field, the best performance was achieved with the CatBoost model trained on integrated data. In three fields (agriculture and fisheries, law, healthcare), the best performance was achieved using the multi-model method trained on integrated data.
Comparing the performance of each field based on MSE and RMSE, the field with the best performance was law (MSE, 0.256; RMSE, 0.506), and that with the poorest performance was science and technology (MSE, 0.516; RMSE, 0.718). In the law field, the best performance was achieved with the CatBoost algorithm trained on integrated data; in the science and technology field, the best performance was achieved when employing the CatBoost algorithm trained on field-specific data.
When comparing the performance of models within the same field, the disaster and safety field showed the largest performance difference (among the 16 fields) across models. Specifically, when considering RMSE, the performance of the best model was approximately 1.343 times superior to that of the poorest model; based on MSE, it was approximately 1.802 times superior. Meanwhile, the agriculture and fisheries field showed the smallest performance difference (among the 16 fields) across models. When considering RMSE, the performance of the best model was approximately 1.074 times superior to that of the poorest model; according to MSE, it was approximately 1.155 times superior. A summary of the superior algorithms, training data, and performance metrics (MSE and RMSE) for each field is presented in Table 4. The scatter plots of the File data by field are shown in Figure A2.
For OpenAPI Data, the scope of data use permission can be interpreted as the most influential variable, indicating that utilization increases or decreases depending on the scope of data use permission. The distribution of red dots in the positive region and blue dots in the negative region indicates that the predicted value increases as the scope of data use permission increases and decreases as the scope of data use permission decreases.
Regarding the open data provider feature, the color of the dot (increasing/decreasing) is meaningless as a nominal variable. The wide distribution of SHAP values on both sides suggests that the predicted value can either increase or decrease depending on the institution’s name. For the open data center API status feature, it can be interpreted that the predicted value is relatively higher if the API is provided by the open data center.
In the case of the data description null count feature, the wide distribution in both negative and positive areas indicates its impact on changing the predicted value. In other words, with red dots distributed in the negative area and blue dots in the positive area, it can be inferred that more missing values decrease the predicted value, while fewer missing values increase it. The summary plot related to the OpenAPI Data feature results is presented in Figure A4.
In addition, 5-fold cross-validation was conducted to evaluate the performance of the single-model. The results indicate that single models achieve high consistency. Based on RMSE, the values of random forest (3.075 ± 0.0613), XGBoost (2.956 ± 0.0761), LightGBM (2.975 ± 0.0926), and CatBoost (2.977 ± 0.0978) were confirmed. Also, the value of multi-model means high consistency as 2.870 ± 0.0861.

5. Discussion

Using open data metadata accumulated from 2012 to 2022 in South Korea, this study applied machine learning techniques to construct predictive models and proposed an alternative approach for evaluating open data utilization in advance. Both single-model (random forest, XGBoost, LightGBM, CatBoost) and multi-model (stacking ensemble) methods were applied. Considering the attributes of the open data used (fields and utilization methods), the training data were selectively utilized as integrated and field-specific data. The results showed that model and method (i.e., single-model and multi-model methods) superiority varied by data attributes. This finding aligns with the research trends and outcomes mentioned by Si et al. [82], emphasizing the importance of considering data attributes in data analysis, as different attributes can influence model performance.
Regarding the implications of distinguishing between two open data utilization methods (i.e., File Data and OpenAPI Data), we observed that the distribution of the target variables was broader in the OpenAPI Data and that the algorithms exhibiting superior performance were more diverse. This corresponds to evidence in prior research [83,84,85,86], which showcases that different models perform better depending on the characteristics of the independent and dependent variables in the data. When using File Data and employing the single-model method, the best performance was achieved using CatBoost in 13 (of 16) fields. This result can be interpreted in light of previous studies [73,87] and suggests the specialized performance of CatBoost in handling categorical data. Additionally, according to the MSE and RMSE metrics, the File Data generally demonstrated superior performance compared with OpenAPI Data. We also observed only a relatively small deviation of the RMSE metrics within each field when using File Data; nevertheless, when using OpenAPI Data, there were significant differences in the performance metrics across fields. For example, when applying the random forest model trained on integrated data in the public administration field using File Data, the difference between the RMSE maximum value of 0.745 and the minimum value of 0.600 was 0.145; this was smaller than the difference of 3.329 for the same conditions when using OpenAPI Data.
Regarding the implications of distinguishing open data by field, the performance of the models trained with integrated data and field-specific data differed because of variations in the quantity of accumulated data and field-specific metadata among the 16 fields. For File Data, better performance was generally achieved when the model was trained with integrated data—with the exception of the science and technology and finance fields. For OpenAPI Data, superior performance was observed when the model was trained with integrated data in eight fields (education, transportation and logistics, agriculture and fisheries, culture and tourism, law, healthcare, industry and employment, environment and meteorology); in the other eight fields (public administration, science and technology, land management, social welfare, food and health, disaster and safety, finance, unification and diplomacy), models trained on field-specific data showed superior performance. These research findings suggest that OpenAPI Data, which are often utilized in real-time and continuous services, can more effectively reflect the characteristics of field-specific data compared with File Data, and this is corroborated by past research [88].
The machine learning model developed through this study is expected to be applicable for evaluating and predicting the various intangible values (e.g., brand, technology, human resources). If this understanding is applied to open data openness and utilization in practical, real-world scenarios and the approach we propose for pre-evaluating and diagnosing open data utilization is implemented, it may help address the ongoing garbage data issues [89,90] related to open data. These findings were expected to serve as a catalyst for accelerating the process of unveiling the details of open data utilization.

6. Conclusions

This study applied machine learning methods to propose an alternative approach for the proactive quantification and evaluation of open data utilization. Regarding academic significance, this research empirically confirmed that building multiple machine learning models and comparing their performance is useful to measure the intangible value of open government data utilization. In so doing, this study overcomes the limitations of previous related studies and expands the horizons of open data utilization measurement, delivering a novel alternative methodology for such procedures. Additionally, this study delivers evidence showing that it is appropriate to consider the attributes of open data (fields and utilization methods) when deciding on which algorithm to apply and training data to use for machine learning approaches.
Regarding practical significance, the proposed approach can be applied in efforts to increase real-world open data utilization. Specifically, its use may enable stakeholders to accurately pinpoint, in advance, the data that they need to disclose to secure high usability for the open data that are made available. From the perspective of consumers, the tool supporting the provision of highly usable open data may then help with the creation of various business opportunities.
Regarding policy implications, the alternative approach proposed in this study may allow for a policy focus shift. In particular, while current open data policies focus on “quantity expansion”, the assessment of open data utilization before data provision may make possible a greater focus on “quality enhancement” in related policies. We also suggest that those involved consider these findings in light of a comprehensive consideration of the indirect factors (law, governance, policy, and technology) emphasized in prior research to influence open data utilization.
Regarding limitations, the machine learning algorithms utilized in this study were tree-based algorithms, and thus other algorithm types (e.g., neural network-based algorithms) were not examined. In addition, for the target indicators of open data utilization, we have established and utilized normalized indicators based on the number of downloads for File Data and the number of applications and calls for OpenAPI Data; there are limitations in the use of these indicators for qualitative analyses focused on tracking and understanding how open data are being used in businesses from an outcome perspective. This limitation may be addressed in the future if digital rights management is applied to open data, as it may enable open data utilization tracking [91]. Furthermore, since we used the official classification system for the public sector in South Korea to classify the 16 fields related to open data utilization, the scalability of the proposed method is limited. Further research should consider open data attributes beyond those analyzed in this study. In addition, although this study confirmed the normalization of the input variable distribution based on prior research, additional feature engineering such as feature selection (e.g., the Boruta algorithm) should be considered.
Moreover, this study can aid in advancing sustainable industrialization and fostering innovation within the framework of the Sustainable Development Goals (SDGs) by advocating for enhanced usability of open data. High-quality and easily accessible open data can serve as a catalyst for innovation across all sectors, thereby contributing to the sustainability of industries. By demonstrating the feasibility of evaluating data usability through metadata, this research offers valuable insights for the development of data governance frameworks. Also, realizing these contributions necessitates coordinated policy efforts across borders.
In terms of sustainability impact, this research enhances the resilience of businesses leveraging open data by introducing a method to assess open data utilization. This approach offers strategic insights to enhance business efficiency and underscores the importance of open data in shaping sustainable practices. Moreover, it proposes an alternative assessment approach that can inform long-term open data policies, shifting focus from mere quantity expansion to quality enhancement. Also, this study promotes practical implications by fostering a sustainable cycle of open data utilization. Our proposed approach empowers open data providers to make informed decisions on data release, thereby boosting open data utilization. On the demand side, increased use of high-quality open data stimulates the creation of diverse business opportunities.
Future research directions include expanding the scope of investigations involving unstructured data (i.e., heavily utilized in high-level artificial intelligence businesses) so that we can predict the utilization of unstructured data. Convergence analyses with data actually containing different open data attributes could also help us overcome the limitations inherent in the metadata used in this study. Finally, researchers are recommended to probe into the relationship between post- and indirect-evaluation factors associated with open data utilization.

Author Contributions

Conceptualization, J.J. and K.C.; methodology, J.J. and K.C.; software, J.J. and K.C.; validation, J.J. and K.C.; formal analysis, J.J. and K.C.; writing—original draft preparation, J.J. and K.C.; writing—review and editing, J.J. and K.C.; supervision, K.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are provided and managed by the South Korean government in the Open Government Data portal (https://www.data.go.kr).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. Scatter plot by field using File Data.
Figure A1. Scatter plot by field using File Data.
Sustainability 16 05880 g0a1
Figure A2. Scatter plot by field using OpenAPI Data.
Figure A2. Scatter plot by field using OpenAPI Data.
Sustainability 16 05880 g0a2
Figure A3. Summary plot illustrating impacts of features (File Data).
Figure A3. Summary plot illustrating impacts of features (File Data).
Sustainability 16 05880 g0a3
Figure A4. Summary plot illustrating impacts of features (OpenAPI Data).
Figure A4. Summary plot illustrating impacts of features (OpenAPI Data).
Sustainability 16 05880 g0a4

References

  1. Gabryelczyk, R. Has COVID-19 accelerated digital transformation? Initial lessons learned for public administrations. Inf. Syst. Manag. 2020, 37, 303–309. [Google Scholar] [CrossRef]
  2. Hamari, J.; Sjöklint, M.; Ukkonen, A. The sharing economy: Why people participate in collaborative consumption. J. Assoc. Inf. Sci. Technol. 2016, 67, 2047–2059. [Google Scholar] [CrossRef]
  3. Niankara, I. Sustainability through open data sharing and reuse in the digital economy. In Proceedings of the 2022 International Arab Conference on Information Technology (ACIT), Abu Dhabi, United Arab Emirates, 22–24 November 2022; pp. 1–11. [Google Scholar]
  4. Helbig, R.; von Höveling, S.; Solsbach, A.; Marx Gómez, J. Strategic analysis of providing corporate sustainability open data. Intell. Syst. Accounting, Finance Manag. 2021, 28, 195–214. [Google Scholar] [CrossRef]
  5. Peled, A. When transparency and collaboration collide: The USA open data program. J. Am. Soc. -Form. Sci. Technol. 2011, 62, 2085–2094. [Google Scholar] [CrossRef]
  6. O’Hara, K. Transparency, open data and trust in government: Shaping the infosphere. In Proceedings of the 4th Annual ACM Web Science Conference, New York, NY, USA, 22–24 June 2012; pp. 223–232. [Google Scholar]
  7. Lnenicka, M.; Nikiforova, A. Transparency-by-design: What is the role of open data portals? Telemat. Inform. 2021, 61, 101605. [Google Scholar] [CrossRef]
  8. Hong, Y. A Study on Policies for Activating the Use of Public Data. J. Korean Data Inf. Sci. Soc. 2014, 25, 769–777. [Google Scholar]
  9. Janssen, M.; Charalabidis, Y.; Zuiderwijk, A. Benefits, adoption barriers and myths of open data and open government. Inf. Syst. Manag. 2012, 29, 258–268. [Google Scholar] [CrossRef]
  10. Weerakkody, V.; Irani, Z.; Kapoor, K.; Sivarajah, U.; Dwivedi, Y.K. Open data and its usability: An empirical view from the Citizen’s perspective. Inf. Syst. Front. 2017, 19, 285–300. [Google Scholar] [CrossRef]
  11. Go, K. Study on Value Creation Strategies of Public Data. Proc. Korean Assoc. Public Adm. 2018, 2018, 3473–3491. [Google Scholar]
  12. Yoon, S.O.; Hyun, J.W. A Study on the Current Status Analysis and Improvement Measures of Public Data Opening Policies: Focusing on the Case of National Priority Data Opening in the Public Data Portal. Korean J. Public Adm. 2019, 33, 219–247. [Google Scholar]
  13. Kim, E.-S. A Study on Legal System Improvement Measures for Promoting the Openness and Utilization of Public Data—Focusing on Cases of Refusal to Provide Public Data. Inf. Policy 2023, 30, 46–67. [Google Scholar]
  14. Kim, M.-H.; Lee, B.-O. Trends and Implications of the Revision of the EU Directive on Public Open Data. Sungkyunkwan Law Rev. 2020, 32, 1–30. [Google Scholar]
  15. Devins, C.; Felin, T.; Kauffman, S.; Koppl, R. The law and big data. Cornell JL Public Policy 2017, 27, 357. [Google Scholar]
  16. Tan, E. Designing an AI compatible open government data ecosystem for public governance. Inf. Polity 2023, 28, 541–557. [Google Scholar] [CrossRef]
  17. Kim, G.-H.; Jung, S.-H.; Yang, J.-D.; Wi, J.-Y. A Policy Study on Public Data for the Past 10 Years Using Big Data Analysis Techniques: Focusing on Comparative Analysis by Administration. Natl. Policy Res. 2023, 37, 45–67. [Google Scholar]
  18. Jetzek, T.; Avital, M.; Bjørn-Andersen, N. Generating Value from Open Government Data, ICIS. 2013. Available online: http://aisel.aisnet.org/icis2013/proceedings/GeneralISTopics/5/ (accessed on 15 December 2013).
  19. Osagie, E.; Waqar, M.; Adebayo, S.; Stasiewicz, A.; Porwol, L.; Ojo, A. Usability evaluation of an open data platform. In Proceedings of the 18th Annual International Conference on Digital Government Research, Staten Island, NY, USA, 7–9 June 2017; pp. 495–504. [Google Scholar]
  20. Máchová, R.; Volejníková, J.; Lněnička, M. Impact of e-government development on the level of corruption: Measuring the effects of related indices in time and dimensions. Rev. Econ. Perspect. 2018, 18, 99–121. [Google Scholar] [CrossRef]
  21. Khurshid, M.M.; Zakaria, N.H.; Rashid, A.; Shafique, M.N. Examining the factors of open government data usability from academician’s perspective. Int. J. Inf. Technol. Proj. Manag. (IJITPM) 2018, 9, 72–85. [Google Scholar] [CrossRef]
  22. Hagen, L.; Keller, T.E.; Yerden, X.; Luna-Reyes, L.F. Open data visualizations and analytics as tools for policy-making. Gov. Inf. Q. 2019, 36, 101387. [Google Scholar] [CrossRef]
  23. Schumpeter, J.A. The Theory of Economic Development: An Inquiry into Profits, Capital, Credit, Interest, and the Business Cycle; Transaction Publishers: New Brunswick, NJ, USA, 1934. [Google Scholar]
  24. Bason, C. Leading Public Sector Innovation; Policy Press: Bristol, UK, 2010; Volume 10. [Google Scholar]
  25. Zuiderwijk, A.; Janssen, M.; Davis, C. Innovation with open data: Essential elements of open data ecosystems. Inf. Polity 2014, 19, 17–33. [Google Scholar] [CrossRef]
  26. Blakemore, M.; Craglia, M. Access to public-sector information in europe: Policy, rights, and obligations. Inf. Soc. 2006, 22, 13–24. [Google Scholar] [CrossRef]
  27. Charalabidis, Y.; Zuiderwijk, A.; Alexopoulos, C.; Janssen, M.; Höchtl, J.; Ferro, E. The world of open data. In Public Administration and Information Technology; Springer International Publishing: Cham, Switzerland, 2018; pp. 978–983. [Google Scholar]
  28. European Commission. Digital Agenda: Turning Government Data into Gold; European Commission: Brussels, Belgium, 2011. [Google Scholar]
  29. Zhang, J.; Dawes, S.S.; Sarkis, J. Exploring stakeholders’ expectations of the benefits and barriers of e-government knowledge sharing. J. Enterp. Inf. Manag. 2005, 18, 548–567. [Google Scholar] [CrossRef]
  30. Kitsios, F.; Papachristos, N.; Kamariotou, M. Business models for open data ecosystem: Challenges and motivations for entrepreneurship and innovation. In Proceedings of the 2017 IEEE 19th Conference on Business Informatics (CBI), Thessaloniki, Greece, 24–27 July 2017; pp. 398–407. [Google Scholar]
  31. European Commission. Digital Agenda: Commission’s Open Data Strategy, Questions & Answers; European Commission: Brussels, Belgium, 2013. [Google Scholar]
  32. Ministry of the Interior and Safety. 2021 Administrative Safety White Paper; Ministry of the Interior and Safety: Sejong, Republic of Korea, 2022. [Google Scholar]
  33. Janssen, M.; Zuiderwijk, A. Infomediary business models for connecting open data providers and users. Soc. Sci. Comput. Rev. 2014, 32, 694–711. [Google Scholar] [CrossRef]
  34. Borgesius, F.Z.; Gray, J.; Van Eechoud, M. Open data, privacy, and fair information principles: Towards a balancing framework. Berkeley Technol. Law J. 2015, 30, 2073–2131. [Google Scholar]
  35. Thompson, N.; Ravindran, R.; Nicosia, S. Government data does not mean data governance: Lessons learned from a public sector application audit. Gov. Inf. Q. 2015, 32, 316–322. [Google Scholar] [CrossRef]
  36. Zuiderwijk, A.; Janssen, M. Open data policies, their implementation and impact: A framework for comparison. Gov. Inf. Q. 2014, 31, 17–29. [Google Scholar] [CrossRef]
  37. Bertot, J.C.; Gorham, U.; Jaeger, P.T.; Sarin, L.C.; Choi, H. Big data, open government and e-government: Issues, policies and recommendations. Inf. Polity 2014, 19, 5–16. [Google Scholar] [CrossRef]
  38. Máchová, R.; Lněnička, M. Evaluating the quality of open data portals on the national level. J. Theor. Appl. Electron. Commer. Res. 2017, 12, 21–41. [Google Scholar] [CrossRef]
  39. Vetrò, A.; Canova, L.; Torchiano, M.; Minotas, C.O.; Iemma, R.; Morando, F. Open data quality measurement framework: Definition and application to Open Government Data. Gov. Inf. Q. 2016, 33, 325–337. [Google Scholar] [CrossRef]
  40. Kourou, K.; Exarchos, T.P.; Exarchos, K.P.; Karamouzis, M.V.; Fotiadis, D.I. Machine learning applications in cancer prognosis and prediction. Comput. Struct. Biotechnol. J. 2015, 13, 8–17. [Google Scholar] [CrossRef]
  41. Kim, E.; Lee, Y.; Choi, J.; Yoo, B.; Chae, K.J.; Lee, C.H. Machine Learning-based Prediction of Relative Regional Air Volume Change from Healthy Human Lung CTs. KSII Trans. Internet Inf. Syst. 2023, 17, 576–590. [Google Scholar] [CrossRef]
  42. Kruppa, J.; Ziegler, A.; König, I.R. Risk estimation and risk prediction using machine-learning methods. Hum. Genet. 2012, 131, 1639–1654. [Google Scholar] [CrossRef] [PubMed]
  43. Xayasouk, T.; Lee, H.; Lee, G. Air pollution prediction using long short-term memory (LSTM) and deep autoencoder (DAE) models. Sustainability 2020, 12, 2570. [Google Scholar] [CrossRef]
  44. Mosavi, A.; Ozturk, P.; Chau, K.-W. Flood prediction using machine learning models: Literature review. Water 2018, 10, 1536. [Google Scholar] [CrossRef]
  45. Ahmed, A.N.; Othman, F.B.; Afan, H.A.; Ibrahim, R.K.; Fai, C.M.; Hossain, S.; Ehteram, M.; Elshafie, A. Machine learning methods for better water quality prediction. J. Hydrol. 2019, 578, 124084. [Google Scholar] [CrossRef]
  46. Lee, D.-S.; Choi, W.I.; Nam, Y.; Park, Y.-S. Predicting potential occurrence of pine wilt disease based on environmental factors in South Korea using machine learning algorithms. Ecol. Inform. 2021, 64, 101378. [Google Scholar] [CrossRef]
  47. Zhang, L.; Wen, J.; Li, Y.; Chen, J.; Ye, Y.; Fu, Y.; Livingood, W. A review of machine learning in building load prediction. Appl. Energy 2021, 285, 116452. [Google Scholar] [CrossRef]
  48. Paltrinieri, N.; Comfort, L.; Reniers, G. Learning about risk: Machine learning for risk assessment. Saf. Sci. 2019, 118, 475–486. [Google Scholar] [CrossRef]
  49. Hegde, J.; Rokseth, B. Applications of machine learning methods for engineering risk assessment—A review. Saf. Sci. 2020, 122, 104492. [Google Scholar] [CrossRef]
  50. Ahuja, G.; Lampert, C.M. Entrepreneurship in the large corporation: A longitudinal study of how established firms create breakthrough inventions. Strat. Manag. J. 2001, 22, 521–543. [Google Scholar] [CrossRef]
  51. Wu, J.-L.; Chang, P.-C.; Tsao, C.-C.; Fan, C.-Y. A patent quality analysis and classification system using self-organizing maps with support vector machine. Appl. Soft Comput. 2016, 41, 305–316. [Google Scholar] [CrossRef]
  52. Cho, H.; Lee, H. Patent Quality Prediction Using Machine Learning Techniques. In Proceedings of the Korean Institute of Industrial Engineers Spring Conference, Changwon, Republic of Korea, 25–27 April 2018; pp. 1343–1350. [Google Scholar]
  53. Erdogan, Z.; Altuntas, S.; Dereli, T. Predicting patent quality based on machine learning approach. IEEE Trans. Eng. Manag. 2022, 71, 3144–3157. [Google Scholar] [CrossRef]
  54. Kim, K.; Hong, J.-S. A hybrid decision tree algorithm for mixed numeric and categorical data in regression analysis. Pattern Recognit. Lett. 2017, 98, 39–45. [Google Scholar] [CrossRef]
  55. Cha, G.-W.; Moon, H.-J.; Kim, Y.-C. Comparison of random forest and gradient boosting machine models for predicting demolition waste based on small datasets and categorical variables. Int. J. Environ. Res. Public Health 2021, 18, 8530. [Google Scholar] [CrossRef] [PubMed]
  56. Foody, G.M.; Mathur, A.; Sanchez-Hernandez, C.; Boyd, D.S. Training set size requirements for the classification of a specific class. Remote. Sens. Environ. 2006, 104, 1–14. [Google Scholar] [CrossRef]
  57. Ramezan, C.A.; Warner, T.A.; Maxwell, A.E.; Price, B.S. Effects of training set size on supervised machine-learning land-cover classification of large-area high-resolution remotely sensed data. Remote. Sens. 2021, 13, 368. [Google Scholar] [CrossRef]
  58. Ahmad, I.; Basheri, M.; Iqbal, M.J.; Rahim, A. Performance comparison of support vector machine, random forest, and extreme learning machine for intrusion detection. IEEE Access 2018, 6, 33789–33795. [Google Scholar] [CrossRef]
  59. Daghistani, T.; Alshammari, R. Comparison of statistical logistic regression and random forest machine learning techniques in predicting diabetes. J. Adv. Inf. Technol. 2020, 11, 78–83. [Google Scholar] [CrossRef]
  60. Suenaga, D.; Takase, Y.; Abe, T.; Orita, G.; Ando, S. Prediction accuracy of Random Forest, XGBoost, LightGBM, and artificial neural network for shear resistance of post-installed anchors. Structures 2023, 50, 1252–1263. [Google Scholar] [CrossRef]
  61. Shehadeh, A.; Alshboul, O.; Al Mamlook, R.E.; Hamedat, O. Machine learning models for predicting the residual value of heavy construction equipment: An evaluation of modified decision tree, LightGBM, and XGBoost regression. Autom. Constr. 2021, 129, 103827. [Google Scholar] [CrossRef]
  62. Muslim, M.A.; Dasril, Y. Company bankruptcy prediction framework based on the most influential features using XGBoost and stacking ensemble learning. Int. J. Electr. Comput. Eng. (IJECE) 2021, 11, 5549–5557. [Google Scholar] [CrossRef]
  63. Rebala, G.; Ravi, A.; Churiwala, S.; Rebala, G.; Ravi, A.; Churiwala, S. Machine learning definition and basics. In An Introduction to Machine Learning; Springer: Berlin/Heidelberg, Germany, 2019; pp. 1–17. [Google Scholar]
  64. West, S.G.; Finch, J.F.; Curran, P.J. Structural equation models with nonnormal variables: Problems and remedies. In Structural Equation Modeling: Concepts, Issues, and Applications; Hoyle, R.H., Ed.; Sage Publications, Inc.: Thousand Oaks, CA, USA, 1995; pp. 56–75. [Google Scholar]
  65. Hong, S.; Malik, M.L.; Lee, M.-K. Testing configural, metric, scalar, and latent mean invariance across genders in sociotropy and autonomy using a non-Western sample. Educ. Psychol. Meas. 2003, 63, 636–654. [Google Scholar] [CrossRef]
  66. Kwon, H.; Park, J.; Lee, Y. Stacking ensemble technique for classifying breast cancer. Health Informatics Res. 2019, 25, 283–288. [Google Scholar] [CrossRef] [PubMed]
  67. Painsky, A.; Rosset, S.; Feder, M. Large alphabet source coding using independent component analysis. IEEE Trans. Inf. Theory 2017, 63, 6514–6529. [Google Scholar] [CrossRef]
  68. Qin, X.; Han, J. Variable selection issues in tree-based regression models. Transp. Res. Rec. J. Transp. Res. Board 2008, 2061, 30–38. [Google Scholar] [CrossRef]
  69. Géron, A. Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2022. [Google Scholar]
  70. Dangeti, P. Statistics for Machine Learning; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  71. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  72. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A highly efficient gradient boosting deci-sion tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3146–3154. [Google Scholar]
  73. Hancock, J.T.; Khoshgoftaar, T.M. CatBoost for big data: An interdisciplinary review. J. Big Data 2020, 7, 94. [Google Scholar] [CrossRef] [PubMed]
  74. Wei, X.; Rao, C.; Xiao, X.; Chen, L.; Goh, M. Risk assessment of cardiovascular disease based on SOLSSA-CatBoost model. Expert Syst. Appl. 2023, 219, 119648. [Google Scholar] [CrossRef]
  75. Jabeur, S.B.; Gharib, C.; Mefteh-Wali, S.; Arfi, W.B. CatBoost model and artificial intelligence techniques for corporate failure prediction. Technol. Forecast. Soc. Chang. 2021, 166, 120658. [Google Scholar] [CrossRef]
  76. Luo, M.; Wang, Y.; Xie, Y.; Zhou, L.; Qiao, J.; Qiu, S.; Sun, Y. Combination of feature selection and catboost for prediction: The first application to the estimation of aboveground biomass. Forests 2021, 12, 216. [Google Scholar] [CrossRef]
  77. Jin, Y.; Ye, X.; Ye, Q.; Wang, T.; Cheng, J.; Yan, X. Demand forecasting of online car-hailing with stacking ensemble learning approach and large-scale datasets. IEEE Access 2020, 8, 199513–199522. [Google Scholar] [CrossRef]
  78. Acquah, J.; Owusu, D.K.; Anafo, A.Y. Application of Stacked Ensemble Techniques for Classifying Recurrent Head and Neck Squamous Cell Carcinoma Prognosis. Asian J. Res. Comput. Sci. 2024, 17, 77–94. [Google Scholar] [CrossRef]
  79. Sahin, E.K.; Demir, S. Greedy-AutoML: A novel greedy-based stacking ensemble learning framework for assessing soil liq-uefaction potential. Eng. Appl. Artif. Intell. 2023, 119, 105732. [Google Scholar] [CrossRef]
  80. Aswin, S.; Geetha, P.; Vinayakumar, R. Deep learning models for the prediction of rainfall. In Proceedings of the 2018 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 3–5 April 2018; pp. 657–661. [Google Scholar]
  81. Almalaq, A.; Edwards, G. A review of deep learning methods applied on load forecasting. In Proceedings of the 2017 16th IEEE international conference on machine learning and applications (ICMLA), Cancun, Mexico, 18–21 December 2017; pp. 511–516. [Google Scholar]
  82. Si, B.; Ni, Z.; Xu, J.; Li, Y.; Liu, F. Interactive effects of hyperparameter optimization techniques and data characteristics on the performance of machine learning algorithms for building energy metamodeling. Case Stud. Therm. Eng. 2024, 55, 104124. [Google Scholar] [CrossRef]
  83. Satoła, A.; Satoła, K. Performance comparison of machine learning models used for predicting subclinical mastitis in dairy cows: Bagging, boosting, stacking, and super-learner ensembles versus single machine learning models. J. Dairy Sci. 2024, 107, 3959–3972. [Google Scholar] [CrossRef] [PubMed]
  84. Zhang, H.; Zhu, T. Stacking Model for Photovoltaic-Power-Generation Prediction. Sustainability 2022, 14, 5669. [Google Scholar] [CrossRef]
  85. Park, U.; Kang, Y.; Lee, H.; Yun, S. A stacking heterogeneous ensemble learning method for the prediction of building con-struction project costs. Appl. Sci. 2022, 12, 9729. [Google Scholar] [CrossRef]
  86. Eom, H.; Kim, J.; Choi, S. Verification of Machine Learning-Based Corporate Bankruptcy Risk Prediction Model and Policy Suggestions: Focused on Improvement through Stacking Ensemble Model. J. Intell. Inf. Syst. Res. 2020, 26, 105–129. [Google Scholar]
  87. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
  88. Sugiyama, M. Introduction to Statistical Machine Learning; Elsevier: Amsterdam, The Netherlands, 2016. [Google Scholar]
  89. Grimes, D.A. Epidemiologic research using administrative databases: Garbage in, garbage out. Obstet. Gynecol. 2010, 116, 1018–1019. [Google Scholar] [CrossRef] [PubMed]
  90. Kilkenny, M.F.; Robinson, K.M. Data quality:“Garbage in–garbage out”. Health Inf. Manag. J. 2018, 47, 103–105. [Google Scholar] [CrossRef] [PubMed]
  91. Hartung, F.; Ramme, F. Digital rights management and watermarking of multimedia content for m-commerce applications. IEEE Commun. Mag. 2000, 38, 78–84. [Google Scholar] [CrossRef]
Figure 1. Concept of the single-model algorithms used in this study.
Figure 1. Concept of the single-model algorithms used in this study.
Sustainability 16 05880 g001
Figure 2. Conceptualization of the multi-model method (stacking ensemble) used in this study.
Figure 2. Conceptualization of the multi-model method (stacking ensemble) used in this study.
Sustainability 16 05880 g002
Figure 3. Workflow for predicting open data utilization using machine learning models.
Figure 3. Workflow for predicting open data utilization using machine learning models.
Sustainability 16 05880 g003
Table 1. Superior models, training data, and performance metrics by field using File Data.
Table 1. Superior models, training data, and performance metrics by field using File Data.
ItemModelTraining DataMSERMSE
Public administrationCatBoostIntegrated0.6650.443
Science and technologyCatBoostField0.7180.516
EducationCatBoostIntegrated0.6250.390
Transportation and logisticsCatBoostIntegrated0.6430.414
Land managementCatBoostIntegrated0.5980.358
Agriculture and fisheriesCatBoostIntegrated0.6520.425
Culture and tourismStacking ensembleIntegrated0.7010.491
LawCatBoostIntegrated0.5060.256
HealthcareStacking ensembleIntegrated0.6270.393
Social welfareCatBoostIntegrated0.6380.407
Industry and employmentCatBoostIntegrated0.6370.405
Food and healthCatBoostIntegrated0.5840.341
Disaster safetyStacking ensembleIntegrated0.6220.387
FinanceCatBoostField0.5570.311
Unification and diplomacyCatBoostIntegrated0.6230.388
Environment and meteorologyCatBoostIntegrated0.6710.450
Table 2. Comparison of model performance based on mean squared error by field using File Data.
Table 2. Comparison of model performance based on mean squared error by field using File Data.
ItemMethod
Single-ModelMulti-Model
(Stacking Ensemble)
AlgorithmRandom ForestXGBoostLightGBMCatBoost
Training DataIntegratedFieldIntegratedFieldIntegratedFieldIntegratedFieldIntegratedField
FieldPublic administration0.5560.5410.5330.5270.5340.5040.4430.5010.5050.494
Science and technology0.7930.5920.7540.5550.7360.6160.6780.5160.7630.607
Education0.4780.5280.4720.5590.4700.5030.3900.4840.4500.509
Transportation and logistics0.5380.6310.5520.6210.5530.6570.4140.5930.5110.618
Land management0.5340.5010.5570.4950.5490.4900.3580.4880.5160.523
Agriculture and fisheries0.4820.4910.4780.4680.4780.4450.4250.4430.4350.470
Culture and tourism0.5350.5810.5190.5390.5460.5520.5010.5570.4910.544
Law0.3850.3740.3350.3440.3130.4080.2560.3260.3470.373
Healthcare0.4350.5230.4290.4960.4160.5080.4660.5040.3930.509
Social welfare0.4860.4830.4330.4870.4500.4790.4070.4800.4330.461
Industry and employment0.4970.5150.4760.5190.4630.4710.4050.4780.4530.476
Food and health0.5440.5180.5150.4290.5180.4500.3410.4350.5000.475
Disaster safety0.4510.6630.4160.6980.4240.6780.5250.6310.3870.635
Finance0.3600.3500.3480.3130.3560.3290.3960.3110.3300.314
Unification and diplomacy0.4750.4650.5130.5280.5250.5560.3880.4980.4750.532
Environment and meteorology0.5280.6010.5190.6220.5170.5910.4500.5750.4820.577
Table 3. Comparison of model performance based on root mean squared error by field using File Data.
Table 3. Comparison of model performance based on root mean squared error by field using File Data.
ItemMethod
Single-ModelMulti-Model
(Stacking Ensemble)
AlgorithmRandom ForestXGBoostLightGBMCatBoost
Training DataIntegratedFieldIntegratedFieldIntegratedFieldIntegratedFieldIntegratedField
FieldPublic administration0.7450.7350.7300.7260.7310.7100.6650.7080.7100.703
Science and technology0.8910.7690.8680.7450.8580.7850.8240.7180.8730.779
Education0.6920.7270.6870.7480.6860.7100.6250.6960.6710.714
Transportation and logistics0.7330.7940.7430.7880.7430.8110.6430.7700.7150.786
Land management0.7310.7080.7460.7040.7410.7000.5980.6980.7180.723
Agriculture and fisheries0.6940.7000.6910.6840.6910.6670.6520.6660.6600.686
Culture and tourism0.7320.7620.7210.7340.7390.7430.7080.7460.7010.738
Law0.6200.6110.5790.5870.5590.6390.5060.5710.5890.611
Healthcare0.6600.7230.6550.7050.6450.7130.6830.7100.6270.713
Social welfare0.6970.6950.6580.6980.6710.6920.6380.6930.6580.679
Industry and employment0.7050.7170.6900.7200.6800.6860.6370.6920.6730.690
Food and health0.7380.7200.7180.6550.7200.6710.5840.6600.7070.689
Disaster safety0.6720.8140.6450.8350.6510.8230.7250.7940.6220.797
Finance0.6000.5920.5900.5600.5970.5730.6290.5570.5750.560
Unification and diplomacy0.6890.6820.7170.7270.7240.7460.6230.7050.6890.730
Environment and meteorology0.7270.7760.7210.7880.7190.7690.6710.7580.6940.760
Table 4. Superior models, training data, and performance metrics by field using OpenAPI Data.
Table 4. Superior models, training data, and performance metrics by field using OpenAPI Data.
ItemModelTraining DataMSERMSE
Public administrationRandom forestField9.7973.130
Science and technologyRandom forestField9.4993.082
EducationXGBoostIntegrated4.8672.206
Transportation and logisticsLightGBMIntegrated14.2043.769
Land managementLightGBMField8.9152.986
Agriculture and fisheriesStacking ensembleIntegrated6.9952.645
Culture and tourismRandom forestIntegrated5.4122.326
LawStacking ensembleIntegrated2.5331.592
HealthcareStacking ensembleIntegrated4.1932.048
Social welfareCatBoostField5.2442.290
Industry and employmentXGBoostIntegrated7.4272.725
Food and healthLightGBMField6.8752.622
Disaster safetyRandom forestField4.5802.140
FinanceRandom forestField11.9213.453
Unification and diplomacyRandom forestField6.0692.463
Environment and meteorologyCatBoostIntegrated4.3802.093
Table 5. Comparison of model performance based on mean squared error by field using OpenAPI Data.
Table 5. Comparison of model performance based on mean squared error by field using OpenAPI Data.
ItemMethod
Single-ModelMulti-Model
(Stacking Ensemble)
AlgorithmRandom ForestXGBoostLightGBMCatBoost
Training DataIntegratedFieldIntegratedFieldIntegratedFieldIntegratedFieldIntegratedField
FieldPublic administration11.3579.79710.78411.22311.10311.60811.38510.83210.20410.869
Science and technology25.3429.49930.01915.41426.60622.80127.60011.13427.83022.103
Education6.7125.3164.8677.5094.9308.0965.5306.3865.0027.271
Transportation and logistics17.98718.60816.56518.70514.20422.42317.20616.98215.82519.322
Land management11.69510.51613.60410.06211.4948.91511.87710.04513.18510.850
Agriculture and fisheries7.6249.4317.8568.8257.23210.4038.2669.2746.9959.945
Culture and tourism5.4127.8305.7136.9005.7608.5355.8597.2495.4708.095
Law2.90629.1639.21748.5479.73614.0114.33228.2682.53335.837
Healthcare5.3476.3484.2397.6884.2596.6684.9726.6724.1934.843
Social welfare6.8255.3356.5346.4377.0957.5656.8455.2446.2956.590
Industry and employment8.81011.3827.42714.1378.05812.2638.67710.5558.03711.853
Food and health8.2277.0057.5947.8058.3096.8759.9907.3477.67712.030
Disaster safety7.4774.5809.8456.4348.9116.1509.3685.5728.3865.418
Finance18.65411.92120.08516.46317.98114.14717.61114.24719.44414.862
Unification and diplomacy12.7906.0698.6139.5628.7927.0757.1257.5356.4807.565
Environment and meteorology6.62811.1395.45911.3874.44710.4584.3809.5065.25210.120
Table 6. Comparison of model performance based on root mean squared error by field using OpenAPI Data.
Table 6. Comparison of model performance based on root mean squared error by field using OpenAPI Data.
ItemMethod
Single-ModelMulti-Model
(Stacking Ensemble)
AlgorithmRandom ForestXGBoostLightGBMCatBoost
Training DataIntegratedFieldIntegratedFieldIntegratedFieldIntegratedFieldIntegratedField
FieldPublic administration3.3703.1303.2843.3503.3323.4073.3743.2913.1943.297
Science and technology5.0343.0825.4793.9265.1584.7755.2543.3375.2754.701
Education2.5912.3062.2062.7402.2202.8452.3522.5272.2372.697
Transportation and logistics4.2414.3144.0704.3253.7694.7354.1484.1213.9784.396
Land management3.4203.2433.6883.1723.3902.9863.4463.1693.6313.294
Agriculture and fisheries2.7613.0712.8032.9712.6893.2252.8753.0452.6453.154
Culture and tourism2.3262.7982.3902.6272.4002.9222.4212.6922.3392.845
Law1.7055.4003.0366.9683.1203.7432.0815.3171.5925.986
Healthcare2.3122.5192.0592.7732.0642.5822.2302.5832.0482.201
Social welfare2.6132.3102.5562.5372.6642.7502.6162.2902.5092.567
Industry and employment2.9683.3742.7253.7602.8393.5022.9463.2492.8353.443
Food and health2.8682.6472.7562.7942.8822.6223.1612.7112.7713.468
Disaster safety2.7342.1403.1382.5372.9852.4803.0612.3612.8962.328
Finance4.3193.4534.4824.0574.2403.7614.1973.7744.4103.855
Unification and diplomacy3.5762.4632.9353.0922.9652.6602.6692.7452.5462.751
Environment and meteorology2.5743.3372.3373.3742.1093.2342.0933.0832.2923.181
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jeong, J.; Cho, K. Proposing Machine Learning Models Suitable for Predicting Open Data Utilization. Sustainability 2024, 16, 5880. https://doi.org/10.3390/su16145880

AMA Style

Jeong J, Cho K. Proposing Machine Learning Models Suitable for Predicting Open Data Utilization. Sustainability. 2024; 16(14):5880. https://doi.org/10.3390/su16145880

Chicago/Turabian Style

Jeong, Junyoung, and Keuntae Cho. 2024. "Proposing Machine Learning Models Suitable for Predicting Open Data Utilization" Sustainability 16, no. 14: 5880. https://doi.org/10.3390/su16145880

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop