Next Article in Journal
What Types of Government Support on Food SMEs Improve Innovation Performance?
Next Article in Special Issue
Case Study Research to Foster the Optimization of Supply Chain Management through the PSS Approach
Previous Article in Journal
Gorilla Troops Optimizer for Electrically Based Single and Double-Diode Models of Solar Photovoltaic Systems
Previous Article in Special Issue
Sustainable Product-Service Systems Customization: A Case Study Research in the Medical Equipment Sector
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Product Service System Configuration Based on a PCA-QPSO-SVM Model

Business School, University of Shanghai for Science and Technology, Shanghai 200093, China
*
Author to whom correspondence should be addressed.
Sustainability 2021, 13(16), 9450; https://doi.org/10.3390/su13169450
Submission received: 12 July 2021 / Revised: 13 August 2021 / Accepted: 17 August 2021 / Published: 23 August 2021

Abstract

:
To achieve sustainable development and improve market competitiveness, many manufacturers are transforming from traditional product manufacturing to service manufacturing. In this trend, the product service system (PSS) has become the mainstream of supply to satisfy customers with individualized products and service combinations. The diversified customer requirements can be realized by the PSS configuration based on modular design. PSS configuration can be deemed as a multi-classification problem. Customer requirements are input, and specific PSS is output. This paper proposes an improved support vector machine (SVM) model optimized by principal component analysis (PCA) and the quantum particle swarm optimization (QPSO) algorithm, which is defined as a PCA-QPSO-SVM model. The model is used to solve the PSS configuration problem. The PCA method is used to reduce the dimension of the customer requirements, and the QPSO is used to optimize the internal parameters of the SVM to improve the prediction accuracy of the SVM classifier. In the case study, a dataset for central air conditioning PSS configuration is used to construct and test the PCA-QPSO-SVM model, and the optimal PSS configuration can be predicted well for specific customer requirements.

1. Introduction

As the product market is nearly saturated, it is getting more challenging for manufacturers to satisfy diversified and individualized consumer requirements. To achieve sustainable growth, many manufacturers have transformed from product-oriented manufacturing to service-oriented manufacturing. This trend of product service system (PSS) to enhance competitiveness and sustainability leads to a massive research effort. The PSS method is based on the simultaneous development of products with tangible features and services surrounding the intangible features, to provide an integrated product that can effectively consider all life cycle stages of the product and related services [1,2].
PSS is described as a hybrid solution that comprises products and services for the purpose of increasing value for customers [3]. On the one hand, customers look forward to being provided with services that aim at enhancing the function and economic performance of the product, such as recycling and maintenance [4]. On the other hand, with ever fierce competition and more diversified customer needs, the low-value-added manufacturing paradigm can no longer meet the requirements of the market and the environment [5]. Therefore, PSS integrates the resources of various parties to meet the needs of customers and reduces the material flow in the consumption process by adding services, which is critical to improve social productivity, living standards, and environmental protection. The emergence of PSS has changed the mode of manufacturing and supply. To create maximum value for customers, PSS has to be individually configured. PSS configuration is to provide a concept by selecting appropriate product modules and service modules among the design modules in advance [6].
The foundation of configuration is to design product modules and service modules by modular design. Modular design is a commonly used approach to realize mass production and individualized design. It combines product elements to form a subsystem with specific functions and uses the subsystem as a versatile module to combine other subsystems to form a system and then produce a product with a certain function. Belkadi et al. [7] proposed a modular design process to support the configuration of production systems that meet the needs of regional markets. Rennpferdt et al. [8] verified the importance of modular product architecture for PSS development. To describe the dependencies between activities, Geng et al. [9] proposed a modular design method based on the fuzzy design structure matrix (FDSM) to obtain the result-oriented PSS. The PSS modular design is a very effective way to improve the response speed and reduce the cost of the industry.
The key in configuration is to combine product modules and service modules to satisfy the customer demands. The existing PSS configuration approaches include genetic algorithm (GA) [10,11], ontology modeling [12,13], convolutional neural network (CNN) [14], multi-objective programming [5,15], and so on. These approaches solve the PSS configuration problem as an optimization model with an objective or a fitness function. The goal of the PSS configuration is to provide a PSS concept that meets certain customer demands. Similar customer demands can be provided with the same PSS configuration. Therefore, the PSS configuration problem is regarded as a classification problem in this paper. The support vector machine (SVM) is an effective tool to solve classification problems. It mainly solves small samples and nonlinear problems. SVM has been successfully applied in the field of classification and regression. Demidova et al. [16] used an SVM based on NSGA-II to predict the effect of medical and technical diagnosis. Pławiak et al. [17] predicted Australian credit scoring based on a deep genetic cascade ensemble of SVM classifiers. Sun et al. [18] proposed a DNN decision-tree-based SVM model for speech emotion classification and recognition.
In this paper, PSS configuration is solved by the SVM method. The PSS configuration problem can be decomposed into three sub-problems: requirement feature extraction, SVM parameter optimization, and PSS configuration classification.
The first task is to reduce the dimension of customer requirements. Among the methods to reduce data dimension, principal component analysis (PCA) shows a strong advantage. PCA utilizes high variance to generate new components that store the most valuable information of elements [19]. Asante-Okyere et al. [20] adopted PCA as a dimension reduction method to improve the performance of the optimized least-squares support vector machine (LSSVM) and the adaptive neuro-fuzzy inference system-subtractive clustering method (ANFIS-SCM). Cao et al. [21] proposed a fault diagnosis method based on PCA and the Gaussian mixed model (GMM), and the PCA was used to reduce the data dimension and increase the feature resolution. Henry et al. [22] proposed an integrated framework based on the use of convolutional neural networks (CNN) and PCA to detect stiction and identify the severity of stiction. The PCA acts as a dimension reduction tool to visualize the extracted features. This paper uses the PCA algorithm to reduce the dimension of customer requirements.
The second task is to optimize SVM parameters. The selection of SVM internal parameters affects the classification performance and fitting effect of the SVM model. The commonly used algorithms are GA, particle swarm optimization (PSO), ant colony algorithm, and so on, while these algorithms have to set many parameters and are easy to fall into the problem of local optimization. Lu et al. [23] proposed a quantum particle swarm optimization (QPSO) algorithm from the perspective of quantum mechanics. The QPSO assumes that the particle swarm optimization system is a particle system that satisfies the basic hypothesis of quantum mechanics. In quantum space, particles do not have a specific trajectory, which allows particles to search for the global optimal solution in the entire feasible solution space. The lack of a defined trajectory means that the speed and position of the particles cannot be determined simultaneously in quantum space. The QPSO algorithm uses the Monte Carlo method to calculate the optimal position of the quantum particle to make sure the global randomness. It has the advantage of a fast convergence rate, and its fitness value is better than that of the traditional PSO [24,25]. This paper takes the advantage of the QPSO algorithm for global random optimization in SVM classification to find the optimal parameters.
The third task is to construct and test a PCA-QPSO-SVM model to predict PSS configuration. The customer requirements and PSS configurations are coded in the model. The PSS configuration is a class label. By comparing the original class label and the predictive class label, the accuracy of the model can be obtained. The PCA-QPSO-SVM model can predict a PSS configuration for specific customer requirements.
The rest of this paper is arranged as follows: In Section 2, the relevant literature is reviewed. The framework of the proposed PSS configuration approach is summarized in Section 3. Section 4 discusses how to use PCA and QPSO to optimize SVM parameters. In Section 5, the PSS configuration model based on PCA-QPSO-SVM is illustrated in detail. Section 6 is a case for an air-conditioner service system to demonstrate the efficacy of the proposed approach. Section 7 is the conclusion of this paper.

2. Literature Review

This section gives a comprehensive literature review consisting of three parts: PSS design, PSS configuration, and PSS configuration optimization.

2.1. PSS Design

PSS is described as a combination of products and services to provide the required utility or function to meet customer needs [26]. To help meet specific customer needs, various PSS development methods have been designed. Durugbo et al. [27] discussed an information flow modeling technique related to the key features of a PSS (such as origin, concepts, and applications). Their proposed model decomposed the iterative process between customers and service providers into different stages to provide total care products. Chiu et al. [28] proposed a process of developing PSS business models to expand current products or services into new market areas. Fargnoli et al. [29] proposed a method based on the collaborative use of PSS quality function deployment, axiomatic design, and service blueprint tools to provide the correlation between customer expectations, PSS components, and PSS modules. Lee et al. [30] utilized a structural services innovation approach to integrating PSS engineering and service engineering in designing intelligent product-service systems. Wang et al. [31] presented a PSS requirements heuristic framework to help engineering designers better improve designs or generate new design concepts.

2.2. PSS Configuration

The system framework for PSS configuration was first proposed by Aurich [32]. Based on the determination of the specific products and service features, technology and service configuration were combined to generate a customized PSS in this framework. The existing PSS configuration approaches include GA, multi-objective programming, ontology modeling, and so on. Sheng et al. [11] proposed a PSS configuration optimization method based on a genetic algorithm. Xuanyuan et al. [10] presented a multi-objective optimization for product configuration and used a multi-objective genetic algorithm to find the Pareto optimal solution set from feasible solutions. Song et al. [5] proposed a multi-objective optimization model for product extension service configuration to solve the problems of too many service solutions and low service delivery efficiency in PSS configuration. Dong et al. [12] proposed an ontology-based service product modeling approach for configuration. Shen et al. [13] presented an ontology-based approach to represent service configuration knowledge and developed a product-extended service configuration system. Based on the combination of quality function deployment and screening life cycle modeling, Haber and Fargnoli [33] proposed a product service system method, which is based on the combination of product service system quality function deployment and screening life cycle modeling tools, which is used to customize solutions for different usage patterns and achieve better environmental performance than independent products.
With the support of artificial intelligence, machine learning algorithms are introduced into the configuration study. To understand customer needs by using quantitative approaches, Yu et al. [34] proposed a knowledge-based artificial neural network (ANN) combined with a decision tree to substantiate customer needs to product specification. Zhou et al. [14] trained the CNN to obtain the complex nonlinear mapping relationship between the customer’s demand attributes and the basic types of product-service systems, to determine the PSS configuration. Shen et al. [35] proposed a method that combines cluster neural network and rule algorithm to extract configuration rules between service parameters and functional requirements, customer features, or product features to gain higher effectiveness of configuration solution.
The SVM, as a kind of machine learning algorithm, has shown excellent results in solving small samples and multi-classification problems. Nowadays, a large number of scholars have used SVM for prediction and classification. Ahlawat et al. [36] developed a hybrid model of a powerful CNN and an SVM for the recognition of handwritten digits from the ministry dataset. Viloria et al. [37] used an SVM to predict the diagnosis of diabetes mellitus (DM). Zhou et al. [38] proposed a multi-model latent space-inducing ensemble SVM classifier for early dementia diagnosis with neuroimaging data. Shao et al. [39] studied the energy consumption of hotel buildings by establishing a support vector machine energy consumption prediction model.

2.3. PSS Configuration Optimization

In the process of predicting the PSS configuration by the SVM method, the optimization of the SVM classifier largely determines the optimization of the PSS configuration scheme.
In the PSS configuration problem, the PSS configuration concepts are classified by the customer requirements. There are a number of requirements that affect classification accuracy. Therefore, the quantity of the characteristics should be reduced, which means the dimension reduction in the SVM algorithm. There are many approaches for feature extraction or dimension reduction, such as neural network, PCA, and so on. Zhang et al. [40] used neural networks to extract more sensitive features with a shallow structure, thereby improving the accuracy of small sample classification. Kontonatsios et al. [41] developed a novel feature extraction method based on a neural network. Unlike the previous unsupervised feature extraction methods, this paper extracts document features in a supervised environment. Xiao et al. [42] presented a feature extractor based on Deep Convolutional Neural Networks to extract patterns shared by a family from entropy graphs automatically. Zhang et al. [43] used a PCA based on the AdaBoost algorithm to detect breast cancer. Ratnasari et al. [44] used a threshold-based region of interest (ROI) and PCA to reduce X-ray images, which obtained the best gray-level threshold of 150. Ma et al. [45] leveraged a deep convolutional neural network to extract image features and leveraged a PCA algorithm to achieve dimension reduction. Negi et al. [46] proposed a method that combines PCA and uncorrelated linear discriminant analysis (ULDA) to obtain the best features that control upper limb motion. Therefore, this paper selects PCA for feature extraction of the dataset. The basic idea of PCA is to use fewer mutually independent features to replace a large amount of information of the original features. It is favored by a large number of scholars.
On the other hand, the accuracy of the SVM classifier prediction depends heavily on the selection of the kernel function and penalty factor. Many scholars have used some optimization algorithms to optimize the parameters of SVM, such as the GA, grid search (GS), PSO, and so on.
(1)
A genetic algorithm is a computational model that simulates the biological evolution process of natural selection and the genetic mechanism of Darwin’s biological evolution theory. It is a method of searching for the optimal solution by simulating the natural evolution process. GA is often combined with SVM to optimize the parameters of SVM. Huang et al. [47] proposed the GA-SVM model to analyze the quantitative contribution of climate change and human activities to changes in vegetation coverage. The model used genetic algorithms to optimize the loss parameters, kernel function parameters, and loss function epsilon values in the SVM. Based on GA-SVM for rapid and effective screening of human papillomaviruses, Chen et al. [48] proposed a Raman spectroscopy technique that improved the accuracy of the model to optimize the penalty factors and nuclear function parameters in the SVM model. Li et al. [49] used the GA-SVM model to identify and classify flip chips.
(2)
Grid search is an exhaustive search method. By looping through the possible values of multiple parameters, it generates the parameter with the best performance, which is the optimal parameter. GS is a common method to optimize the parameters of SVM. Lv et al. [50] used PSO-SVM and GS-SVM to predict the corrosion rate of a steel cross-section. Tan et al. [51] proposed a method combining a successive projections algorithm (SPA) with an SVM based on GS-SVM to classify and identify apple samples with different degrees of bruising. Kong et al. [52] used the GS-SVM model to assess marine eutrophication states of coastal waters.
(3)
Particle swarm optimization is an optimization algorithm that simulates the predation behavior of bird swarms. The iteration process forms the optimal position and optimal direction, hence updating the particle swarm. Many scholars apply the PSO algorithm to optimize SVM parameters. García Nieto et al. [53] proposed a hybrid PSO optimized SVM model to predict the successful growth cycle of spirulina. Liu et al. [54] developed the PSO-SVM model to predict the daily PM2.5 level. Bonah et al. [55] combined Vis-NIR hyperspectral imaging with pixel analysis and a new CARS-PSO-SVM model to classify foodborne bacterial pathogens.
In the PSO algorithm, the position and moving speed of the particles co-determine the movement trajectory of the particles. In quantum space, particles do not have a specific trajectory, which allows particles to search for the global optimal solution in the entire feasible solution space. The lack of a defined trajectory means that the speed and position of the particles cannot be determined simultaneously. Ch et al. [56] employed the SVM-QPSO model to forecast the streamflow values of Vijayawada station and Polavaram station of Andhra Pradesh in India. Li et al. [57] presented the use of an LSSVM algorithm based on quantum-behaved particle swarm optimization to establish the nonlinear relationship of slope stability. They verified that QPSO-SVM can provide a high degree of accuracy and reliability. Therefore, this paper applies the QPSO to optimize the SVM parameters to solve the PSS configuration problem.

3. Research Framework

PSS configuration design is based on the modular design concept. Product modules and service modules compose varied PSS configurations. In the design and sales departments, there are many historical data about customers’ demands and their purchases schemes. This paper takes full advantage of the historical data to construct and test a PCA-QPSO-SVM model. In this model, PCA is used to reduce the dimension of customer requirements to extract the important features. QPSO is applied to optimize the SVM parameters, and a multi-classification SVM classifies the PSS configurations. The research framework is shown in Figure 1.
(1)
Data preparation and preprocessing
Customer requirements and the corresponding PSS configurations are obtained from the design and sales database. The representative historical dataset is extracted.
(2)
Reduction of the requirement dimension
The PCA algorithm is used to reduce the dimension of the customer requirements, then the processed dataset is used in the training set and testing set.
(3)
Construction of the QPSO-SVM model
The QPSO algorithm is used to optimize the kernel function σ and the penalty factor C for SVM. Functions that satisfy Mercer’s theorem [58] can be used as the kernel function. The kernel function is used to map the samples from the original space to the higher-dimension feature space and make them linearly separable in the feature space. The penalty factor is used to control the balance between margin maximization and deviation minimization [59]. By optimizing relevant parameters, the QPSO-SVM model is constructed and tested to verify the accuracy. In the QPSO-SVM model, the customer requirements are the input, and the PSS configuration meeting the customer requirements is the output.
(4)
Prediction of the PSS configuration scheme
According to the PCA-QPSO-SVM model, PSS configuration can be predicted according to the inputting of new customer requirements.

4. Construction of a PCA-QPSO-SVM Model

4.1. Principal Component Analysis

The principal component analysis is a multivariate statistical method for analyzing the correlation between multiple features. The essential idea of PCA is to reveal the internal structure among multiple features through a few principal components. In other words, multiple features in the original data are reduced to a few features, simultaneously the relationships among the original data features are retained as much as possible. PCA has been playing an important role in the fields of artificial intelligence, data mining and image recognition, and so on. The mathematical derivation process of PCA is as follows.
Assuming that the projection of a sample point xi is W T x i on the hyperplane in the new space, then the covariance matrix of the sample points is i W T x i x i T W after projection. If the projections of all sample points can be separated as much as possible, then the variance of the sample points would be maximized, so the optimization goal can be written as:
max W   t r ( W T X X T W )
s . t .   W T W = I
where, W = (w1, w2, …, wd).
By using the Lagrange multiplier method for Formula (1), we can obtain:
X X T w i = λ i w i
Then the obtained eigenvalues are sorted: λ1λ2 ≥ … ≥ λd. Finally, the eigenvector corresponding to the top d’ eigenvalues forms W = (w1, w2, …, wd).
The steps of the PCA algorithm are as follows:
  • Step 1: Set the initial dataset D = {x1, x2, …, xm} and the low-dimensional space dimension d’.
  • Step 2: Centralize all samples: xi x i 1 m i = 1 m x i .
  • Step 3: Calculate the sample covariance matrix X X T and decompose the eigenvalues of the covariance matrix X X T .
  • Step 4: Take the eigenvector corresponding to the top d’ eigenvalues w1, w2, …, wd’.

4.2. Quantum Particle Swarm Optimization Algorithm

The PSO algorithm was proposed by Kennedy and Eberhart [60]. It is a random search algorithm based on swarm intelligence developed by simulating the foraging behavior of birds. Its basic idea is to find the optimal solution through collaboration and information sharing among individuals in a group. The mathematical description of the PSO algorithm is as follows:
Assuming that the particle swarm size is M, the current optimal position of particle i in the n-dimensional space is expressed as:
p b e s t i = ( p i 1 , p i 2 , , p i n )
The current optimal position of the entire particle swarm is:
g b e s t = ( g 1 , g 2 , , g n )
During each iteration, the velocity update formula of each particle is:
v i ( t + 1 ) = w v i ( t ) + c 1 r a n d 1 ( p b e s t i ( t ) x i ( t ) ) + c 2 r a n d 2 ( g be s t ( t ) x i ( t ) )
where, c1,c2 are learning factors; w is an inertia factor; rand1, rand2 is a random value in [0, 1]; vi(t) is the velocity of the i-th particle, xi(t) is the position of the i-th particle; t is the number of iterations.
The position update formula of each particle is:
x i ( t + 1 ) = v i ( t + 1 ) + x i ( t )
The PSO algorithm needs to set too many parameters, such as inertia factor w, learning factors c1, c2, etc. It is difficult to find the optimal parameters. Moreover, it is easy to fall into the dilemma of a local optimum due to the lack of randomness in the particle position change. To redress these problems, this paper applies the QPSO algorithm with higher performance. The QPSO algorithm cancels the particle’s moving direction attribute and increases the randomness of the particle position change. The calculation process of the QPSO algorithm is as follows:
The term mbest is introduced in the QPSO algorithm, which represents the average value of pbest. It can be expressed as:
m b e s t = 1 M i = 1 M p b e s t i
The current optimal position of particle i is:
p i = ϕ p b e s t i + ( 1 ϕ ) g b e s t
where Φ is uniform value in [0, 1].
The position update formula of each particle is:
x i = p i ± a l p h a | m b e s t x i | ln ( 1 μ )
where μ is a uniform value in [0, 1] and is the new parameter, which can control the degree of contraction and expansion of the particle position generally. The value of alpha is less than 1.

4.3. Support Vector Machine

The SVM was originally proposed by Cortes and Vapnik [61]. It is a supervised learning algorithm that classifies datasets. The essential idea of the SVM is to construct a hyperplane in the sample space as a decision line that divides two different samples. The SVM can solve both binary classification problems and multi-class problems. In this paper, PSS configuration is regarded as a multi-class problem. The derivation process of the SVM algorithm is as follows.
Given a set S containing N training samples, S = {(xi, yi), i = 1, 2, …, N}, the expression of classification hyperplane is:
f ( x ) = w x + b
where w is the normal vector of the hyperplane and b is the translation distance of the hyperplane.
The objective function of partition hyperplane with the “maximum interval” can be expressed as:
min w , b   1 2 w 2
y i ( w x i + b ) 1 0 , i { 1 , 2 , , N }
To enhance the error tolerance of the SVM classifier, a relaxation variable ξi and the penalty factor C are introduced into the objective function. The objective function is expressed as:
min w , b 1 2 | | w | | 2 + C i = 1 n ξ i
s . t . y i ( w x + b ) 1 ξ i
ξ i 0 , i = 1 , 2 , , N
where ξi is a non-negative slack variable, which is used to improve the generalization ability of the model. C is the penalty factor, which is used to control the balance between margin maximization and deviation minimization.
For solving the non-linear classification problem, it is necessary to introduce the kernel function in the SVM. The kernel function is used to map the samples from the original space to the higher-dimension feature space and make them linearly separable in the feature space. In this paper, the Gaussian kernel function is used:
k ( x i , x j ) = exp ( ( x i , x j ) 2 σ 2 )
The Lagrange multiplier is brought into the objective function. The optimization problem can be expressed as:
max i = 1 n α i 1 2 i = 1 n j = 1 n α i α j y i y j k ( x i , x j )
s . t . i = 1 n α i y i = 0
0 α i C , i = 1 , 2 , , N
where αi is the Lagrangian coefficient.
The final optimized hyperplane is expressed as:
f ( x ) = i = 1 n α i y i k ( x i , x j ) + b , i = 1 , 2 , , N

4.4. Optimization of the SVM Parameters

The quantity of the feature attributes affects the accuracy of the SVM model. The dimension of the feature attributes should be reduced first by the PCA algorithm. Then, the QPSO algorithm is used to optimize the C and σ of the SVM classifier based on the processes data. Finally, the optimal parameters are brought into the SVM model to realize the PCA-QPSO-SVM modeling. The PCA-QPSO-SVM optimization process is shown in Figure 2. The steps of constructing the PCA-QPSO-SVM model are as follows:
  • Step 1: Use a PCA algorithm to reduce the dimension of dataset Q to get a new dataset Q’.
  • Step 2: Determine the initial parameters of the QPSO, such as the number of particle swarms, the range of the parameters, the alpha value, and so on.
  • Step 3: Set the fitness function in QPSO. In this paper, the fitness function is the average of SVM cross-validation (CV), and its value represents the classification accuracy of the model. The optimal value pbest and the global optimal value gbest for each particle are updated by iterating the fitness function, where pbest is the penalty factor C, gbest is the kernel function σ.
  • Step 4: Calculate the optimal position mbest of the particle swarm and update the new position of each particle.
  • Step 5: Determine the end condition. When the optimal search reaches the maximum number of iterations, the optimal search ends; otherwise, go to Step 3.
  • Step 6: The optimal parameters (C, σ) are brought into the SVM model to conduct prediction.

5. PSS Configuration Based on the PCA-QPSO-SVM Model

This paper proposes a PCA-QPSO-SVM model and uses it to solve the PSS configuration problem. The process includes two parts: data collection and construction of the PCA-QPSO-SVM model.

5.1. Data Collecting and Processing

In real life, customers would have different needs for PSS, which generates diversified requirements for PSS configuration. It is vital to collect customer requirements features. Taking central air conditioning as an example, consumers usually use strong stability, high energy saving, and strong reliability to express their requirements. The vague requirements have to be understood and realized by a specific PSS configuration. The proposed PCA-QPSO-SVM model is used to classify customer requirements using PSS configuration as a class label. Individualized PSS configuration can be obtained, which will support the PSS design.
In the PCA-QPSO-SVM model, the input must be a real vector. We assume requirement features of customer Ci are represented by En = {CNi|1 < i < n, CN1, …, CNi, …, CNn}, and each feature CNi has t selectable values. Then numbers can be used to encode the t optional values. For example, E1 = {0, 0, 2,…, 2, 1, −1} represents the requirement features of customer C1.
A PSS solution is comprised of product modules and service modules. Assuming there are s product modules in the PSS configuration framework, the module features for the product part can be expressed as P = {Pi|1 < i < s}. There is one or more instances for each module characteristic. For example, there exist t available instances for module Pi, and the set is {pij|1 < j < t}. Similarly, the module features for the service part can be expressed as S = {Sv|1 < v < u}. Assume there exist q available instances for module Sv, and the set is {suw|1 < w < q}. The product instances and service instances make up diversified PSS configuration instances. Assume there exist l PSS configuration instances, and PSS configuration instances can be expressed as PSS = {PSSk|1 < k < l}. Product and service modules have different selectable instance values, which can be represented by discrete numbers as tags. For example, a PSS configuration instance is expressed as {p12, p22,…, pi3, s11, s22,…, su3}.

5.2. Construction of the PCA-QPSO-SVM Model for PSS Configuration

  • Step 1: Determine the product modules and service modules, then combine the corresponding instances to form different PSS configurations. According to the relevant historical data, the ‘requirements-configuration’ samples are collected to construct the model.
  • Step 2: Reduce the dimension of requirement features by using the PCA algorithm. QPSO is used to perform k-fold cross-validation (CV) to find the best Gaussian kernel function σ and the penalty factor C. For k-fold CV, the entire training set is divided into k subsets with an equal number of samples. One of the subsets is selected as the testing set, and the remaining k-1 subsets are the training set.
  • Step 3: Construct the multi-class SVM model by using the best parameter combination (C, σ) to test the testing set. After constructing a reliable classification model, PSS configuration can be predicted by inputting new customer requirements.

6. Case Study

Company A is a central air-conditioning manufacturer. To enhance the company’s competitiveness and meet the diversified requirements of customers, company A decides to transform towards a service-oriented enterprise and provides customers with individualized PSS configuration. According to the modular design platform, and the PSS configuration design can be realized by combining different product instances and service instances. To quickly and accurately recommend the schemes to customers among diversified PSS configuration instances, the proposed PCA-QPSO-SVM model is applied in company A.
First, 100 data samples are extracted from the design and sales database. Then, the 100 data samples are divided into a training set and a testing set. The number of samples in the training set is 75 and that in the testing set is 25. Finally, the multi-classification PCA-QPSO-SVM model is used to predict the PSS configuration. The customer’s need is used as the input of the PCA-QPSO-SVM model, and the PSS configuration scheme is used as the output of the model.

6.1. Data Coding and Features Analysis

According to customers’ using conditions, 11 requirement features were identified by experts, including environmental protection, stability, intelligence, simplicity, convenience, adaptability, reliability, comfort, energy-saving, safety, and heat dissipation. Five-point requirement levels are defined as {L, ML, M, MH, H}, which mean {low, medium-low, medium, medium-high, high}. The corresponding code is {−2, −1, 0, 1, 2}. The description of requirement features is shown in Table 1. For instance, if a customer requires high energy savings, then the value for CN9 can be represented by the code ‘2′. By consulting the experts, the modules and relevant instances of the central air conditioner are shown in Table 2. The modules and relevant instances of the service are shown in Table 3.
The product modules and service modules constitute different PSS configuration schemes. In this case, the PSS configuration scheme {A1, B2, C2, D2, E1, F1, F3, G1, H1, I1, J2, K1, L2, M2, N1, N2} is coded ‘1′, {A2, B1, C2, D1, E2, F2, F4, G2, H2, I2, J1, K3, L1, M1, N3, N4} is coded ‘2′, {A3, B3, C3, D2, E2, F1, F3, G3, H2, I1, J1, K1, L3, M2, N1, N2} is coded‘3′, {A4, B2, C1, D2, E1, F2, F4, G1, H1, I2, J3, K2, L2, M3, N3, N4} is coded ‘4′, and {A1, B1, C2, D2, E2, F1, F3, G3, H1, I2, J3, K2, L3, M2, N1, N2} is coded ‘5′.
The original design data for constructing the PCA-QPSO-SVM model are shown in Table 4. The distribution of each feature is displayed with a frequency diagram and box plots, as shown in Figure 3 and Figure 4.
It can be seen from Figure 3 and Figure 4 that most of the datasets are relatively evenly distributed and have no abnormal feature attributes. Only features CN8, CN9, CN11 have few outliers, which has little effect on the entire datasets, so the features cannot be eliminated easily. To see the correlation between each feature more intuitively, the establishment of a feature relationship heat map is shown in Figure 5.
In Figure 5, the correlation between features (CN1, CN9) and (CN4, CN10) is higher. Next, by establishing two coordinate systems with CN1 as the x-axis, CN9 as the y-axis and CN4 as the x-axis, and CN10 as the y-axis, we can observe whether they are linearly or nonlinearly related in the plane distribution. The result of feature correlation is shown in Figure 6.
It can be seen from Figure 6 that the features (CN1, CN9) and (CN4, CN10) do not show a linear correlation, so the features CN1, CN9, CN4, and CN10 cannot be eliminated easily.

6.2. PCA-QPSO-SVM Model Construction for PSS Configuration

6.2.1. Dimension Reduction of Requirement Feature

To improve the classification performance of the SVM model, the original feature set is processed to reduce dimension. First, the PCA algorithm is used to reduce the dimension of 11 requirement features in the original 100 sample data. Because the values of the data samples are all in {−2, -1, 0, 1, 2}, there is no need for StandardScaler processing. Second, the principal components are determined by calculating the covariance of each feature. The result of the calculation is shown in Figure 7.
In Figure 7, the first and second principal components account for most of the variance, accounting for 63.62%. Therefore, the parameter set of PCA is n_components = 2. The data after PCA dimension reduction is shown in Table 5.
To demonstrate the distribution of these 100 samples more clearly after PCA dimensionality reduction, this paper visualizes these 100 samples. Take X1 as the x-axis and X2 as the y-axis, and the established rectangular coordinate system is shown in Figure 8.
To visualize the data, the original 11-dimensional features are reduced to 2-dimensional features through the PCA algorithm, thereby reducing nine attribute features. To verify the relationship between the 2-dimensional feature and the original 11-dimensional feature, and we can establish a heat map to reflect the relationship intuitively between the original features and PCA principal components, as shown in Figure 9.
In Figure 9, the color from dark to light represents a value from −0.4 to 0.5, if the value of a feature is positive, it means that it is positively related to the principal component. If it is negative, then the opposite. It can be seen from Figure 9 that the correlation coefficient of X1 is extremely low only with CN1, CN2, CN9, CN10, and the correlation coefficients of X2 are extremely low only with CN3, CN4, CN5, CN10. Then it indicates that the substitution effect of X1 and X2 is better.

6.2.2. QPSO-SVM Model Construction and Parameters Setting

First, the 2-dimensional dataset obtained by the PCA algorithm is used as the dataset of the multi-class SVM model. Then, the QPSO algorithm is used for parameter optimization of 2-fold CV to find out the optimal parameter pair (C, σ). In the parameter optimization process, 75 randomly selected samples were used to train the QPSO-SVM model. In the QPSO algorithm, a parameter pair (C, σ) was obtained by each iteration, when the number of iterations reaches the maximum and output the optimal parameter pair (C, σ). The specific parameter settings of the QPSO algorithm are shown in Table 6.
The optimal parameter pair (C, σ) was calculated by setting the parameter value of the QPSO algorithm, and then it was used as the parameter sets the value of the multi-class SVM model for prediction analysis. The entire process was completed by Pycharm 2019.

6.3. Prediction and Comparative Analysis of PCA-QPSO-SVM Model

To evaluate the performance of the training model, classification accuracy is used as an index. In each iteration of the PCA-QPSO-SVM model, the training accuracy of training samples is shown in Figure 10.
According to QPSO parameter optimization, when the number of iterations reaches 8, the classification accuracy under 2-fold CV is the highest (95.01%), and the optimal parameter values C = 3.01, σ = 2.41 are obtained at this time. Then the optimal parameters are used in the multi-class SVM model, and 25 samples in the testing set are used to test the classification accuracy of the model. The test results are shown in Figure 11. It turns out that all 25 test samples are predicted correctly, and the accuracy rate is 100%, which verifies that the performance of the QPSO-SVM model is excellent.
To verify the validity of the PCA-QPSO-SVM model prediction, this paper compares PCA-QPSO-SVM with PCA-PSO-SVM, PSO-SVM, GA-SVM, and GS-SVM. In the experiment, various prediction algorithms are run 20 times, respectively. For each method, the optimal prediction result is selected as the final result. The final results are compared as shown in Table 7. The simulation results show that the prediction accuracy of the PCA-QPSO-SVM model is 100%, which is better than other algorithms. Therefore, the proposed model has a better adaptability and predictive ability. In addition, the mean square error of PCA-QPSO-SVM is 0, indicating that the PCA-QPSO-SVM model has a good predictive modeling effect.
To verify the reliability of the model to solve actual problems, this article will test the consistency of the PSS configuration provided by a new customer’s demand with the actual situation. The first new customer needs to configure central air-conditioning for the hospital. Its requirements are good stability and low noise (that is, the environmental protection is very high). Because the patient is considered, the comfort is also very high. The input is {2, 2, −1, −1, 0, 2, 1, 2, 1, 1, 0}, and the output result is the configuration plan “4”, namely {A4, B2, C1, D2, E1, F2, F4, G1, H1, I2, J3, K2, L2, M3, N3, N4}. In practice, A1 has low noise and good stability. Because E1 has the characteristics of low wind pressure and small air volume, it is more comfortable. Configuration plan “4” meets the needs of customers and verifies the reliability of the model.

6.4. Discussion of Results

In the case study, there exist unimportant requirement features, so it is vital to reduce the dimension of customer requirement features. In this paper, we retain 95% information on requirement features. To facilitate the visualization, customer requirements are reduced from 11 features to 2 principal components by the PCA algorithm. Because the PCA algorithm can filter out some small influencing features by reducing the dimension, it will greatly improve the calculation speed and prediction accuracy of SVM. By comparing it with the PSO-SVM model, the PCA-PSO-SVM reaches the optimal result in the 13th iteration, the PSO-SVM model reaches the optimal result in the 21st iteration. In addition, the PCA-PSO-SVM model has the higher prediction accuracy (92%) than the PSO-SVM model (64%). Moreover, the PCA-PSO-SVM model has the lower mean square error (1.0) than the PSO-SVM model (1.04).
The QPSO algorithm is used to optimize the penalty factor C and the kernel function σ of SVM in this paper. The experimental results show that the optimal parameters (C = 3.01, σ = 2.41) and the highest classification accuracy (95.01%) are obtained by the PCA-QPSO-SVM model. By comparing with the PCA-PSO-SVM model, the proposed approach in this paper reaches the optimal result in the ninth iteration, which is faster than the PCA-PSO-SVM model (the 18th iteration). In terms of prediction accuracy, the proposed approach is 100%, which is higher than the PCA-PSO-SVM model (92%). In addition, the proposed approach has the lower mean square error (0) than the PCA-PSO-SVM model (1.0). These verified that the QPSO algorithm has the advantages of fast calculation ability and strong optimization ability in comparison with other optimization approaches.
By comparing with the GA-SVM model, the PSO algorithm can find the optimal parameters faster and lower mean square error. The PSO algorithm reached the optimal in the 21st iteration, which is faster than the GA-SVM model (the 27th iteration). Although the prediction accuracy the PSO-SVM model is similar to the GA-SVM, the mean square error of the PSO-SVM model is 1.04, which is lower than the GA-SVM model (1.56).
Similarly, by comparing with the GS-SVM model, the PSO algorithm can find the optimal parameters faster and with a better prediction accuracy. The PSO algorithm reached the optimal in the 21st iteration, which is faster than the GA-SVM model (the 38th iteration). The prediction accuracy of the PSO-SVM model is 64%, which is higher than the GS-SVM model. In addition, the mean square error of the PSO-SVM model is 1.04, which is lower than the GS-SVM model (3.32).
In general, the experimental results show that it is effective for predicting PSS configuration by the proposed PCA-QPSO-SVM model. The proposed approach has the highest accuracy and best performance by comparing with other approaches, which can provide a more accurate PSS configuration that meets customer requirements. In the product service area, this is beneficial to customers who will receive a better value offering and the manufacturer who can identify the main aspects of its solution more effectively.
The limitation of the dataset is that the sample are relatively small. The optimal parameters and the highest classification accuracy are reached through 50 iterations.

7. Conclusions and Future Research

With the increasingly fierce market competition, more and more manufacturers combine products and services into PSS to provide customers with greater value. Manufacturers must accurately determine which configuration meets customer needs better. The PSS configuration is deemed as a multi-classification problem in this paper. A PCA-QPSO-SVM model is proposed to solve this problem. The validation of the proposed model in air conditioner configuration indicates that it can be used as an effective prediction model for PSS configuration. To sum up, the proposed approach reveals the following features.
The PSS configuration design can be realized by machine learning. The historical design and sales data have great value in supporting individualized PSS design. A multi-classification SVM helps PSS providers to easily predict configuration schemes with higher customer satisfaction.
Compared with the conventional SVM model, the proposed model considers the impact of feature dimension and the selection of key parameters. The PCA algorithm has the advantage of convenience in reducing the dimensions of the dataset. In the process of optimizing SVM parameters, QPSO does not fall into the problem of local optimization and has a better fitness value. The usage of QPSO in optimizing SVM parameters can improve classification accuracy.
Although the proposed optimization method can predict the PSS configuration well, it is restricted by the training data. The predicted model performs effectively when there are adequate and reliable data. Completely new customer requirements are hard to process and generate the appropriate PSS configuration. Therefore, in future research, we can calculate the similarity between new customer requirements and previous customer requirements, and the PSS configuration corresponding to the maximum similarity can be recommended to the new customer.

Author Contributions

Conceptualization, Z.C. and X.G.; methodology, Z.C. and X.G.; writing—review and editing, Z.C. and X.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 71301104, in part by the Humanity and Social Science Foundation of Ministry Education of China under Grant 19YJA630021, and in part by the Specialized Research Fund for the Doctoral Program of Higher Education of China under Grant 20133120120002.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Matschewsky, J.; Kambanou, M.L.; Sakao, T. Designing and providing integrated product-service systems-challenges, opportunities and solutions resulting from prescriptive approaches in two industrial companies. Int. J. Prod. Res 2018, 56, 2150–2168. [Google Scholar] [CrossRef] [Green Version]
  2. Haber, N.; Fargnoli, M. Designing product-service systems: A review towards a unified approach. In Proceedings of the 7th International Conference on Industrial Engineering and Operations Management (IEOM), Rabat, Morocco, 11–13 April 2017; pp. 817–837. [Google Scholar]
  3. Shimomura, Y.; Nemoto, Y.; Kimita, K. A Method for Analysing Conceptual Design Process of Product-Service System. CIRP Ann. Manuf. Technol. 2015, 64, 145–148. [Google Scholar] [CrossRef]
  4. Schweitzer, E.; Aurich, J.C. Continuous Improvement of Industrial Product–service Systems. CIRP J. Manuf. Sci. Technol. 2010, 3, 158–164. [Google Scholar] [CrossRef]
  5. Song, W.; Chan, F.T.S. Multi-objective configuration optimization for product-extension service. J. Manuf. Syst. 2015, 37, 113–125. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Chai, N.; Ostrosi, E.; Shang, Y. Extraction of association rules in the schematic design of product service system based on Pareto-MODGDFA. Comput. Ind. Eng. 2019, 129, 392–403. [Google Scholar] [CrossRef]
  7. Belkadi, F.; Colledani, M.; Urgo, M.; Bernard, A.; Colombo, G.; Borzi, G.; Ascheri, A. Modular design of production systems tailored to regional market requirements: A Frugal Innovation perspective. IFAC-Pap. 2018, 51, 96–101. [Google Scholar] [CrossRef]
  8. Rennpferdt, C.; Greve, E.; Krause, D. The Impact of Modular Product Architectures in PSS Design: A systematic Literature Review. Procedia CIRP 2019, 84, 290–295. [Google Scholar] [CrossRef]
  9. Geng, X.; Jin, Y.; Zhang, Y. Result-oriented PSS Modular Design Method based on FDSM. Procedia CIRP 2019, 83, 610–615. [Google Scholar] [CrossRef]
  10. Xuanyuan, S.; Jiang, Z.; Patil, L.; Li, Y.; Li, Z. Multi-objective optimization of product configuration. In Proceedings of the ASME 2008 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Brooklyn, NY, USA, 3–6 August 2008; pp. 961–968. [Google Scholar]
  11. Sheng, Z.; Xu, T.; Song, J. Configuration design of product service system for CNC machine tools. Adv. Mech. Eng. 2015, 2, 222–231. [Google Scholar] [CrossRef] [Green Version]
  12. Dong, M.; Yang, D.; Su, L. Ontology-based service product configuration system modeling and development. Expert Syst. Appl. 2011, 38, 11770–11786. [Google Scholar] [CrossRef]
  13. Shen, J.; Wang, L.; Sun, Y. Configuration of Product Extension Services in Servitisation Using an Ontology-based Approach. Int. J. Prod. Res. 2012, 50, 6469–6488. [Google Scholar] [CrossRef]
  14. Zhou, Y.; Liu, W.; Chai, N.; Fan, B.; Zhang, Z. Base type selection of product service system based on convolutional neural network. Procedia CIRP 2019, 83, 601–605. [Google Scholar] [CrossRef]
  15. Wei, W.; Fan, W.; Li, Z. Multi-objective optimization and evaluation method of modular product configuration design scheme. Int. J. Adv. Manuf. Technol. 2014, 75, 1527–1536. [Google Scholar] [CrossRef]
  16. Demidova, L.A.; Egin, M.M.; Tishkin, R.V. A Self-tuning Multiobjective Genetic Algorithm with Application in the SVM Classification. Procedia Comput. Sci. 2019, 150, 503–510. [Google Scholar] [CrossRef]
  17. Pławiak, P.; Abdar, M.; Acharya, U.R. Application of new deep genetic cascade ensemble of SVM classifiers to predict the Australian credit scoring. Appl. Soft. Comput. 2019, 84, 105740. [Google Scholar] [CrossRef]
  18. Sun, L.; Zou, B.; Fu, S.; Chen, J.; Wang, F. Speech emotion recognition based on DNN-decision tree SVM model. Speech Commun. 2019, 115, 29–37. [Google Scholar] [CrossRef]
  19. Guyon, I.; Gunn, S.; Nikravesh, M. Feature Extraction: Foundations and Applications; Springer Science Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  20. Asante-Okyere, S.; Shen, C.; Ziggah, Y.Y.; Rulegeya, M.M.; Zhu, X. Principal component analysis (PCA) based hybrid models for the accurate estimation of reservoir water saturation. Comput. Geosci. UK 2020, 145, 104555. [Google Scholar] [CrossRef]
  21. Cao, S.; Hu, Z.; Luo, X.; Wang, H. Research on fault diagnosis technology of centrifugal pump blade crack based on PCA and GMM. Measurement 2021, 173, 108558. [Google Scholar] [CrossRef]
  22. Henry, Y.Y.S.; Aldrich, C.; Zabiri, H. Detection and severity identification of control valve stiction in industrial loops using integrated partially retrained CNN-PCA frameworks. Chemom. Intell. Lab. Syst. 2020, 206, 104143. [Google Scholar] [CrossRef]
  23. Lu, S.; Sun, C.; Lu, Z. An improved quantum-behaved particle-swarm optimization method for short-term combined economic emission hydrothermal scheduling. Energy. Convers. Manag. 2010, 51, 561–571. [Google Scholar] [CrossRef]
  24. Chen, M.; Ruan, J.; Xi, D. Micro grid scheduling optimization based on quantum particle swarm optimization (QPSO) algorithm. In Proceedings of the 2018 Chinese Control and Decision Conference (CCDC), Shenyang, China, 9–11 June 2018; pp. 6470–6475. [Google Scholar]
  25. Zhu, S.; Luo, P.; Yang, Y.; Lu, Q.; Chen, Q. Optimal dispatch for gridconnecting microgrid considering shiftable and adjustable loads. In Proceedings of the IECON 2017—43rd Annual Conference of the IEEE Industrial Electronics Society, Beijing, China, 29 October–1 November 2017; pp. 5575–5580. [Google Scholar]
  26. Baines, T.S.; Lightfoot, H.W.; Evans, S.; Neely, A.; Greenough, R.; Peppard, J.; Alcock, J.R. State-of-the-art in product-service systems. Proc. Inst. Mech. Eng. Part B: J. Eng. Manuf. 2007, 221, 1543–1552. [Google Scholar] [CrossRef] [Green Version]
  27. Durugbo, C.; Tiwari, A.; Alcock, J.R. A review of information flow diagrammatic models for product-service systems. Int. J. Adv. Manuf. Technol. 2011, 52, 1193–1208. [Google Scholar] [CrossRef]
  28. Chiu, M.C.; Kuo, M.Y.; Kuo, T.C. A systematic methodology to develop business model for a product service system. Int. J. Indus. Eng. 2015, 22, 369–381. [Google Scholar]
  29. Fargnoli, M.; Haber, N.; Sakao, T. PSS modularisation: A customer-driven integrated approach. Int. J. Prod. Res. 2019, 57, 4061–4077. [Google Scholar] [CrossRef] [Green Version]
  30. Lee, C.H.; Chen, C.H.; Trappey, A.J. A structural service innovation approach for designing smart product service systems: Case study of smart beauty service. Adv. Eng. Inf. 2019, 40, 154–167. [Google Scholar] [CrossRef]
  31. Wang, Z.; Chen, C.H.; Zheng, P.; Li, X.; Khoo, L.P. A novel data-driven graph-based requirement elicitation framework in the smart product-service system context. Adv. Eng. Inform. 2019, 42, 100983. [Google Scholar] [CrossRef]
  32. Aurich, J.C.; Wolf, N.; Siener, M.; Schweitzer, E. Configuration of product-service systems. J. Manuf. Technol. Manag. 2009, 20, 591–605. [Google Scholar] [CrossRef]
  33. Haber, N.; Fargnoli, M. Sustainable product-service systems customization: A case study research in the medical equipment sector. Sustainability 2021, 13, 6624. [Google Scholar] [CrossRef]
  34. Yu, L.; Wang, L.; Yu, J. Identification of Product Definition Patterns in Mass Customization Using a Learning-Based Hybrid Approach. Int. J. Adv. Manuf. Technol. 2008, 38, 1061–1074. [Google Scholar] [CrossRef]
  35. Shen, J.; Wang, L. Configuration Rules Acquisition for Product Extension Services using Local Cluster Neural Network and Rulex Algorithm. In Proceedings of the 2010 International Conference on Artificial Intelligence and Computational Intelligence, Sanya, China, 23–24 October 2010; pp. 196–199. [Google Scholar]
  36. Ahlawat, S.; Choudhary, A. Hybrid CNN-SVM Classifier for Handwritten Digit Recognition. Procedia Comput. Sci. 2020, 167, 2554–2560. [Google Scholar] [CrossRef]
  37. Viloria, A.; Herazo-Beltran, Y.; Cabrera, D.; Pineda, O.B. Diabetes Diagnostic Prediction Using Vector Support Machines. Procedia Comput Sci. 2020, 170, 376–381. [Google Scholar] [CrossRef]
  38. Zhou, T.; Thung, K.H.; Liu, M.; Shi, F.; Zhang, C.; Shen, D. Multi-modal latent space inducing ensemble SVM classifier for early dementia diagnosis with neuroimaging data. Med. Image Anal. 2020, 60, 101630. [Google Scholar] [CrossRef]
  39. Shao, M.; Wang, X.; Bu, Z.; Chen, X.; Wang, Y. Prediction of energy consumption in hotel buildings via support vector machines. Sustain. Cities. Soc. 2020, 57, 102128. [Google Scholar] [CrossRef]
  40. Zhang, K.; Chen, J.; Zhang, T.; Zhou, Z. A Compact Convolutional Neural Network Augmented with Multiscale Feature Extraction of Acquired Monitoring Data for Mechanical Intelligent Fault Diagnosis. J. Manuf. Syst. 2020, 55, 273–284. [Google Scholar] [CrossRef]
  41. Kontonatsios, G.; Spencer, S.; Matthew, P.; Korkontzelos, I. Using a neural network-based feature extraction method to facilitate citation screening for systematic reviews. Expert. Syst. Appl. X 2020, 6, 100030. [Google Scholar] [CrossRef]
  42. Xiao, G.; Li, J.; Chen, Y.; Li, K. MalFCS: An effective malware classification framework with automated feature extraction based on deep convolutional neural networks. J. Parallel Distrib. Comput. 2020, 141, 49–58. [Google Scholar] [CrossRef]
  43. Zhang, D.; Zou, L.; Zhou, X.; He, F. Integrating feature selection and feature extraction methods with deep learning to predict clinical outcome of breast cancer. IEEE Access 2018, 6, 28936–28944. [Google Scholar] [CrossRef]
  44. Ratnasari, N.R.; Susanto, A.; Soesanti, I. Thoracic X-ray features extraction using thresholding-based ROI template and PCA-based features selection for lung TB classification purposes. In Proceedings of the 2013 3rd international conference on instrumentation, commmunications. Information Technology and Biomedical Engineering (ICICI-BME), Bandung, Indonesia, 7–8 November 2013. [Google Scholar]
  45. Ma, J.; Yuan, Y. Dimension reduction of image deep feature using PCA. J. Vis. Commun. Image Represent. 2019, 63, 102578. [Google Scholar] [CrossRef]
  46. Negi, S.; Kumar, Y.; Mishra, V.M. Feature extraction and classification for EMG signals using linear discriminant analysis. In Proceedings of the 2016 2nd international conference on advances in computing, communication, & automation (ICACCA), Bareilly, India, 30 September–1 October 2016. [Google Scholar]
  47. Huang, S.; Zheng, X.; Ma, L.; Wang, H.; Huang, Q.; Leng, Q.; Meng, E.; Guo, Y. Quantitative contribution of climate change and human activities to vegetation cover variations based on GA-SVM model. J. Hydrol. 2020, 584, 124687. [Google Scholar] [CrossRef]
  48. Chen, C.; Wang, J.; Chen, C.; Tang, J.; Lv, X.; Ma, C. Rapid and efficient screening of human papillomavirus by Raman spectroscopy based on GA-SVM. Optik 2020, 210, 164514. [Google Scholar] [CrossRef]
  49. Li, K.; Wang, L.; Wu, J.; Zhang, Q.; Liao, G.; Su, L. Using GA-SVM for defect inspection of flip chips based on vibration signals. Microelectron. Reliab. 2018, 81, 159–166. [Google Scholar] [CrossRef]
  50. Lv, Y.; Wang, J.; Wang, J.; Xiong, C.; Zou, L.; Li, L.; Li, D. Steel corrosion prediction based on support vector machines. Chaos Solitons Fractals 2020, 136, 109807. [Google Scholar] [CrossRef]
  51. Tan, W.; Sun, L.; Yang, F.; Che, W.; Ye, D.; Zhang, D.; Zou, B. Study on bruising degree classification of apples using hyperspectral imaging and GS-SVM. Optik 2018, 154, 581–592. [Google Scholar] [CrossRef]
  52. Kong, X.; Sun, Y.; Su, R.; Shi, X. Real-time eutrophication status evaluation of coastal waters using support vector machine with grid search algorithm. Mar. Pollut. Bull. 2017, 119, 307–319. [Google Scholar] [CrossRef] [PubMed]
  53. Nieto, P.J.G.; Gonzalo, E.G.; Fernández, J.R.A.; Muñiz, C.D. A hybrid PSO optimized SVM-based model for predicting a successful growth cycle of the Spirulina platensis from raceway experiments data. J. Comput. Appl. Math. 2016, 291, 293–303. [Google Scholar] [CrossRef]
  54. Liu, W.; Guo, G.; Chen, F.; Chen, Y. Meteorological pattern analysis assisted daily PM2.5 grades prediction using SVM optimized by PSO algorithm. Atmos. Pollut. Res. 2019, 10, 1482–1491. [Google Scholar] [CrossRef]
  55. Bonah, E.; Huang, X.; Yi, R.; Aheto, J.H.; Yu, S. Vis-NIR hyperspectral imaging for the classification of bacterial foodborne pathogens based on pixel-wise analysis and a novel CARS-PSO-SVM model. Infrared Phys. Technol. 2020, 105, 103220. [Google Scholar] [CrossRef]
  56. Ch, S.; Anand, N.; Panigrahi, B.K.; Mathur, S. Streamflow forecasting by SVM with quantum behaved particle swarm optimization. Neurocomputing 2013, 101, 18–23. [Google Scholar] [CrossRef]
  57. Li, B.; Li, D.; Zhang, Z.; Yang, S.; Wang, F. Slope stability analysis based on quantum-behaved particle swarm optimization and least squares support vector machine. Appl. Math. Model. 2015, 39, 5253–5264. [Google Scholar] [CrossRef]
  58. Vapnik, V.N. The Nature of Statistical Learning Theory; Springer: New York, NY, USA, 2000. [Google Scholar]
  59. Long, H.J.; Wang, L.Y.; Shen, J.; Wu, M.X.; Jiang, Z.B. Product service system configuration based on support vector machine considering customer perception. Int. J. Prod. Res. 2013, 51, 5450–5468. [Google Scholar] [CrossRef]
  60. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  61. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn 1995, 20, 273–297. [Google Scholar] [CrossRef]
Figure 1. Research framework.
Figure 1. Research framework.
Sustainability 13 09450 g001
Figure 2. PCA-QPSO-SVM optimization process.
Figure 2. PCA-QPSO-SVM optimization process.
Sustainability 13 09450 g002
Figure 3. Features frequency diagram.
Figure 3. Features frequency diagram.
Sustainability 13 09450 g003
Figure 4. Features box plots.
Figure 4. Features box plots.
Sustainability 13 09450 g004
Figure 5. Feature relationship heat map.
Figure 5. Feature relationship heat map.
Sustainability 13 09450 g005
Figure 6. Feature correlation graph.
Figure 6. Feature correlation graph.
Sustainability 13 09450 g006
Figure 7. The principal component ratio chart.
Figure 7. The principal component ratio chart.
Sustainability 13 09450 g007
Figure 8. Distribution of PCA dimensionality reduction dataset.
Figure 8. Distribution of PCA dimensionality reduction dataset.
Sustainability 13 09450 g008
Figure 9. The heat map of the relationship between principal components and eigenvalues.
Figure 9. The heat map of the relationship between principal components and eigenvalues.
Sustainability 13 09450 g009
Figure 10. QPSO-SVM parameter optimization.
Figure 10. QPSO-SVM parameter optimization.
Sustainability 13 09450 g010
Figure 11. Model performance.
Figure 11. Model performance.
Sustainability 13 09450 g011
Table 1. Descriptions of requirement features.
Table 1. Descriptions of requirement features.
FeatureOptionCode
CN1Environmental protection{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN2Stability{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN3Intelligence{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN4Simplicity{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN5Convenience{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN6Adaptability{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN7Reliability{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN8Comfort{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN9Energy saving{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN10Safety{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
CN11Heat dissipation{L, ML, M, MH, H}{−2, −1, 0, 1, 2}
Table 2. Product module description.
Table 2. Product module description.
Product ModuleInstanceCode
CompressorPermanent magnet synchronous frequency conversion screw typeA1
Photovoltaic direct-drive frequency conversion centrifugalA2
DC frequency conversionA3
Permanent magnet synchronous frequency conversion centrifugalA4
CondenserWater-cooled condenserB1
Air-cooled condenserB2
Evaporative condenserB3
EvaporatorHorizontal evaporatorC1
Vertical tube evaporatorC2
Throttling partsCapillaryD1
ThrottleD2
FanAxial fanE1
Centrifugal fanE2
ReservoirunidirectionalF1
BidirectionalF2
VerticalF3
HorizontalF4
Filter drierLoose filling dry filterG1
Block filterG2
Compact bead dryer filterG3
Cooling TowerDry cooling towerH1
Temperature cooling towerH2
Table 3. Service module description.
Table 3. Service module description.
Service ModuleInstanceCode
Recycling serviceHome inspectionI1
High price recyclingI2
Cash transactionI3
Maintenance serviceAnnual maintenanceJ1
Quarterly maintenanceJ2
Monthly maintenanceJ3
Spare parts serviceOriginal parts supplyK1
Non-original parts supplyK2
Replacement of faulty spare partsK3
Spare parts upgradeK4
Install serviceRemote installation and debuggingL1
On-site installation and commissioningL2
Fully commissioned installation and commissioningL3
Control Technology ServiceAdaptive location and weatherM1
Self-regulation of demandM2
Predictive self-diagnosisM3
Cleaning serviceDuct cleaningN1
Parts cleaningN2
Cooling tower cleaningN3
Condenser cleaningN4
Table 4. The original data.
Table 4. The original data.
SamplesInputsOutputs
CN1CN2CN3CN4CN5CN6CN7CN8CN9CN10CN11
0100212−11121−11
0222−1−1−112121−13
03111200212205
0411020−12012−12
05−112112120−124
……………………………………………………………………
960−1112−10021−11
9700202−12120−11
980202102112−22
991212102122−12
10012210−1111215
Table 5. PCA dimensionality reduction data.
Table 5. PCA dimensionality reduction data.
SamplesInputsOutputs
X1X2
01−0.224301033−2.4042749491
02−1.7995168653.144920113
03−0.986306315−0.2902091735
04−1.898471273−0.4228950362
053.8827740240.5807983514
……………………
96−0.270004463−2.444215151
970.094105412−1.8243768291
98−1.867236441−0.2506690222
99−1.777916303−0.4431489832
100−0.14995178−0.5557396065
Table 6. QPSO parameter settings.
Table 6. QPSO parameter settings.
ParameterSettings
Number of particles50
Particle dimension2
The maximum number of iterations50
Alpha0.6
Maximum parameter15
Minimum parameter0.01
Fitness function2-fold CV classification accuracy
Algorithm stop conditionThe number of iterations > 50
Table 7. Precision comparison of PCA-PSO-SVM, PSO-SVM, GA-SVM, and GS-SVM.
Table 7. Precision comparison of PCA-PSO-SVM, PSO-SVM, GA-SVM, and GS-SVM.
ModelPCA-QPSO-SVMPCA-PSO-SVMPSO-SVMGA-SVMGS-SVM
Number of tests2525252525
Number of errors029910
Tests accuracy100%92%64%64%60%
Mean square error01.01.041.563.32
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cui, Z.; Geng, X. Product Service System Configuration Based on a PCA-QPSO-SVM Model. Sustainability 2021, 13, 9450. https://doi.org/10.3390/su13169450

AMA Style

Cui Z, Geng X. Product Service System Configuration Based on a PCA-QPSO-SVM Model. Sustainability. 2021; 13(16):9450. https://doi.org/10.3390/su13169450

Chicago/Turabian Style

Cui, Zhaoyi, and Xiuli Geng. 2021. "Product Service System Configuration Based on a PCA-QPSO-SVM Model" Sustainability 13, no. 16: 9450. https://doi.org/10.3390/su13169450

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop