1. Introduction
Competition in manufacturing industries is intensifying around the world, and these industries are facing changes, such as the diversification of customer requirements, shorter product life cycles, and the transition to a multiproduct small-volume production system. In addition, a new area, such as smart manufacturing, is emerging with the 4th Industrial Revolution. To enhance manufacturing competitiveness and improve productivity, several manufacturers are focusing on reducing the production lead time and minimizing inventory. To achieve these goals, production management and scheduling techniques have emerged as core competencies [
1]. Numerous scheduling problems have been studied for various types of products that require assembly operations, considering the production of components and subsequent assembly processes together [
2]. For example, Lee et al. [
3] introduced a scheduling study of a manufacturing system that produces and assembles the body and vehicle of a fire engine, and Potts et al. [
4] dealt with a scheduling problem in make-to-order manufacturing systems, such as PC assembly. In addition, there are scheduling problems with assembly operations arising from various manufacturing systems (refrigerators [
5], clothing [
6], food [
7], and semiconductors [
8,
9,
10,
11]).
This study deals with a scheduling problem in a two-stage assembly-type flow shop. The first stage is the component production process in which each component is independently fabricated on a dedicated machine. The components made in the first phase are moved to the second stage and assembled into the final product. The assembly operation in the second stage can start after all necessary components have completed in the first stage. The completion time of each product is defined as the time when the assembly operation completes in the second stage. To enhance the flow shop productivity, this study aims to minimize the makespan, which is equal to the completion time of the last scheduled job.
This study constrains the waiting times between stages 1 and 2. This implies that components completed in the first stage must be entered into the second stage for the assembly operation within the limited waiting times. Therefore, it is necessary to complete the operations of the first stage considering the time when the assembly operation can start in the second stage.
Figure 1a,b illustrates example schedules with two jobs in the assembly flow shop with and without limited waiting times, respectively. The starts in the first stage in
Figure 1a are delayed observing the limited waiting times
w21 and
w22, whereas the second job in
Figure 1b can start without any delay. In general, waiting times are limited to prevent quality deterioration owing to increased waiting times. There are various cases considering the waiting time constraints in manufacturing industries, such as semiconductors, batteries, food, steelmaking, and biotechnology [
12,
13,
14].
Since Lee et al. [
3] introduced a two-stage assembly-type flow shop problem, there have been many studies on the scheduling problems, and two review papers (Komaki et al. [
2] and Framinan et al. [
15]) were recently published. Many studies have sought to minimize the makespan. The makespan minimization problem was demonstrated to be strongly NP-hard in [
3], even though only two machines existed in the first stage. Lee et al. [
3] also developed three heuristics based on Johnson’s algorithm [
16] and suggested a branch-and-bound (B&B) algorithm. Potts et al. [
4] showed that an optimal solution for the problem with more than two machines in the first stage is always in permutation schedules where the sequences on all machines in the flow shop are the same. Additionally, Hariri and Potts [
17] and Haouari and Daouas [
18] provided B&B algorithms, while Sun et al. [
19] proposed heuristic algorithms. On the other hand, Koulamas and Kyparisis [
20] considered the extended problem in which there are three stages and they developed heuristic algorithms. Additionally, Sung and Juhn [
21] proposed a B&B algorithm for the problem of outsourcing one type of component subject to the job-dependent lead time in the first stage. Wu et al. [
22] addressed two scenario-dependent jobs processing times and proposed a B&B algorithm and metaheuristics to minimize the robust makespan, and Wu et al. [
23] considered a processing time-based learning effect and proposed a B&B algorithm and several metaheuristics. On the contrary, for other objectives, Lee and Bang [
24] proposed a B&B algorithm to minimize total tardiness, whereas Lee [
25,
26] developed B&B algorithms to minimize total completion times and total weighted tardiness, respectively. Additionally, there have been various studies on the assembly-type flow shop problems [
27,
28,
29,
30,
31,
32].
Although typical assembly-type flow shop scheduling problems have been studied by many researchers, only a few studies have considered waiting time constraints or time lags. According to recent review papers, it is necessary to study scheduling problems with waiting time constraints, but there are only a few studies that considered no-wait constraints. Mozdgir et al. [
33] introduced a no-wait two-stage assembly flow shop and proposed three metaheuristic algorithms: a genetic algorithm, differential evolution algorithm, and population-based variable neighborhood algorithm. In addition, Ji et al. [
34] provided a hybrid particle swarm optimization algorithm, and Li et al. [
35] proposed an iterated local search method. Shao et al. [
36] suggested an iterated local search method and a variable neighborhood search. Most of the studies related to the no-wait assembly flow shop considered metaheuristic algorithms [
37,
38,
39,
40] as solution methodologies because these problems are NP-hard.
The considered scheduling problem is defined as
AF(
m,1)|
max-
wait|
Cmax in three-field notation suggested by Graham et al. [
41].
AF(
m,1) represents an assembly-type flow shop where there are
m machines to fabricate components in stage 1 and a single assembly machine in stage 2.
Max-
wait denotes the limited waiting constraints between the two stages.
Cmax represents the makespan, which is the objective function of the scheduling problem. Assuming that there is only one machine in the first stage, this scheduling problem becomes a two-machine flow shop problem with limited waiting time constraints (
F2|
max-
wait|
Cmax), which has been proven to be NP-hard in [
42]. Therefore, this problem is also NP-hard.
To the best of the authors’ knowledge, no assembly flow shop study has considered the limited waiting time constraints that are generalized from no-wait. Thus, as its main contribution, this study is the first attempt to consider the assembly flow shop with limited waiting times to fill part of the existing research gap. Additionally, this paper proposes various solution methodologies, including the mathematical formulation, priority rule-based list scheduling, constructive heuristic, and three metaheuristic algorithms. Furthermore, we suggest a method that converts a sequence obtained by the heuristics into a complete schedule with waiting time constraints.
This study aims to minimize the makespan in a two-stage assembly flow shop with limited waiting time constraints. We provide a mixed-integer programming (MIP) to define the considered problem clearly and use a commercial optimization solver, CPLEX 12.10, to solve the MIP. However, because this problem is NP-hard, there is a limitation in obtaining an optimal solution despite using the solver. Thus, we propose various heuristic algorithms, such as priority rule-based list scheduling, constructive heuristic, and metaheuristic to solve the large-sized problem within a short computation time. To evaluate the proposed algorithms, we performed a series of computational experiments, including the calibration of the metaheuristics, on randomly generated problem instances and reported the results.
The remainder of this paper is organized as follows.
Section 2 provides the assumptions made in this study and notation for a clear description and provides a mathematical model, while
Section 3 proposes heuristic algorithms.
Section 4 reports the computational tests to evaluate the performance of the heuristic algorithms as well as the mathematical model. Finally,
Section 5 concludes this study and presents future research.
4. Computational Experiments
This section evaluates the proposed algorithms on randomly generated problem instances. All the algorithms were coded in Java, and the tests were conducted on a personal computer with Intel Core i7-8700 CPU (3.2 GHz) and 16 GB RAM. To generate problem instances, we used the same method in [
25]. In the first stage, there are
m dedicated machines, and three levels of
m were used, that is,
m = (2, 5, 10), whereas there is a single machine in the second stage. Thus, there are a total of (
m + 1) machines in the tests. The processing times were generated from a uniform distribution.
Table 2 shows the discrete uniform distribution
U(
a,
b) with a range between
a and
b. In Set A, the workloads of the two stages are equal. On the contrary, in Sets B and C, the second and first stages become bottlenecks, respectively. Limited waiting times were also generated from
U(1, 100).
The first experiment evaluated the mixed integer programming formulation (MIP) through a commercial optimization solver CPLEX 12.10. In the experiment, we considered five levels of jobs (
n = 10, 20, 30, 40, and 50) and generated five instances for each of the (
m,
n) combinations. In addition, the computation time limit (
TL) for each instance was set to 1800 s to avoid excessive computation times. A summary of the results is presented in
Table 3, including the average computation times and the number of instances not terminated until the time limit
TL. As can be observed from the table, the ACPUT in Set A increased exponentially as (
m,
n) increased. For
m = (5 and 10), CPLEX found an optimal solution only for instances with
n = 10 within 1800 s. On the contrary, CPLEX solved instances of Sets B and C relatively faster than those in Set A. One drawback of CPLEX in this test is its wide range of computation times. For (
m = 10,
n = 20) in all three sets, the minimum CPU time was only a few seconds, whereas the maximum time was 1800 s. If CPLEX shows a significant range of CPU times, the estimates for solving times are inaccurate and even impossible; hence, using CPLEX may be impractical.
Next, we tested the priority rule-based list scheduling methods to determine the best method for offering an initial seed sequence for MNEH. The same instances with n = 10, 20, 30, 40, and 50 were used in these tests. To compare the list scheduling methods, the relative deviation index (RDI) was used as a measure, defined as (obj# − objbest)/(objworst − objbest) for each instance where obj# represents the makespan using priority rule #, and objbest and objworst represent the best and worst, respectively. According to the definition of RDI, the lower the RDI, the better the performance.
A summary of the results is in
Table 4 which provides the average RDI and number of instances (out of 75) that LS# found the best solutions. As observed from the table, LS1 outperformed the other rules. This indicates that the first stage on which LS1 focuses is more important than the second stage, probably because the decision on sequencing jobs on the parallel machines in the first stage must consider the limited waiting time constraints. However, in the second stage, there is a single assembly machine, and the sequence is equivalent to the sequence in the first stage because this study considers only permutation schedules. Thus, if a shop manager wants to focus on either stage, the result recommends the first stage. In summary, LS1 showed the best performance among the proposed list scheduling rules, therefore LS1 will offer an initial seed sequence for MNEH in the subsequence experiments.
To check the efficiency of the interchange method in the proposed MNEH, comparison tests were performed on the instances with up to
n = 50. In this test, initial seed sequences were obtained using LS1.
Table 5 and
Figure 6 show the results that show the effectiveness of the interchange method in MNEH. The table reports the percentage change defined as 100 × (
objMNEH −
objNEH)/
objNEH. Note that the negative percentage change indicates that MNEH found a better solution than NEH. As revealed in the table, MNEH worked better than NEH, which indicates that the simple interchange method can improve the solution quality. In addition, the improvement was more apparent when the problem size increased. The same result was obtained in Set A than in Sets B and C as shown in
Figure 6. Thus, it can be concluded that the interchange method used in MNEH worked well. In addition, although the interchange procedure was added, MNEH provided a solution for all instances in less than a second.
To achieve the best performance of the proposed metaheuristic algorithms (GA, IG, and SA), calibration experiments were performed. In general, calibration before using a metaheuristic is essential because the performance depends significantly on the control parameters. In other to calibrate the metaheuristics, we generated three instances for each combination of (m, n), with three levels of machines (m = 2, 5, and 10) and four levels of jobs (n = 50, 100, 300, and 500), on Set A. The algorithms solved each instance independently three times, and the average value resulting from the three runs was used to compute the RDI for comparison. For the termination condition, we set the maximum CPU time defined as tmax = n(m + 1)(tf)/2 ms, where tf denotes the time factor and tf = 60 in this calibration.
The proposed GA has eight operators and parameters which are listed below.
Selection: two levels (tournament and roulette).
Crossover: six levels (OX1, OX2, SJX, SJX2, SBX, and SBX2).
Mutation (pIS): five levels (0, 0.25, 0.5, 0.75, and 1).
Population size (Psize): five levels (20, 30, 40, 50, and 60).
Crossover probability (pc): five levels (0.0, 0.1, 0.2, 0.3, and 0.4).
Mutation probability (pm): five levels (0.6, 0.7, 0.8, 0.9, and 1.0).
Local search (Pl): five levels (30, 40, 50, 60, and 70).
Restart (Pr): five levels (50, 60, 70, 80, and 90).
For the operators and parameters, 468,750 combinations were possible. To simplify this calibration, we performed two separate experiments. The first experiment was to select the best combination for selection, crossover, and mutation, and the second experiment was to find the best combination for the other parameters.
For the first experiment, a full factorial design for selection, crossover, and mutation was considered; hence, 60 different algorithms were tested. The remaining parameters were fixed as (
Psize,
pc,
pm,
Pl,
Pr) = (50, 1, 1, ∞, ∞). These values were intended to make the parameters meaningless and to focus on the effects of selection, crossover, and mutation on the solutions.
Figure 7 illustrates the main effect plot to determine the best operators with lower RDI values. As shown in the figure, (tournament, OX1 and 0.5) showed better performance. This was likely because many changes by crossover operation did not lead to better sequences in this problem with limited waiting time constraints. For mutation, the use of insertion and interchange with the same probability provided good results. Therefore, tournament, OX1 and
pIS = 0.5 were used in subsequent experiments.
The second experiment examined the remaining parameters (
Psize,
pc,
pm,
Pl,
Pr). The full factorial design of the parameters could make 3125 combinations. However, because testing these combinations requires significant time, we adopted the Taguchi L
25 orthogonal array design instead of a full factorial design. The L
25 design is provided in
Table 6; that is, the 25 combinations provided in the table were tested. Note that the Taguchi design developed by Genichi Taguchi can achieve parametric research and optimization while reducing the number of experiments or numerical tests [
54].
Figure 8 shows that the lowest RDIs for the parameters were achieved at (
Psize,
pc,
pm,
Pl and
Pr) = (30, 0.2, 1, 60 and 60). Remarkably, the probabilities of crossover and mutation were opposite. With these probabilities, the GA occasionally executes crossover, but it always performs mutation operations. Considering the calibration results, i.e., OX1 and
pc = 0.2, the mutual interchange between two solutions may not help search for better solutions to this problem. In contrast, the mutation was more useful for improving the solution. The values of
Pl and
Pr were the same as 60. This indicates that the local search and restart procedures start when the best solution did not improve during 60 generations.
The suggested IG has two parameters:
d and
pIS. This calibration considered five levels of each parameter, that is,
d = (6, 8, 10, 12, and 14) and
pIS = (0, 0.25, 0.5, 0.75, and 1). Thus, we tested 25 combinations. A summary of the results is presented in
Figure 9. As shown in the figure, RDI had the lowest at
d = 10 and the difference in RDIs between
d values was very clear. This value differed significantly from the original value of 4 which was experimentally found in [
47]. It is likely because computing power improved compared to in the past. Note that the destruction size
d is determined by considering the trade-off between the solution quality and computation time. That is, a large
d can provide good solutions but requires a large amount of computation. Therefore, in a limited CPU time, if the
d value is too large, IG may provide a poor solution. The effect of the
pIS was not significant. This may be because IG is based on the insertion mechanism. Consequently, the IG uses
d = 10 and
pIS = 0.75.
For calibration of the proposed SA, we considered five levels of each of the three parameters (
α,
L, and
pIS). In this experiment, the epoch length
L was defined as
L =
n ×
l, where
l is a control parameter. The design of this calibration was also based on the Taguchi L
25 orthogonal array (
Table 7).
Figure 10 shows the calibration results. Three parameters (
α,
l, and
pIS) had the lowest RDI values at (0.995, 15, and 0.25), respectively. In contrast to GA and IG, the results showed bath-curves for all the parameters. In other words, both higher and lower values from the best values resulted in worse outcomes. This is probably because SA generates a neighborhood solution using only a local search of insertion and interchange without any other methods. From the value of
pIS = 0.25, the interchange appears to be more effective in SA. Therefore, the SA used
α = 0.995,
l = 15, and
pIS = 0.25 in subsequence tests.
After calibrating the proposed metaheuristics, we examined the solutions obtained from the heuristics by comparing them with those from CPLEX. Here, we considered four levels of
n = 20, 30, 40, and 50 because the destruction size
d of IG was determined to be 10 in the calibration. Metaheuristics independently solved each instance five times, and the average values of the five independent results were used. In addition, for the termination condition
tmax n(
m + 1)(
tf)/2 ms, three levels of
tf = (30, 60, and 90) were considered, and let
Atf be the algorithm
A using
tf = (30, 60, and 90).
Table 8 summarizes the comparison results, which shows the average percentage gaps of heuristic solutions from a solution using CPLEX, and
Figure 11 represents the number of instances for which the heuristic solutions are better than or equal to the CPLEX solutions. The negative percentage gap indicates that the heuristic solutions are better than those of the CPLEX. Overall, the IG algorithm exhibited the best performance. For 179 out of 180 instances, IG found solutions better than or equal to those of CPLEX. Both algorithms were better than CPLEX, and SA outperformed GA. For the three levels of
tf, only a few changes were observed in all cases. This indicates that the proposed metaheuristics can find a solution close to an optimal solution within a short computation time for instances of up to
n = 50. These experimental results verified the effectiveness and efficiency of the proposed metaheuristics. Meanwhile, MNEH showed percentage gaps of approximately 2% in Set A and less than 1% in Sets B and C. Although MNEH was overwhelmed by the metaheuristics, its performance is acceptable, bearing in mind that MNEH is a constructive heuristic.
We also examined the performance of the heuristics on large instances with
n = 100, 200, 300, 400, and 500. A summary of the comparison results is in
Figure 12 which shows the RDI values of the heuristic algorithms and
Figure 13 which represents the number of instances that the heuristic found the best solution. Overall, SA outperformed the other algorithms, and even SA
30 was better than GA
90 and IG
90. These results were remarkably different from those of the previous test with small instances of up to
n = 50. This is probably because of the trade-off relationship between solution quality and computation time. In IG, the greedy reinsertion method checks all possible positions to find the best one, and thus, this mechanism is very effective in small instances. However, this technique becomes computationally burdensome when
n increases. In contrast, the SA procedure, which uses only a local search method with insertion and interchange, is relatively simple. That is, SA generates a new solution by slightly changing the current solution, and the time to create a new solution rarely increases even if
n increases. Therefore, in terms of the trade-off relation, it can be said that SA is much more effective and efficient. On the contrary, for the instances with
m = (2 and 5) of Sets B and C, there was no difference in the performance of all algorithms, including MNEH. However, for instances with
m = 10, SA overwhelmed the other algorithms, and IG performed better than GA. A summary of the experimental results thus far reveals that attempts to improve the solution by changing the sequence slightly are desirable in this scheduling problem.
In addition, we performed statistical tests to see statistically significant differences in the performance of the proposed metaheuristics. This test used RDI values obtained from GA
90, IG
90, and SA
90 on large-size instances of
n = 100, 200, 300, 400, and 500.
Figure 14 shows the 95% confidence interval for the mean RDIs of the algorithms. As shown in
Figure 14, SA
90 performed significantly better in terms of the mean RDI. We also performed Kruskal–Wallis tests which is a non-parametric method.
Table 9 shows the results of pairwise comparisons from the tests. For the tests in Set A, SA
90 significantly outperformed GA
90 and IG
90. However, there is no significant difference between SA and IG in Sets B and C. This is likely because the RDIs for
m = 2 and 5 were zero.
The last analysis shows the effect of the waiting time constraints on solutions. For this analysis, we generated 10 instances of (
m,
n) = (5, 300) in Set A. The waiting times were generated randomly with a range of [1, 100], and these values were treated by multiplying the ratios (0.5, 1.0, 1.5, 2.0, 2.5, and 3.0). We analyzed the results obtained by solving each instance independently five times using SA
90 which showed the best performance. As shown in
Figure 15, the makespan decreased as the waiting time ratio increased. This is because tight limited waiting times caused delayed starts; hence, the makespan increased accordingly. The decline of the makespan slowed when the ratio exceeded 2. That is, if the limited waiting times are twice the processing times, the effect of the waiting time limit on the solution is reduced. These results suggest the need for manufacturing technologies and systems that can achieve product quality while setting enough limited waiting times. It may also provide a guideline for considering the trade-off between production efficiency and product quality. Meanwhile, the range of the objective values increased as the ratio increased. This may be due to the loose waiting time limits that increased the time range in which jobs could start. In other words, tight waiting time limits reduced the solution space for start times in a problem.