Abstract
In this paper, a hybrid deep neural network scheduler (HDNNS) is proposed to solve jobshop scheduling problems (JSSPs). In order to mine the state information of schedule processing, a jobshop scheduling problem is divided into several classificationbased subproblems. And a deep learning framework is used for solving these subproblems. HDNNS applies the convolution twodimensional transformation method (CTDT) to transform irregular scheduling information into regular features so that the convolution operation of deep learning can be introduced into dealing with JSSP. The simulation experiments designed for testing HDNNS are in the context of JSSPs with different scales of machines and jobs as well as different time distributions for processing procedures. The results show that the MAKESPAN index of HDNNS is 9% better than that of HNN and the index is also 4% better than that of ANN in ZLP dataset. With the same neural network structure, the training time of the HDNNS method is obviously shorter than that of the DEEPRM method. In addition, the scheduler has an excellent generalization performance, which can address largescale scheduling problems with only smallscale training data.
1. Introduction
Jobshop scheduling problem (JSSP) [1] is one of the most famous problems in the industrial production, and it is categorized as a large class of intractable numerical problems known as NPhard [2]. The solution space for an JSSP (where m is the number of machines and n is the number of jobs) is [3].
As it will be discussed in Section 2, many scholars have tried to solve this type of problems with populationbased methods [4], genebased methods [5], and heuristic methods [6]. However, in the face of largescale problems, the response rate of the above methods has no distinct advantages. Many current researches show that data mining and machine learning methods have great potential in effect and efficiency [7]. In this paper, a hybrid deep neural network scheduler (HDNNS) is put forward to promote the scheduling capability. And convolution twodimensional transformation (CTDT) is developed to convert JSSP’s state information into regular information so that the process can be simplified in the convolutional network.
HDNNS has contributions in the following aspects:(i)Based on the work of Weckman [3], Metan et al. [8], and Paolo et al. [9], HDNNS transforms JSSP into several classification subproblems. HDNNS’s main innovation is the classification of the processing sequence of each job on each machine. The more precise classification method makes HDNNS more effective in the largescale problems.(ii)Convolution twodimensional transformation (CTDT) comes up in this paper. The function of CTDT is to convert the irregular scheduling data into regular multidimensional data with the form of Cartesian product. The transformed multidimensional data can be effectively processed in the deep convolution networks.(iii)HDNNS designs a hybrid neural network combining the deep convolution network [10] and the BP neural network [11]. In the first half of the network structure, convolution network and BP network are used to deal with the structural features and irregular features, respectively. After a certain number of layers of network processing, HDNNS merges these two networks with flattening operation for further feature extraction.
Our experimental results prove that the scheduling results of HDNNS are superior to many learningbased methods (ANN, HNN, and reinforcement learning methods), traditional classification methods (SVM, GOSS, and others), and attributeoriented induction methods (AOI) [9] for the MAKESPAN index. HDNNS can occupy an advantage in the JSSPs compared with populationbased methods (GA) and optimization methods (BBM). The value of HDNNS is not negated in the tests because GA and BBM are time consuming in computation. Besides, unlike GA and BBM, the HDNNS method has strong generalization performance. Our experiments certificate that a model trained by the data of smallscale JSSPs can address a largescale one.
Although the training of the model requires extra time, the training process can be finished in advance. When the application environment remains stable, the model may not even need further updates. Such characteristics can increase the application value of the model to a certain extent. The training process can also be effectively accelerated by hardware such as GPU. Also, with the appropriate hardware (such as GPU and FPGA), the training speed will be significantly boosted.
The structure of this paper is as follows. In Section 2, a part of the most related work on the solution methods for JSSP has been reviewed along with neural network and other approaches available in the literature. In Section 3, the mathematical model of JSSP is proposed. In Section 4, the framework of the HDNNS is introduced, which includes scheduler structure, convolution twodimensional transformation, and the basis of deep neural network. In Section 5, a 6 8 JSSP example is applied to explain our method. In Section 6, six experiments are utilized to test the effectiveness and the generalization performance of the proposed method.
2. Related Works
2.1. PopulationBased and GeneBased Methods for JSSP
Over the last decades, JSSP has attracted much attention in the academia. Hence, a wide range of approaches have been developed for JSSP. Recently, populationbased and genebased methods are investigated to find optimal or nearoptimal solutions.
Zhao et al. [12] proposed an improved particle swarm optimization with a decline disturbance index to improve the ability of particles in exploring global and local optimum solutions and to reduce the probability of particles being trapped into a local one. Peng et al. [13] combined a tabu search procedure with path relinking and showed that their method had a high performance in solving benchmark problem instances. Asadzadeh [14] tried to improve the efficiency of the genetic algorithm in solving JSSP by parallelizing populations and using an agentbased approach. Kurdi et al. [15] presented a modified island model genetic algorithm (IMGA) for JSSP. In this model, a natureinspired evolutionary method and a migration selection mechanism have been added to the classical IMGA to improve diversification and delay premature convergence. Park et al. [16] proposed a dynamic JSSP and applied genetic programmingbased hyperheuristic methods with ensemble combination schemes to solve it. The investigated schemes had majority voting, linear combination, weighted majority voting, and weighted linear combination. It was concluded from the experiments that for the dynamic JSSP, the linear combination outperformed the other methods. Jiang et al. [17] employed the grey wolf optimization (GWO) to deal with two combinatorial optimization problems in the manufacturing field: jobshop and flexible jobshop scheduling cases. The discrete GWO algorithm was compared with other published algorithms for two scheduling cases. Experimental results demonstrate that our algorithm outperforms other algorithms for the scheduling problems under study. Fu et al. [18] proposed a fireworks algorithm with special strategies to solute the flowshop scheduling problem under the consideration of multiple objectives, timedependent processing time, and uncertainty. Sharma et al. [19] developed a variant of the ABC algorithm inspired from beer froth decay phenomenon to deal with jobshop scheduling problems.
There is no doubt that populationbased and genebased strategies are effective to solve JSSPs. However, faced with largescale problems, the number of repeated iterations and updating operations often take a long time. Therefore, it is of great value to study a learningbased scheduler with fast response.
2.2. LearningBased and Neural NetworkBased Methods for Solving JSSP
With the further development of machine learning, some scholars try to solve JSSPs with learningbased methods. In this field, researches can be divided into two categories.
In the first category, learning methods are used to optimize populationbased and genebased methods. Learning methods optimize the updates of solutions, which thus improve the efficiency of optimization. Yang and Lu et al. [20] proposed a hybrid dynamic preemptive and competitive NN approach called the advanced preventive competitive NN method. A CNN was used to classify the system conditions into 50 groups. For each production interval, the current system status group was determined by CNN. Shiue et al. [21] extended the previous work by considering both the input control and the dispatching rule, such as those in a wafer fabrication manufacturing environment. In a novel recent work by Mirshekarian and Sormaz [22], a statistical study of the relationship between JSSP feature and optimal MAKESPAN was conducted. Ramanan et al. [23] proposed an artificial neural networkbased heuristic method. This method utilized ANN to generate a solution of JSSP and then took it as the initial sequence to a heuristic proposed by Suliman. Adibi et al. [24] used a trained artificial neural network (ANN) to update parameters of a metaheuristic method at any rescheduling point in a dynamic JSSP according to the problem condition. Maroosi et al. [25] proposed an approach which utilizes the parallel membrane computing method and the harmony search method to solve flexible job shop problems. Information from the best solutions was used to boost the speed of convergence while preventing premature convergence to a local minimum.
In the second category, a reinforcement learning or machine learning framework is applied to build a learningbased model (ANN [11], SVM [26], CNN [27], or others). Then, the model is trained to master scheduling rules and complete automatic scheduling tasks. Weckman et al. [3] developed a neural network (NN) scheduler for JSSP in which the genetic algorithm was used to generate optimal or nearoptimal solutions for a benchmark problem instance, and then, an NN was used to capture the predictive knowledge regarding the sequence of operations. Chen et al. [28] proposed a ruledriven dispatching method based on data envelopment analysis and reinforcement learning for the multiobjective scheduling problem. Mao et al. [29] presented the deep reinforcement learning method (DEEPRM) and translated the problem of packing tasks with multiple resource demands into a learning problem. This solution has an essential inspiration for solving the JSSP. Moreover, the initial results show that DEEPRM performs comparably to stateoftheart heuristics, adapts to different conditions, converges quickly, and learns strategies that are sensible in hindsight. Shahrabi et al. [30] proposed a reinforcement learning (RL) with a Qfactor algorithm to enhance the performance of the scheduling method proposed for dynamic JSSP which considered random job arrivals and machine breakdowns. Nasiri et al. [31] used discrete event simulation and multilayer perceptron artificial neural network to solve the openshop scheduling problem. Mohammad et al. [9] proposed a data miningbased approach to generate an improved initial population for populationbased heuristics solving the JSSP. This method applied a combination of “attributeoriented induction” and “association rule mining” techniques to extract the rules behind the optimal or nearoptimal schedules of JSSP. Finally, their experiments verify the significant amount of FEs that can be saved using the proposed approach and the superiority of the proposed method in comparison with the method of Koonce and Tsai [32].
According to the retrospective literature, none of the previous studies directly applied deep learning frameworks to JSSP. This paper creates a convolution twodimensional transformation and designs network structure to solve JSSP.
3. Mixed Integer Programming Model of JSSP
Jobshop scheduling problem (JSSP) can be described as a mixed integer programming problem.
The mathematical description is [33]
The decision variables are defined as follows:(i) is equal to 1 if job j is scheduled at the kth position on machine i(ii) denotes the start time of the job at the kth position of machine i
The parameters are defined as follows:(i)J is the set of the jobs, and M is the set of the machines(ii)n is the number of the jobs, and (iii)m is the number of the machines, and (iv) is a nonnegative integer which represents the processing time of job j and machine i(v) if the kth position of job j requires machine i
The objective function is in (1). Constraint (2) ensures that each position on each machine is assigned to exactly one job. Constraint (3) ensures that each job only gets one position on a machine. Constraint (4) states that the start time of a job on a machine should be larger than the completion time of the job scheduled at the previous position. Constraint (5) is the precedence constraint. It ensures that all operations of a job are executed in the given order. In (5), V is since the completion time of any operation cannot exceed the summation of the processing times from all the operations. Constraint (6) ensures that the MAKESPAN is at least the largest completion time of the last job on all machines. Constraint (7) ensures that the start time of all jobs at all positions is greater or equal to 0.
4. Hybrid Deep Neural Network Scheduler
4.1. Scheduler Structure
A hybrid deep neural network scheduler (HDNNS) is designed based on convolution twodimensional transformation (CTDT) and hybrid deep neural network.
The structure of the scheduler is shown in Figure 1.
HDNNS is divided into two sections: training section and scheduling section.
The training section has six steps (Step 1.1–Step 1.6). First, a large number of JSSPs are generated according to the JSSP description in Step 1.1. The description includes the number of machines m, the number of jobs n, and the distribution function of processing time . Next, the generated problems are solved by stateoftheart methods (BBM or GA in this paper). Moreover, corresponding scheduling results are generated in Step 1.2. In Step 1.3, each JSSP is divided into several subproblems, described as the features of a job processing and the priority in the machine. Features of job processing generate the 1D and 2D input data with CTDT in Step 1.4. Moreover, the priority in the machine generates onehot target data in Step 1.5. Finally, the scheduler training is in Step 1.6.
The training section has five steps (Step 2.1–Step 2.5). First, Step 2.1 is started when a new JSSP requires to schedule. Then, 1D input and 2D input can be produced by generating subproblem operation (same as Step 1.3) and convolution twodimensional transformation operations (same as Step 1.4). In Step 2.4, we use a trained neural network to obtain the priority of each process in each job corresponding to the input of two groups of the neural network. In Step 2.5, a complete scheduling result is created with all priority results taken into account.
4.2. Mathematical Representation of Standard Solver and Division of Subproblems
Combined with the MIP description of JSSP in Section 3, all solvers are abstracted as follows:
In (9), X is the set of 01 decision variables , H is the set of integer decision variables , P is the set of processing time data , and R is the set of operation requiring data . And can be any scheduler for JSSP, such as genetic algorithm (GA) [14], branch and bound method (BBM) [34], and tabu search algorithm [13].
In order to improve the generalization performance of the model, HDNNS classifies a complete JSSP into several subproblems. Specifically, each subproblem determines the priority category on machine of the job processing process:
In (10), is the processing feature of job j’s kth position in machine i and the relationship between the job’s position and the machine is given by R. is a subproblem scheduler from the in (9), and is the integer priority of job processing on the machine (if in schedule result X, job j is processed in the order k in machine i, then ). The generation of and will be introduced in Sections 4.3 and 4.4.
The subproblem generation process is shown in Figure 2.
4.3. Convolution TwoDimensional Transformation
4.3.1. Definition of OneDimensional Features
This paper designs a convolution twodimensional transformation (CTDT) to extract scheduling features. Convolution operation is commonly used to extract features in the field of artificial intelligence and image processing [35, 36]. Many scholars believe that deep convolution operation is an effective way to extract complex combined features [37]. The CTDT is proposed to transform the irregular data in scheduling process (which cannot be convoluted directly) into regular data by the form of Cartesian product.
First, we define the 1dimensional matrix relative machine processing time from P as
In (11), is the ratio of processing time of job to that of job , which are represented as follows: will provide the scheduler with relative information about the processing time of jobs.
Then, we define the 1dimensional matrix’s earliest start time from P and R as
In (13), , is the earliest start time of job j’s kth position shown as follows:
In (14), provides the urgency information of jobs.
Similarly, we define the 1dimensional other features as
In (15), consists of a series of important features in reference and application. is the number of the features, and in this paper, . The features are given in Table 1.
In Table 1, the variables in tables are defined as follows:
In (16)–(18), is the total processing time, is the processing time of machine i, and is the processing time of job j.
4.3.2. Convolution TwoDimensional Transformation and Definition of TwoDimensional Matrix
Cartesian product operation can combine linear features and convert onedimensional feature data into twodimensional feature data. This paper designs convolution twodimensional transformation (CTDT) based on Cartesian product.
The transformation is described in
In (19), and are the two onedimensional features and is the sign of Cartesian product; the mathematical definition is shown in (20). α and β are the parameters of this transformation. is a nonlinear activation function. This function will match the model parameters and extract new features in different horizons. The sigmoid function is shown in
In (21), x is a matrix.
An example of a function is shown in Figure 3.
In Figure 3, and are the two onedimensional data like , , and in Section 4.3.1. The Cartesian product of and is . Three sets of parameters are used to normalize in Figure 3. Different parameters mean that the model pays attention to different data scales, which helps the model to discover the characteristics of different scales.
4.4. Training Labels
HDNNS transforms the scheduling problem into classification problems. So, this paper uses onehot encoding [38] to define training labels.
Job i’s kth position (one machine j) onehot priority label is shown in (22). Three examples are given in Figure 4:
In (22), is the number of positions in job i machine j, defined in Section 4.2.
4.5. Structure of Hybrid Deep Neural Network Scheduler
In this section, an innovative hybrid deep neural network structure for JSSP is introduced.
As shown in Figure 1, the inputs of the hybrid deep neural network scheduler are onedimensional input , twodimensional input , and target input .
The expression is shown in
The general structure of the network is shown in Figure 5.
In Figure 5, the left side of the structure diagram is the input part of the network.
For , HDNNS uses layers (fully connected layer) [39] (FCL in the figure) to preliminarily extract onedimensional features. As shown in Figure 5, the output of the th layer is defined as and the output of the final layer is . The fully connected layer is a typical combination of neurons in the deep convolution network [39].
For , HDNNS uses layers (convolutional layer) [39] (CL in the figure) to preliminarily extract twodimensional features. The size of the convolution kernel [39] is set to 3 3. As shown in Figure 5, the output of the th layer is defined as to in different features in (23) and the weight of the th layer is defined as to .
At the layer of the network, onedimensional features and twodimensional features are combined by flattening operation in the flattened layer [40], described as
In (24), are the network weights of layers , and q is a neural index.
After fully connected layers, the feature passes through a Softmax layer [39] containing only n neurons. This layer converts the feature signal into a meaningful probability description . has the same shape with the target . However, is not a 01 variable, but a continuous quantity, which satisfies , where . can be interpreted as the possibility of selecting priority p.
After defining the structure of the neural network, we use the error backpropagation (BP) method [39] to train the network parameter.
A trained neural network can be described as a function mapping in the scheduling section of Figure 1, which is shown in the following formula:
In (25), is the deep neural network scheduler and the input function is and in Figure 1 and (23). The is the possibility that the current subproblem belongs to each priority. There are n priorities, so there are n elements in , each of which is described as
4.6. Scheduling Sequence Generation Method
In (10), the whole problem is decomposed into several subproblems. In this part, a scheduling sequence generation algorithm combines the solutions of the subproblems into a complete solution of JSSP.
The pseudocode description of the method is shown in Algorithm 1.

In Algorithm 1, each cycle for i will determine the scheduling order of one machine. Each cycle for j will determine the scheduling sequence of one job in the machine i.
When determining the order of jobs, in Step 7, the algorithm first chooses the most assured judgment of the neural network scheduler, and the most reliable judgment is the output probability closest to 1. In Step 8 and Step 9, when the job is selected as priority , the other data of job and priority are set to 0 according to constraints (2) and (3) to avoid the conflict in the next loop. In Step 9, the algorithm updates the value of the output matrix.
4.7. Generalization Performance of HDNNS
HDNNS algorithm has a reliable generalization. Specifically, we can easily extend the training results of smallerscale JSSPs (the number of machines is small) to solve largerscale JSSPs (the number of machines is significant). Such characteristics give HDNNS a unique advantage. When the solution of a largescale problem is difficult to be generated by the existing methods, HDNNS can use the solution of a smallscale problem to train the network and then use the trained model to schedule a largescale problem.
In (26), the input parameters of the trained scheduler are composed of two sets of data, one of which is onedimensional data and the other is twodimensional data generated by CTDT. For all inputs, as the number of machines increases, the input and output structures of the neural network will not change.
Although the absolute value of the parameter changes, the correlation between the parameters still exists. The scheduler will use these features with relationship to complete the scheduling. Of course, the more significant the gap between the scale of training data and the scale of actual scheduling data, the bigger the error of results. This paper will discuss it in the experiment.
5. An Example of HDNNS
In this section, we illustrate HDNNS with an example ().
In the training section of Figure 1, we generate a series of JSSP and solve them as the training data.
An example of algorithm generation of JSSP is described as Tables 2 and 3.
Tables 2 and 3 describe a 6 8 JSSP, and Figure 6 shows the Gantt chart of the optimal solution (with BBM). In Table 2, the number in line j and column k is the time required for job j’s kth position. In Table 3, the number in line j and column k is the machine required for job j’s kth position.
In Figure 6, the horizontal axis is the time axis and the ordinate axis is the machines axis. Each block represents a processing, and different colors represent different jobs. The text in the boxes means that the processing of job j’s kth position starts at the time of the left side of the block and ends at the right side of the block.
Then, 48 subproblems are generated according to (10). Onedimensional and twodimensional features are extracted for each subproblem, and training data such as (23) are generated in Table 4.
Six groups of twodimensional features are selected for visual display, and the pictures are shown in Figure 7.
(a)
(b)
(c)
(d)
(e)
(f)
Six groups of matrices generated by CTDT are shown in Figure 7. Among them, (a), (c), and (e) have a high priority and the other three have a low priority.
In this extreme case of the highest priority and the lowest priority, it is easy to find that images with the same priority have a lot in common. In general, the hue of matrix , , and in (a), (c), and (d) is darker and that of matrix , , in (b), (d), and (e) is brighter. The remaining three matrices describe the whole problem rather than the subproblem. Therefore, the same graphics are shown in different subproblems.
Although identifying similar priority categories is more difficult for human beings, our deep learningbased scheduler can effectively extract the priority information.
After training the network with the data in Table 4, we get a scheduler that can respond quickly. When a new scheduling problem arrives, the scheduler processes the problem according to (25) and gets the priority matrix O. For this problem, the output example of matrix O is shown in Table 5.
Finally, the scheduling sequence generation algorithm is used to process the output matrix and the scheduling order X and the time result in Figure 6 can be obtained.
6. Results and Discussion
6.1. Parameters and Effect Experiment
In this part, the training process of HDNNS and the influence of different parameters on HDNNS are discussed.
Dataset ZLP (7 7) [41] is used in this part to validate the effectiveness of the method effectively. ZLP (7 7) dataset contains 2000 7 7 () JSSPs, and it corresponds to solutions.
This experiment trains the scheduler with the first 1500 questions and labels and then tests the scheduler with the last 500 questions. The learning rate of the network is 0.01. The training process curve is plotted in Figures 8–10.
In Figures 8–10, the horizontal axis is the number of training loops and the vertical axis is the classification correctness, classification loss, and MAKESPAN [22] (completion time of processing). The curves of different colors represent the experimental results obtained by choosing different model parameters and . Among them, the loss evaluation index calculation formula is
In (27), is the target output, o is the probabilistic description of current features belonging to various classifications, and is the loss function. is the bool value, and this value indicates whether the target class of input features instance is k. is the probability of input features belonging to class k predicted by the model. There is a onetoone mathematical relationship between and in (26).
The three figures show that the performance of the model improves gradually with the increase of the number of training cycles. This improvement can be achieved until the classification accuracy reaches more than 90% and the model loss reaches less than 1. Moreover, the disparity between the scheduling result and the optimal solution reaches less than 5%. The above experiments show that the HDNNS can effectively train the scheduler to complete the JSSP scheduling task.
Seven groups of different model parameters were selected and tested. The experimental results show that among all the parameters, and have better results. When and , the classification accuracy of the centralized test is more than 93%, the loss is less than 85%, and the gap of MAKESPAN is less than 4%.
6.2. Confusion Matrix of the Result
In order to measure the effectiveness of HDNNS, this paper compares it with the classical ANN [3] method.
Dataset ZLP (7 7) [41] is used in this part to train two kinds of neural networks. This experiment trains the scheduler with the first 1500 questions and labels. Then, this experiment tests the scheduler with the last 500 questions. For HDNNS, the size of the convolution kernel is 3 3, and , , and learning rate is 0.01. For the ANN method, the ANN structure is 1112107 and learning rate is 0.01.
The classification confusion matrix of ANN and HDNNS (the output of Step 2.4 in Figure 1) is shown in Tables 6 and 7.
In Tables 6 and 7, the line i and column j is the number of times that the job with the ith position of the machine has been assigned to the jth position of the machine. The priority of job 2 in machine 1 is 2, meaning that this job is in the second position of machine 1’s processing. If a scheduler classifies the location of job 2 in the machine 1 as 2, one will be added to the second row and the second column of the confusion matrix. If a scheduler classifies the location of job 2 in the machine 1 as 3, one will be added to the second row and the third column of the confusion matrix. Therefore, the larger the number on the diagonal line, the higher the accuracy of the model.
The bar figure of the confusion matrix is shown in Figures 11 and 12.
Tables 6 and 7 and Figures 11 and 12 show that the classification performance of HDNNS is better than ANN. On the stability of classification, two methods can classify the highest and lowest priority jobs more accurately because the boundary of classification will introduce less noise interference. However, the classification accuracy of each priority of the HDNNS method is more stable. The accuracy of classification results of the HDNNS method fluctuates between 88% and 98%. In terms of classification accuracy, HDNNS can achieve 90% classification accuracy. It is better than 60% of the ANN method.
The essence of the learningbased method is to estimate the probability from input to output by finding the implicit relationship between them. Because the ANN method does not consider the depth of local features combination, its effect is not ideal. For example, the has no noticeable effect on its priority. However, the combination of , the , and the has a more significant impact on the final output.
The traditional neural network does not have strong ability to deal with combined features. Thus, ANN is difficult to achieve effective training because of the disappearance of the gradient [36]. In this paper, deep convolutional network is introduced into the scheduling problem to solve the problem of learning and training combined features, which improves the accuracy of network classification.
6.3. MAKESPAN and Time Consumption Comparisons in ZLP Dataset
JSSP scheduling methods are divided into two categories: populationbased (genebased) method and learningbased method. The populationbased (genebased) method obtains the nearoptimal solution by updating the solutions set. The effect of this method is often better than the other two algorithms. Because iteration will produce a lot of time cost, this kind of method can often get excellent scheduling results. Therefore, this subsection does not compare populationbased (genebased) methods.
This subsection will discuss the performance of HDNNS algorithm from the above two aspects. This part tests the performance of deep reinforcement learning (DEEPRL) [29, 42], deep Q learning (DQN) [43], artificial neural network (ANN) [3], Hopfield neural network (HNN) [44], stochastic processing time (SHPT) [45] method, and shortest processing time (SPT) [46] method.
The dataset is ZLP dataset [41], which contains 2000 8 8 JSSPs (, ) and 13 13 JSSPs (, ). The solution of JSSPs above is generated with the BBM method [47]. The processing time matrix of each JSSP satisfies the uniform distribution . For HDNNS, ANN and the first 1500 JSSPs are used in the training section. The last 500 JSSPs are used in scheduling section to test the performance of the model. The DEEPRM method is trained by interacting with the JSSP model. The size of the convolution kernel is 3 3, and and . The learning rate of HDNNS, ANN, and DEEPRL is 0.01. The number of learning epochs is 100. For HNN, SHPT, and SPT, the performance of the method is tested directly with the last 500 data. The experimental environment is Lenovo k4450, Ubuntu 16, CPU i4700 2.1 MHZ, Python, and Tensorflow.
The results of the experiment are shown in Table 8. Four indexes are discussed in the table: average MAKESPAN, scheduling score, scheduling time, and training time. The scheduling score is calculated according to
In (28), the scheduling capacity of current algorithm A in database D is defined as the ratio of the average optimal MAKESPAN to the current algorithm’s MAKESPAN .
Eight groups of JSSPs are tested in Table 8. The first four groups are 8 8 JSSPs, and the last four groups are 13 13 JSSPs. Each problem is generated by randombased function, and its processing time satisfies the uniform distribution .
For the learningbased method (HDNNS, DQN, DEEPRL, and HNN), the time consumption is divided into training time and scheduling time. The training time is the total time needed for 100 epochs of model training. The scheduling time recorded the total time of testing 500 JSSPs.
The ANN method is tested in two cases, one (ANN (1D) in proposed in [3]) using only onedimensional feature as the input feature and the other (ANN (ALL)) using flattened onedimensional features and twodimensional features as input features. ANN (1D) has a smaller network structure, so it has faster training efficiency and scheduling efficiency. Although the scheduling results of ANN (ALL) are better than that of ANN (1D), its training time is significantly improved with the JSSP scale. It is because the network structure of ANN has no advantage in dealing with complex scheduling information. Moreover, it cannot adequately deal with the relationship between the combination features and the output. In general, the scheduling effect of ANN network is better than the SPT method and STPT method.
Hopfield neural network (HNN) can also effectively obtain the scheduling results. But, unlike ANN, HNN seeks stability point through evolution and achieves the purpose of scheduling. HNN has a good effect on smallscale problems but suffers from the resolution of largescale problems.
DEEPRM and DQN are scheduling methods based on reinforcement learning (DEEPRM’s network structure is the same as that of HDNNS, and DQN use a standard deep network). These methods do not need labeled training data in the training section, but they need much interaction with the scheduling environment. In most cases, interaction learning is much slower than learning through training data. For JSSP which has easy access to label data, DEEPRM and DQN have disadvantages in training efficiency.
HDNNS is stable in different processing time distributions and different problem scales m and n. Moreover, the scheduling ability is maintained at 90% of the optimal solution, which is superior to the same ANN and HNN. Although the training time of HDNNS is longer than that of ANN (1D), it does not affect the realtime scheduling of the scheduler in applications because the training phase can be completed beforehand. Considering the scheduling performance of all the algorithms, HDNNS has significant advantages.
6.4. MAKESPAN and Time Consumption Comparisons in Traditional Dataset
This subsection uses the same methods as in Section 6.3 to solve the classical JSSPs, which include ft10 [48], ft20 [48], la24 [49], la36 [49], abz7 [50], and yn1 [51].
The experimental procedure is as follows. First, 2000 JSSPs of the same scale as the under test JSSP are generated. Then, the stateoftheart method is used to find the optimal solution (nearoptimal solution) as the training data. In this experiment, the solution of smaller JSSP (ft10, ft20, la24) is generated by the BBM method [47]. Moreover, the solution of larger JSSP (la36, abz7, yn1) is generated by the GA method. The first 1500 JSSPs are used in the training section. The last 500 JSSPs are used in scheduling section to test the performance of the mode.
The DEEPRM method is trained by interacting with the JSSP model. The learning rate of HDNNS, ANN, and DEEPRM is 0.01. The number of learning epochs is 100. The experimental environment is Lenovo k4450, Ubuntu 16, and CPU i4700 2.1 MHZ.
The test results are shown in Table 9.
The structure of Table 9 is the same as that of Table 8. The first column shows the optimal solution. Six popular JSSPs are tested in Table 9. The brackets below the JSSP name indicate the size of the problem.
Testing with separate test questions introduces randomness, so we recommend using the average of a large number of test results to measure the effectiveness of the algorithm (like ZLP datasets).
6.5. MAKESPAN Comparisons with Traditional Classification Algorithms
In this subsection, several traditional classification methods are used to compare with HDNNS. HDNNS is essentially a classificationbased method, so it is necessary to compare it with some traditional classification methods. We replace the deep neural network scheduler in Figure 1 with other classification methods and measure its effect.
In this experiment, we test knearest neighbor (KNN) [52], support vector machine (SVM) [26], decision tree (DT) [53], extremely randomized trees (ERT) [54], and Gaussian model (GOSS) [55].
The test dataset is ZLP dataset [41], which contains 2000 15 12 JSSPs (, ) and 2000 15 18 JSSPs (, ). The solution of JSSPs above is generated with the GA method. The first 1500 JSSPs are used as the training section. Then, the performance of the method is tested with the last 500 data. The parameters of the above classification methods are the default parameters of Python 3’s sklearn tool kit. The result of the solution is shown in Table 10.
The experimental results show that HDNNS has a significant advantage over traditional classification algorithms. Although the traditional method has an advantage in efficiency, it can only achieve the 80% of nearoptimal solution. Therefore, HDNNS has a big advantage in the framework of this paper.
6.6. Analysis of Generalization Performance
HDNNS has a good scalability, and a trained scheduler can be used to solve problems of different machine numbers m. In other words, models trained with less complex problems can be used to solve more complex problems. Based on this premise, it is necessary to measure the generalization of models at different levels of complexity.
This subsection discusses the performance of models trained with smallscale data in solving largescale problems. In the experiment, 1500 JSSPs (labels are generated with GA) are used as the training section. Then, groups of larger problems (larger machine number m) are applied to test the scheduling capability of HDNNS. In order to get a credible conclusion, the experiment generates 500 different JSSPs and corresponding nearoptimal solution for each group.
The box diagram of the experiment is shown in Figure 13.
In Figure 13, each box in the diagram represents a test result of a group. The top and bottom multiplication symbols represent the maximum and minimum values in the test. Moreover, the top and bottom triangles between the multiplication symbol is the 1% point and the 99% point of the 200 data. The lower and upper bounds of the boxes are 25% and 75% of the 200 data. The horizontal longer line in the middle of the box is the median number, and the horizontal shorter line is the average number.
Figure 13 shows that the closer the scale of test problems and training problems is, the better their performance wil l be. The average ratio of MAKESPAN obtained by HDNNS to GA is 0.97, and the MAKESPAN of the scheduling result is also stable.
With the increase in the number of machines, the model’s efficiency gradually decreases, which is embodied in the decline of the excellent degree of the solution and the stability of the solution. However, the decline in solving ability is not rapid and unacceptable.
We are happy to see that our scheduler can extract scheduling knowledge from a simple JSSP and use it successfully in a more complex scheduling problem. Specifically, the excellent degree of solutions of all test problems is greater than 0.86 (average).
7. Conclusion
A hybrid deep neural network scheduler with the characteristics of offline training and online realtime scheduling is created in this paper. In this scheduler, we present two innovations based on the machine learning framework. One is the convolution twodimensional transformation (CTDT), which converts the irregular data in the scheduling process into regular data; this enables deep convolutional operation to be used to solve JSSP. Another is hybrid deep neural network structure including convolution layer, fully connected layer, and flattening layer. And, this structure can effectively complete the extraction of scheduling knowledge.
The results show that the MAKESPAN index of HDNNS is 9% better than that of HNN and is 4% better than that of ANN in ZLP dataset. The training time of the HDNNS method is obviously faster than that of the DEEPRM method with the same neural network structure. Besides, the scheduler has brilliant generalization ability, which can solve largescale scheduling issues with smallscale training data.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (No. 61873240).