Abstract

This paper addresses a single-machine scheduling problem with sequence-dependent family setup times. In this problem the jobs are classified into families according to their similarity characteristics. Setup times are required on each occasion when the machine switches from processing jobs in one family to jobs in another family. The performance measure to be minimized is the total tardiness with respect to the given due dates of the jobs. The problem is classified as -hard in the ordinary sense. Since the computational complexity associated with the mathematical formulation of the problem makes it difficult for optimization solvers to deal with large-sized instances in reasonable solution time, efficient heuristic algorithms are needed to obtain near-optimal solutions. In this work we propose three heuristics based on the Iterated Local Search (ILS) metaheuristic. The first heuristic is a basic ILS, the second uses a dynamic perturbation size, and the third uses a Path Relinking (PR) technique as an intensification strategy. We carry out comprehensive computational and statistical experiments in order to analyze the performance of the proposed heuristics. The computational experiments show that the ILS heuristics outperform a genetic algorithm proposed in the literature. The ILS heuristic with dynamic perturbation size and PR intensification has a superior performance compared to other heuristics.

1. Introduction

Scheduling is a very important decision-making process that occurs in manufacturing systems. Scheduling problems deal with the allocation of resources to jobs over given time periods and its goal is to optimize one or more performance measures. This type of problems has been thoroughly studied since the mid-1950s [1]. Nowadays, scheduling problems are one of the most studied problems because they have great practical and theoretical importance. These problems have many applications in several industries (like chemical, metallurgic, and textile) and most of these problems belong to the class of -hard problems.

The scheduling problem focused on in this paper is stated as follows. There is a set of jobs to be processed on a single machine without interruption or preemption. Jobs are classified into families and are available at time zero. Each family has jobs (), such that . denotes the family of job . The processing time () and due date () of job are previously known. There is a family setup time between jobs and if job is processed immediately after job and . If jobs and belong to the same family (), . The setup times are sequence-dependent; that is, may not be equal to , , .

Setup time is the time required to prepare the necessary resource (machines) to perform a task (operation or job) [2]. Setup operations include obtaining tools, positioning work in process material, return tooling, cleanup, setting the required jigs and fixtures, adjusting tools, and inspecting material [1]. In the problem under study we do not consider the group technology (GT) assumption; that is, a family of jobs is not necessarily processed as a single batch [3]. Therefore a family of jobs could be divided into multiple nonconsecutive batches in an optimal sequence and each of the batches incurs a setup time.

The goal of the problem is to find a production schedule (sequence of jobs) that minimizes the total tardiness with respect to the given due dates of the jobs. The tardiness of a job (defined by ) is defined as the completion time of the job () minus the due date for the job if the job is completed after the due date, and the tardiness is equal to zero if the job is completed before the due date. can be expressed as . If job is processed immediately after job , then if , and if . The total tardiness of jobs (objective function) is computed as

The single-machine total tardiness (SMTT) problem with sequence-dependent family setup times is denoted by following the three-field notation presented by Graham et al. [4] and adapted by Allahverdi et al. [1], where is the single-machine environment, is information of sequence-dependent family or batch setup time, and is the total tardiness objective function.

The total tardiness criterion is very important in manufacture systems because several costs may exist when a job is delivered with tardiness. Among these the following can be quoted: contractual penalties, loss of credibility resulting in a high probability of losing a client for some or for all the possible futures jobs, and damage in the company’s reputation that may distance other clients.

Since the SMTT problem without sequence-dependent setup times is a binary -hard problem [5], it follows that the problem considered in this paper is at least -hard in the ordinary sense [6].

Scheduling problems that consider explicitly setup times are of great importance in manufacturing systems. Extensive literature reviews for these problems, considering different shop environments, were presented by [1, 2]. The SMTT problem with sequence-dependent setup times (without considering the grouping of jobs into families), denoted by , is one of the most researched topics in the scheduling literature. For this problem there are many studies on metaheuristics such as genetic algorithms [710], Simulated Annealing [11], Ant Colony Optimization [12, 13], Greedy Randomized Adaptive Search Procedure [14, 15], Iterated Local Search [16], and iterated greedy [17]. There are also some studies on exact algorithms for [1821].

Some studies were realized on single-machine scheduling problems with family consideration and sequence-independent family setup times (). To better understand , let us suppose that the jobs ,  , and belong to different families. If job (or ) is processed immediately before job , then (or ), where is the considered sequence-independent setup time. Note that the setup time does not depend on the family (or ) previously processed; it only depends on family . Kacem [22] addressed the problem and proposed a set of methods to obtain lower bounds for the optimal total tardiness. Schaller [23] developed Branch and Bound and heuristic algorithms for the problem. Exact methods were used by Pan and Su [24] and Baker and Magazine [25] to solve the problem, where is the maximum lateness. To minimize the total earliness and tardiness cost (), exact and heuristic methods were developed by Schaller and Gupta [26]. Gupta and Chantaravarapan [27] considered the problem where jobs in each family are processed together; that is, the group technology assumption is considered. These authors provided a mixed-integer linear programming (MILP) model capable of solving small-sized problems.

Single-machine scheduling problems involving sequence-dependent family setup times have been much less studied [2]. van der Veen et al. [28] addressed the problem, where is the maximum completion time (makespan). These authors proposed a polynomial time algorithm for this problem. For the problem , where is the total completion time, Karabati and Akkan [29] proposed a Branch and Bound algorithm. Jin et al. [30] addressed the problem and developed a Simulated Annealing heuristic. For the same problem, Jin et al. [31, 32] proposed dominance relations and Tabu Search heuristics. Sels and Vanhoucke [33] considered the problem where each job has a release time (the earliest time at which the job can star its processing). They developed a genetic algorithm with local search. Recently, Herr and Goel [34] addressed a SMTT problem with sequence-dependent family setup times and resource constraints; that is, each job requires a certain amount of resource that is supplied through upstream processes. Schedules must be generated in such a way that the total resource demand does not exceed the resource supply up to any point in time. This problem can be denoted as , where is the quantity of a resource required by the machine to process the job . The authors proposed a MILP model and a heuristic algorithm to solve the problem.

To the best of our knowledge, there is only a paper in the literature discussing the problem. Chantaravarapan et al. [6] propose a MILP model and a hybrid genetic algorithm (HGA) for this problem. The effectiveness of the HGA was evaluated by instances with up to 60 jobs. The experiments presented in [6] showed that HGA performs better than other heuristics.

Motivated by the computational complexity and the practical relevance of the problem, in this work we test the applicability of the metaheuristic Iterated Local Search (ILS) to find good quality solutions for realistic size problem instances. We use ILS because it is a simple and generally applicable heuristic that has proved to be very effective for solving a wide range of difficult combinatorial optimization problems, especially vehicle routing problems [3539] and scheduling problems [4045]. Furthermore, ILS has very few control parameters and it does not require specific knowledge of the problem as in sophisticated heuristic algorithms.

The traditional ILS consists of an iterative process that combines a perturbation phase and a local search phase [46, 47]. During the perturbation phase, the current solution is modified in a probabilistic fashion similar to the mutation operator used in genetic algorithms. In the local search phase, the perturbed solution is improved leading to a local minimum solution. An acceptance criterion is used to decide whether the search continues from the local minimum solution or from the one that served as the starting point of the most recent perturbation phase.

Besides the standard ILS heuristic, this paper presents other contributions: the ILS is enriched by two special features. The first feature consists in using a variable perturbation size to escape from local optimal solutions. This feature is inspired from the Variable Neighborhood Search (VNS) heuristic that systematically changes the neighborhood within the search [48]. The perturbation size at the beginning is fixed at and incremented by 1, if the solution is not improved until a value (the maximum perturbation size). If a solution improves in any perturbation size, it is again fixed at . The second feature consists in using a Path Relinking (PR) technique [49] to intensify the search of good solutions. PR generates new solutions by exploring paths that connect elite solutions. The performance of our ILS heuristic with the different features is carefully analyzed.

The remainder of the paper is organized as follows. In Section 2, we present the mathematic formulation for the problem under study. In Section 3, we describe all the phases of the proposed ILS heuristics. In Section 4, we show the design of instances and the calibration of our heuristics and report the computational results. Finally, in Section 5, we conclude this paper and give future directions.

2. Problem Model

In this section we present a mixed-integer linear programming (MILP) model of the problem. This mathematical model is originally from [6]. The resulting MILP is used to assess the performance of the developed heuristics for small-size problem instances. The goal of the model is to determine the optimal sequence of jobs to be processed on the single machine. A sequence is a permutation of jobs , where is the th job of the processing sequence (i.e., is the job processed in the position ).

The following parameters are used in the model (input data): = number of families, = number of jobs in family (), = total number of jobs (), = due date of the th job in family (, ), = processing time of the th job in family (, ), = sequence-dependent setup time of preceding family and following family (), = setup time of family before the first position ().

The following decision variables are used within the model: = completion time of job at position . = tardiness of job at position .

The resulting MILP model is as follows:

The objective function (the total tardiness) to be minimized is defined in (2). Constraint set (3) assures that one position of the sequence can contain only one job. Constraint set (4) guarantees that each job should be processed only once. Constraint sets (5) and (6) calculate the completion time of job in the first position of the sequence (). Constraint set (5) controls the setup time of the first position, resulting in the completion time of the first position in (6). Constraint sets (7) and (8) calculate the completion times from the second position to the last position of the sequence. For any two consecutive jobs, constraint set (7) checks whether or not the preceding job and the following job are from the same family. If so, there is no setup time between them (). Otherwise, setup time exists (). Constraint (9) computes the tardiness value () of jobs at each position . Constraints (10) represent nonnegativity conditions of the decision variables and , while constraints (11) and (12) ensure that and are binary variables.

3. Proposed ILS Heuristics

To solve the problem, in this section, we propose three heuristics based on the Iterated Local Search (ILS) metaheuristic. ILS is a simple and generally applicable metaheuristic that iteratively applies a perturbation procedure as a diversification mechanism and a local search (LS) as an improvement heuristic. At each iteration, a new initial solution is generated by randomly performing an appropriate modification, called perturbation, to a good locally optimal solution previously found (current solution). Instead of generating a new initial solution from scratch, the perturbation mechanism generates a promising initial solution by retaining part of the structure that made the current solution a good solution. The perturbed solution is improved by the LS heuristic obtaining a new solution . The solution is accepted as the new current solution under some conditions defined by the acceptance criteria. A detailed explanation of the ILS metaheuristic can be found in [47].

The first proposed heuristic, called ILS_BASIC, is a standard implementation of ILS. The pseudocode of ILS_BASIC is described in Algorithm 1. To implement the basic ILS algorithm, four procedures are specified: (i) CONSTRUCTION, where an initial solution is constructed; (ii) LOCAL_SEARCH, which improves the solution initially obtained and the perturbed solution; (iii) PERTURBATION, where a new starter point is generated through a perturbation of the current solution; (iv) ACCEPTANCE_CRITERION, which determines from which solution the search should continue. The best solution found over all iterations is returned by the ILS algorithm.

output: Best Solution
() begin
()      ≔ CONSTRUCTION();
()      ≔ LOCAL_SEARCH(, );
()     ; //the best solution
()     while  stop_condition  do
()       ≔ PERTURBATION(, );
()       ≔ LOCAL_SEARCH(, , );
()      if    then
()       ;
()   if    then
()    ;
()   else
()     ACCEPTANCE_CRITERION(, , );
()  return  

Our ILS_BASIC algorithm has four input parameters: , , , , and . Parameter defines the perturbation size. Parameter is used to reduce the size of the neighborhood in the LS procedure (this parameter defines the probability to generate a neighbor solution). defines the way of selecting the next neighbor solution in the local search process (two selection strategies are tested: first-improvement and best-improvement). Parameter is used in the ACCEPTANCE_CRITERION ( defines the probability to accept a worse solution). The iterations of the ILS algorithm are computed until a stopping condition (defined by the parameter ) is satisfied.

It is important to mention that the perturbation parameter used in the ILS_BASIC algorithm is static; that is, is fixed a priori before the beginning of the ILS search. The performance of the algorithm strongly depends on the intensity of the perturbation mechanisms. If it is small, not many new solutions will be explored, while if it is too large, it will adopt almost randomly starting solutions.

The second proposed heuristic, called ILS_DP, improves ILS_BASIC by using a dynamic and adaptive perturbation; that is, the value of (perturbation size) is defined dynamically during the search according to updating of the best solution found by the algorithm. ILS_DP algorithm, besides using the parameters , , and of ILS_BASIC, uses two new parameters: and . controls the increment frequency of , and defines the maximum value for . The steps of ILS_DP are summarized in Algorithm 2. In Step (), an initial solution is constructed and it is improved by local search (LS) procedure (Step ). In Step () the perturbation size is initialized (). The iterations of the algorithm ILS_DP are computed in Steps () to () until a stopping condition is satisfied. During each iteration, the current solution is perturbed (Step ) and improved by local search, obtaining a solution (Step ). In Steps () to (), if solution improves the best solution obtained so far, the perturbation size is set to its lowest value (). If the best solution is not improved during consecutive iterations, the perturbation size is incremented (Steps to ). The maximum value of is . In Steps () to (), if solution improves the current solution , it is accepted as the new current solution; otherwise solution is accepted if it meets the acceptance criterion.

output: Best solution
()    begin
()      CONSTRUCTION();
()      LOCAL_SEARCH(, , );
()     ;
()     ;
()     ; //the best solution
()     while  stop_condition  do
()       ≔ PERTURBATION(, );
()       ≔ LOCAL_SEARCH(, , );
()   if    then
()    ;
()    ;
()    ;
()   else
()    ;
()    if     then
()     ;
()   if    then
()    ;
()   else
()      ≔ ACCEPTANCE_CRITERION(, , );
()  return  

The third proposed heuristic, named ILS_DP+PR, is an extension of the second heuristic. ILS_DP+PR combines ILS and Path Relinking (PR) generating a hybrid heuristic. PR is an approach that generates new solutions by exploring trajectories that connect high-quality solutions [50]. PR uses a set of high-quality solutions (). We adapted PR in the context of ILS as a form of intensification, which consists in finding a path between the solution returned by the LS procedure and an elite solution. A general pseudocode description of ILS_DP+PR is presented in Algorithm 3. The algorithm has an additional parameter: (the maximum size of the elite set). The algorithm ILS_DP+PR differs from ILS_DP in the following steps. In Step (), the elite set EliteSet is initialized with the best solution obtained at the beginning of the algorithm. During each iteration, in Step (), the PATH_RELINKING intensification is applied between the solution (returned by local search) and the solution selected randomly from (Step ). The PR returns the best solution found (). In Step (), the elite set is updated with solution .

output: Best solution
()    begin
()      CONSTRUCTION();
()      ≔ LOCAL_SEARCH(, , );
()     ;
()     ;
()     ; //the best solution
()     ;
()     while  stop_condition  do
()       ≔ PERTURBATION(, );
()    ≔ LOCAL_SEARCH(, , );
()    ≔ Randomly select a solution from ;
()    ≔ PATH_RELINKING(, );
()    ≔ UPDATE_ELITE_SET(, );
()   if    then
()    ;
()    ;
()    ;
()   else
()    ;
()    if     then
()     ;
()   if    then
()    ;
()   else
()     ≔ ACCEPTANCE_CRITERION(, , );
()  return  

The next subsections provide a detailed explanation of the main components that are used in the proposed ILS heuristics.

3.1. Representation of a Solution

A solution (schedule) of the problem is represented by a permutation of jobs. For an instance with and , in Table 1 is presented an example of an input data. For each job , the family, processing time, and due date are presented in Table 1(a). The setup times between the families are shown in Table 1(b).

For the schedule , illustrated in Figure 1, the completion time and tardiness of each job are , , , , , , and , , , , , , , respectively. Therefore, the total tardiness for this schedule is . We can see that the schedule forms four batches and three setup times are required (, , and ). In this example, setup time is not considered at the beginning of the sequence.

3.2. Construction of the Initial Solution

The CONSTRUCTION procedure used by the ILS algorithms is based on the NEH heuristic [51]. In this work, to generate an initial solution, we adapt the NEH heuristic for single-machine scheduling and total tardiness minimization. In this heuristic first, the jobs are arranged in nondecreasing order of the due dates (Earliest Due Date dispatching rule) forming an ordered list of jobs . The first job of is selected to form a partial sequence (). For each , the next job of is inserted in all possible positions of generating partial sequences. For example, for , we obtain two partial sequences by inserting job in the two positions of : and . The best partial sequence (in relation to the partial total tardiness) is selected to replace . The heuristic finishes when a complete sequence is obtained (i.e., all the jobs of were inserted into ).

3.3. Local Search

The LOCAL_SEARCH procedure is used to improve a solution (which can be the initial solution or the perturbed solution) and is based on the insertion neighborhood. This procedure is shown in Algorithm 4. The iterations of the local search algorithm are determined in Steps () to (). In each iteration, neighbor solutions of the current sequence are constructed by considering the set of all jobs. Each job is removed from the current sequence (at random and without repetition (Steps and )) and then inserted into all possible positions of (Step ). The current solution is replaced by the best solution () among the possible ones, only if an improvement of can be obtained (Steps and ). If , the entire neighborhood is evaluated; that is, all jobs are selected to be inserted in all possible positions. In this case, the size of the neighborhood is . The parameter defines the job selection probability. To reduce the size of the neighborhood, we can use . We also test two strategies to select the solution to be explored in next iteration: first-improvement and best-improvement. If (Step ), the selection strategy is the first-improvement; otherwise, it is the best-improvement. Note that, with the first-improvement strategy (), a new iteration is started from the first improved solution . The iterations of the LOCAL_SEARCH procedure are repeated while there is improvement in the current solution (Steps ); that is, the procedure ends when the solution is a local optimum with respect to the insertion neighborhood.

output: Improved solution
()    begin
()     ;
()     while    do
()       false;
()       set of jobs determined at random;
()      for    to    do
()        ≔ randomly select a job from ;
()       ;
()       if    then
()      ≔ best solution obtained by inserting job in all position of ;
()     if    then
()      ;
()       ≔ true;
()      if    then
()       Go to step () / First Improvement /
()  return  
3.4. Perturbation

The goal of the perturbation method in an ILS heuristic is escape from a local optimal solution. In this work, we used a perturbation mechanism based on jobs exchanges. The perturbation size (parameter ) defines the number of exchanges to be made. To perturb a solution , a position () is first randomly chosen. Then, exchanges are applied between the jobs of positions and (). That is, the PERTURBATION procedure executes exchanges. For example, let us consider a schedule with jobs, . Let be a position randomly chosen. For , two pairs of jobs are exchanged: and . For , three pairs of jobs are exchanged: , , and . For and , the perturbed schedules are and , respectively.

In ILS_BASIC algorithm, parameter has a fixed value. In ILS_DP and ILS_DP+PR, the value of varies in the interval , where is a parameter to be tuned, such that .

3.5. Path Relinking Intensification

The Path Relinking (PR) technique was originally proposed by [50] as a mechanism to combine intensification and diversification by exploring trajectories connecting high-quality (elite) solutions previously produced during the search. PR needs a pair of solutions, say (initiating solution) and (guiding solution), . A path that links to is generated by applying neighborhood movements to the initiating solution, which progressively introduces attributes from the guiding solution [49].

In ILS_DP+PR algorithm, we use the PR variant called Mixed Path Relinking (MPR) [52]. Instead of starting from a solution and gradually transforming it into the solution , the MPR procedure performs one step from to , obtaining an intermediate solution . Then becomes the initiating solution and the guiding solution, obtaining a new intermediate solution . In the next step of the procedure, becomes the initiating solution and the guiding solution, obtaining an intermediate solution , and so on until the intermediate solution is equal to the guiding solution. The main advantage of this strategy is that it explores deeply neighborhoods of both input solutions. The MPR procedure returns the best solution obtained during the construction of the paths that connect the solutions and .

In this paper, the paths are generated applying exchange moves. For example, let us consider the solutions and ( jobs). Since these solutions have only one job in the same position (job 2 is in the first position in both solutions), the distance (or difference) between and is 6. Starting from , six neighbor solutions are generated by exchanging two jobs so that the distance with respect to is reduced by one: , , , , , and . From these solutions, the best one is chosen to be the intermediate solution (e.g., ). Considering the initiating solution and the guiding solution, five neighbor solutions are generated. From these five solutions the best one is chosen to be the new intermediate solution (e.g., ). Note that the difference between and is 4. The procedure ends when the difference between the solutions initiating and guiding is 1.

The initiating solution is the solution returned by the LOCAL_SEARCH procedure and the guiding solution is selected at random from the elite set (EliteSet). This set represents the pool of the best different solutions found by the algorithm. The maximum size of EliteSet is . The UPDATE_ELITE_SET procedure tests if a solution will be added or not to . When the elite set is full (), a solution is added to if and is better than the worst solution in . In this case, is added to in place of the most similar worst solution, that is, the solution with the smallest distance (or difference) from .

4. Computational Experiments

In this section, we describe the experiments carried out in order to study the behavior of the developed heuristic algorithms ILS_BASIC, ILS_DP, and ILS_DP+PR.

To the best of our knowledge, Chantaravarapan et al.’s study [6] is the only previous study to consider the problem addressed in this paper. Chantaravarapan et al. [6] proposed a hybrid genetic algorithm (HGA) which uses a local search heuristic in order to enhance performance. Once the best-fit chromosome in the offspring pool is determined, the local search heuristic is applied to that chromosome to improve the fitness before it is replaced into the population pool. The local search heuristic is based on two interchange methods: Adjacent Pairwise Interchange (API) and Randomized Pairwise Interchange (RdPI). The API swaps two consecutive jobs from the first position of the sequence to the last position. The RdPI randomly chooses any two jobs and switches the positions. If the solution is improved after switching, the initial sequence is updated and the process starts from the first position of the sequence. The local search continues until there is no improvement in solution.

We compare our ILS heuristics against HGA on 1440 randomly generated test instances. For small instances we compare the performance of the best proposed heuristic with respect to the MILP model solved by the commercial optimization solver IBM-ILOG CPLEX 12.5, which we will simply refer to as CPLEX. All of the heuristic algorithms and the MILP model have been coded in C++ and run on a PC with an Intel(R) Xeon(R) CPU X5650, 2.67 GHz with 48 GB of RAM running Ubuntu Linux 14.04.

In the following subsections, we first describe the random instance generation; then we present the parameter setting of our ILS algorithms; next we analyze the results obtained on large and small instances, and finally we analyze the convergence of the algorithms.

4.1. Random Test Instances

To test the performance of the proposed heuristic algorithms, random instances of the problem are generated according to Jin et al. [32]. The number of jobs is classified into two sets: (large-size instances) and (small-size instances). The number of families is set to be (large-size instances) and (small-size instances).

The processing times of jobs () are uniformly distributed in the interval . Three classes of setup times () are considered. Small (S), medium (M), and large (L) setup times are uniformly distributed in the ranges , , and , respectively. The due dates () of jobs are integer numbers uniformly distributed in the interval , where is a factor used to control the range of due dates, . Small and large due dates are generated with and , respectively.

The factors and the levels of the generated instances are shown in Table 2. Combining the factors , , , and ST, there are 144 categories of large instances and 108 categories of small instances. For each category of small and large instances, 10 and 2 instances are generated, respectively. Therefore, a total of 1440 large instances and 216 small instances were generated. All the generated instances are available for download at http://www.dpi.ufv.br/projetos/scheduling/smtt.htm.

4.2. Performance Measures

The most common performance measure used in the literature to compare the results of heuristic methods is the Relative Percentage Deviation (RPD) from the best known solution [53]. The RPD is computed according to the following equation:where is the solution (objective function value) obtained by a given method and is the best solution obtained among all the methods or the best known solution, possibly optimal. With this performance measure, good methods will have RPD close to 0.

In the problem, the minimum total tardiness (best solution) could be zero, so the RPD gives a division by zero. In this case, the Relative Deviation Index (RDI) measure is used to avoid division by zero. The RDI is computed as follows:where and are the best and the worst solutions obtained among all the compared methods, respectively. With this performance measure, a RDI between 0 and 100 is obtained for each method such that a good method will have a RDI close to 0. Note that if the worst and the best solutions take the same value, all the compared methods provide the best (same) solution and, hence, the RDI value will be 0 (best index value) for all the methods.

4.3. Calibration of Parameters

In this subsection we present the parameter setting of the developed heuristic algorithms ILS_BASIC, ILS_DP, and ILS_DP+PR. In order to calibrate the parameters of the heuristics, we generate a set of 100 random large instances (calibration instances) according to Section 4.1.

To make a fair comparison, the stopping condition () for all the heuristics is set to a maximum CPU elapsed time equal to seconds. Setting the time limit in this way allows more computation effort as the number of jobs increases [53]. In the calibration experiments, all the heuristics are run five independent times to obtain a final average of the results.

A Design of Experiments (DOE, [54]) is carried out to calibrate the proposed heuristics, where the factors are the parameters that need calibration and the response variable is the performance of the different algorithm configurations. Since the best known solution of each calibration instance is nonzero, the performance of each algorithm configuration is measured by RPD determined by (13). The experimental results were analyzed by means of a multifactor analysis of variance (ANOVA) technique [54]. In this statistical technique two hypotheses are tested, (i) null hypothesis: the medians (or performances) of all algorithms are equal; and (ii) alternative hypothesis: the median of at least one algorithm is different. To apply ANOVA, we checked its three main hypotheses, that is, normality, homoscedasticity, and independence of residuals. Statistical analysis showed that all of the three hypotheses could be accepted.

The first heuristic ILS_BASIC depends on four parameters: (perturbation size), (probability to generate a neighbor solution in the LS procedure), (which defines the selection strategy, best-improvement (), or first-improvement ()), and (which defines the probability to accept a worse solution). We carry out a full factorial experiment with these four parameters. A series of preliminary experiments were conducted in order to determine a suitable set of levels to test. The following levels were tested: , , , and . All combinations of values of the four parameters give 160 configurations of ILS_BASIC. Each ILS_BASIC configuration is run five different times on each calibration instance. The results of ANOVA technique indicate that there is statistically significant difference between the obtained results at a 95% confidence level. Since the of ANOVA test is less than , the null hypothesis should be rejected in favor of the alternative hypothesis. Figure 2 shows, for some configurations of ILS_BASIC, the means plot and Tukey’s Honestly Significant Difference (HSD) confidence intervals with 95% confidence level from the statistical test. For easy viewing, Figure 2 only shows the confidence intervals of configurations . Overlapping intervals indicate that no statistically significant difference exists among the overlapped means. We can clearly see that there are statistically significant differences among some configurations of ILS_BASIC. For example, the configurations 99, 100, and 104 are statistically different. Configuration 99 is statistically better than configuration 100 which in turn is statistically better than configuration 104. Also, we can see that configurations 99, 101, 102, and 103 are statistically equivalent. However, configuration 143 presents the smallest average RPD. This configuration of ILS_BASIC corresponds to , , , and . Thus, these parameter values will be used for ILS_BASIC in the next experiments.

The ILS_DP algorithm is calibrated in the same way as ILS_BASIC. For ILS_DP, five parameters should be calibrated: , , , , and . The first three parameters are the same parameters of ILS_BASIC. Parameter controls the increment frequency of the perturbation size () and is the maximum value of . For the five parameters of ILS_DP, the following levels were tested: , , , , and . The combination of these parameters (factors) yields a total of 256 different configurations of ILS_DP algorithm. Each configuration is run five different times on each calibration instance. The results of ANOVA test indicate that there is statistically significant difference between the obtained results at a 95% confidence level. The results of 16 configurations (including the best configuration) are shown in Figure 3. This figure shows the means plot and Tukey’s HSD confidence intervals with 95% confidence level from the statistical test. We can see that there are statistically significant differences among some configurations of ILS_DP. For example, the configurations 141, 142, and 143 are statistically better than the other configurations. These three configurations are statistically equivalent. However, configuration 142 presents the smallest average RPD. This configuration corresponds to following parameters: , , , , and . Since the value found for parameter is a threshold value, other levels were tested for this parameter: . The best results were obtained with . Thus, the parameters , , , , and will be used for ILS_DP in the next experiments.

Finally, we calibrate the ILS_DP+PR algorithm. Six parameters affect the performance of this algorithm: , , , , , and . Parameter is the maximum size of the elite set used in the Path Relinking intensification. Since the first-improvement strategy () gives the best results in the algorithms ILS_BASIC and ILS_DP, parameter is set to be 1 in ILS_DP+PR. For the other parameters of ILS_DP+PR, the following levels were tested: , , , , and . All combinations of values of these parameters give 256 configurations of ILS_DP+PR. For some configurations, Figure 4 shows the means plot and Tukey’s HSD confidence intervals with 95% confidence level from the ANOVA test. We can see that the configurations 121, 123, and 132 present the best results. However, configuration 123 (with , , , , and ) presents the smallest average RPD. Since the values found for parameters and are threshold values, other levels were tested for these parameters: and . New six configurations of ILS_DP+PR are generated by combining the values of these two parameters. The best configuration was obtained with and .

Based on the above results, the final parameter settings used in the proposed heuristic algorithms are shown in Table 3.

4.4. Comparison of the Heuristics on Large-Size Instances

In the first experiment, we present the comparison of solution quality generated by the proposed three ILS heuristics, ILS_BASIC, ILS_DP, and ILS_DP+PR, on 1440 large instances. We also compare our heuristics against the heuristic HGA proposed in [6].

Since the heuristics are stochastic methods, each instance is solved 30 times by each heuristic. All the heuristics are run with the same stopping criterion which is based on an amount of CPU time ( seconds).

The performances of all the heuristics are measured by the RDI from the best and worst known solutions obtained among all the methods. The RDI is computed for each instance according to (14), where and are the best and the worst solutions obtained among all the compared heuristics, respectively. Since each heuristic is run 30 times for each instance, for each heuristic three RDI values are presented: (i) the Best, (ii) the Worst, and (iii) the Average (Avg).

The obtained numerical results of the heuristics are reported in Table 4, where we have averaged the 120 instances of each group . For the heuristics HGA, ILS_BASIC, ILS_DP, and ILS_DP+PR, the Best, Worst, and Average (Avg) results are presented.

Considering the Best RDI values, the three ILS heuristics are better than HGA heuristic. For a total of 1440 large instances tested, the heuristics HGA, ILS_BASIC, ILS_DP, and ILS_DP+PR find the best known solution in 1182 (82.1%), 1343 (93.2%), 1351 (93.8%), and 1398 (97.1%) instances, respectively. These results show that all the heuristics, in 30 runs, are able to find the best known solution for most instances.

For all groups of instances, the heuristic ILS_DP+PR presents low Worst RDIs. For all groups of instances, except for groups and , the Worst RDIs values of ILS_DP are better than the respective values of ILS_BASIC.

Considering the Avg RDI values, the performance of the heuristic ILS_DP+PR is notoriously better than the other heuristics, except for groups of instances and . The heuristic ILS_DP outperforms, on average, ILS_BASIC for all groups of instances with and jobs. ILS_BASIC is better than ILS_DP for instances with jobs. The obtained average results indicate that the proposed ILS heuristics have a superior performance compared to HGA heuristic. Considering all groups of instances, the heuristics HGA, ILS_BASIC, ILS_DP, and ILS_DP+PR present overall Average (Avg) RDIs of 35.26%, 4.83%, 4.51%, and 3.34%, respectively.

Tables 5, 6, 7, and 8 show the Average RDIs of the heuristics, where the instances are grouped by the number of jobs (), number of families (), due date range (), and setup class (ST), respectively. In Tables 5 and 6 we can see that the Average RDI values of the heuristics increase as the number of jobs () and number of families () increase, respectively.

In Table 7 we can clearly see that the performance of the heuristics improves as the due date range increases. For all the heuristics, the greatest variation in the results occurs when the due date ranges vary. All the heuristics present short RDIs for instances with large due dates. Generally, for instances with large due dates, zero total tardiness is determined. The heuristic ILS_DP+PR presents the best results for all group of instances, except for instances with large due dates ( and ), where the ILS_BASIC presents the best results (Table 7). Instances with large due dates, generally, are easy to solve because almost all jobs are completed before their due dates. In these instances, ILS_DP+PR is inferior to ILS_BASIC probably because the dynamic perturbation used in ILS_DP+PR is relatively strong that easily allows leaving local minimum solutions.

In Table 8 we can observe that ILS_DP+PR presents the best results and all heuristics have large RDIs for instances with large (L) setup times.

In order to guarantee that the observed differences in the average results are indeed statistically significant, we carry out a parametric ANOVA statistical test using the RDI measure as response variable. We check the three main hypotheses of ANOVA test: normality, homoscedasticity, and independence of the residuals. No significant deviations were found in the fulfillment of the hypotheses. The results of this statistical test indicate that there is statistically significant difference between the obtained results at a 95% confidence level, with a . Since the ANOVA test does not specify which heuristics differ significantly, we carry out a Multiple Comparisons test to compare each pair of heuristics with a 95% confidence level. Table 9 shows the result of this test. Column Difference displays the sample mean of the first heuristic minus that of the second. Column Limits shows an uncertainty interval for the difference. Any pair for which the absolute value of the difference exceeds the limit (|Difference| > Limits) is statistically significant at the selected confidence level and is indicated by () in the column Significant. In Table 9, we can see that there is a significant difference between each pair of heuristics, and the heuristics ILS_BASIC and ILS_DP present the smallest difference (Difference − Limits = ).

The same analysis can be displayed in Figure 5. This figure shows, for the four heuristics and all the large instances, the means plot and Tukey’s HSD confidence intervals with 95% confidence level from the ANOVA test. We can see that ILS_DP+PR is statistically the best heuristic because its confidence interval does not overlap with the intervals of any of the other heuristics. The second best algorithm is ILS_DP followed by ILS_BASIC. The proposed ILS heuristics are statistically better than HGA heuristic.

In order to have a clearer picture, the heuristic HGA is removed for subsequent statistical analyses and we carry out another statistical test considering only the three proposed ILS heuristics. In Figure 6 we can clearly see that ILS_DP+PR is statistically better than ILS_DP which in turn is statistically better than ILS_BASIC. These results show that a basic ILS heuristic can be considerably improved by using a dynamic or variable perturbation size and Path Relinking intensification technique.

The obtained average results show that the effectiveness of the heuristic ILS_BASIC is considerably improved by using a dynamic perturbation and an intensification mechanism based on the Path Relinking technique. We believe that our ILS heuristics are better than HGA because we use a local search procedure based on the insertion neighborhood. HGA uses a local search heuristic based on the interchange neighborhoods API and RPI. In scheduling problems which have a permutation representation, insertion neighborhoods are efficient and strongly preferable over the interchange neighborhoods [55]. Furthermore, local search based metaheuristics are generally more effective than population based algorithms such as Ant Colony Optimization, Particle Swarm Optimization, and genetic algorithm [40, 56]. In order to facilitate follow-up research, the results obtained in this paper are available at http://www.dpi.ufv.br/projetos/scheduling/smtt.htm.

4.5. Experimental Results on Small-Size Instances

In this subsection the best ILS heuristic (i.e., ILS_DP+PR) is compared with CPLEX solver on the 216 small-size instances. This solver is applied to solve the MILP formulation (presented in Section 2) with a threshold CPU time of 1800 seconds for each small instance. That is, if after the established time no optimal solution is obtained, the best current solution (upper bound) is returned by CPLEX. Each small instance is solved 30 times by ILS_DP+PR heuristic using seconds as stopping criterion. In the comparison analysis, for each instance, we consider the best solution obtained by each heuristic. The CPLEX is only run one time for each each instance, because it uses a deterministic method. Since the used computer has 12 cores, the CPLEX is run using all the cores and the heuristic is run using a single core.

For a total of 216 small instances, 201 optimal solutions were generated by CPLEX, from which all these optimal solutions were found by ILS_DP+PR heuristic. On 11 instances with jobs, CPLEX was not able to generate a feasible solution, and on 5 instances ILS_DP+PR was better than CPLEX. Table 10 shows the results for these 5 instances. This table reports the factors of the instances (first column), the total tardiness () values obtained by ILS_DP+PR and CPLEX (second and third columns, resp.), and the relative percentage improvement (RPI) of ILS_DP+PR with respect to CPLEX (final column). The RPI is determined by . As can be seen, the solutions obtained by ILS_DP+PR are on average 12.64% better than the solutions (upper bounds) generated by CPLEX. The improvement of ILS_DP+PR varies from 6.44% to 20.87%.

The average CPU times spent by CPLEX and ILS_DP+PR are presented in Table 11. The CPU times of ILS_DP+PR are equal to seconds. We can see that CPLEX spends much more CPU time than ILS_DP+PR for all small instances. When the number of jobs is increased, the performance of CPLEX (considering solution quality and computational time) worsens significantly. We tested CPLEX solver on instances with jobs. For most of these instances CPLEX was not able to find feasible solutions in the established CPU time. Thus, it can be asserted that to solve large instances of the problem under study the most appropriate methods are heuristic algorithms, although they do not guarantee finding optimal solutions.

4.6. Time Analysis

The experiments presented in previous subsections were done considering solution quality of the heuristics. To give additional information concerning the computational effort required by the heuristic algorithms, we use the time-to-target (TTT) plot analysis [57]. A TTT plot is generated by executing an algorithm times and measuring the computational time required to reach a solution at least as good as a target solution. The running times are sorted in increasing order. The th running time is associated with a probability and the points , , are plotted. Each plotted point indicates the probability (vertical axis) for the algorithm to achieve the target solution in a given running time (horizontal axis). The more to the left is the plot, the better is the corresponding algorithm.

For this analysis we used an instance with jobs and families, where the heuristic algorithms HGA, ILS_BASIC, ILS_DP, and ILS_DP+PR found the best known solution (target solution) at least one time. We performed 100 runs for each algorithm, varying the random number generator seed. The algorithms were made to stop whenever a solution better than or equal to the target solution was found. Figure 7 displays the TTT plot for the four algorithms. It can be seen that the proposed ILS algorithms find solutions as good as the target solution clearly faster than HGA. All ILS algorithms have similar behaviors and a 2-second run time is enough to ensure a 98% probability of obtaining the target solution. The probability of HGA finding the target solution in 2 seconds is about 20%. Moreover, within 14 seconds, HGA reaches the target solution with a maximum probability of 98%. This analysis indicates that the proposed ILS algorithms converge to good solutions with short CPU times.

5. Conclusions

In this paper we have considered a single-machine scheduling problem with sequence-dependent family setup times, so as to minimize the total tardiness. The main contribution of this work was the development of three effective heuristics based on ILS metaheuristic to solve the problem. The ILS_DP extended the standard ILS_BASIC heuristic by using a dynamic perturbation size. Moreover, the ILS_DP+PR heuristic extended ILS_DP by using a Path Relinking intensification procedure that keeps a set of elite solutions. To the best of our knowledge, ILS heuristic has not been applied to solve single-machine scheduling problems with sequence-dependent family setup times. To improve the efficiency of the proposed heuristics, their parameters were fine-tuned using a Design of Experiments methodology. The heuristics were compared on the basis of computational experiments performed on a comprehensive set of large-size problem instances. The experiments demonstrated the effectiveness of our ILS heuristics. They outperformed a hybrid genetic algorithm (HGA) consistently. The heuristics ILS_DP and ILS_DP+PR performed better than the basic ILS (ILS_BASIC); that is, the results of the standard ILS were improved significantly by using a dynamic perturbation size and Path Relinking intensification. All the obtained results have been statistically validated.

We believe that the LS_DP+PR heuristic presented in this paper is a significant contribution, worthy of future study. Future research is to apply the proposed heuristic to other single-machine scheduling problems, for example, the problem with sequence-dependent family setup times and resource constraints, as the one studied very recently by Herr and Goel [34]. Other objectives could also be explored, such as the total completion times and total earliness and tardiness.

Competing Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

The authors are thankful to the financial support of CNPq and FAPEMIG, Brazilian research agencies.