Abstract

Workforce scheduling process consists of three major phases: workload prediction, shift generation, and staff rostering. Shift generation is the process of transforming the determined workload into shifts as accurately as possible. The Shift Minimization Personnel Task Scheduling Problem (SMPTSP) is a problem in which a set of tasks with fixed start and finish times must be allocated to a heterogeneous workforce. We show that the presented three-phase metaheuristic can successfully solve the most challenging SMPTSP benchmark instances. The metaheuristic was able to solve 44 of the 47 instances to optimality. The metaheuristic produced the best overall results compared to the previously published methods. The results were generated as a by-product when solving a more complicated General Task-based Shift Generation Problem. The metaheuristic generated comparable results to the methods using commercial MILP solvers as part of the solution process. The presented method is suitable for application in large real-world scenarios. Application areas include cleaning, home care, guarding, manufacturing, and delivery of goods.

1. Introduction

Workforce scheduling process consists of three major phases: workload prediction, shift generation, and staff rostering (see, e.g., [1]). Workload prediction is the phase of determining the workload based on known and predicted events. Shift generation is the process of transforming the determined workload into shifts as accurately as possible. The generated shifts form an input for the staff rostering phase, where employees are assigned to the shifts.

From the practical point of view, the workforce scheduling process relies on both optimization resources and human resources. It links the organization together, optimizing processes and streamlining decision-making. A shift scheduler should consider legal as well as human aspects of shift work. From an employee perspective, shift work is associated with pressure on social and family life, health issues, motivation, and loyalty. When the input data for the shift generation and for the staff rostering are well validated, significant benefits in financial efficiency and employee satisfaction can be achieved by applying relevant optimization methods. A good literature review of workforce scheduling can be found in [2].

The shift generation phase creates a shift structure including the tasks to be carried out in the shifts, the timing of the tasks and breaks, and the skills required in the shifts. The generation of shifts is based on either the varying number of required employees working during the planning horizon or the tasks that the shifts must cover. We call these employee-based and task-based shift generation problems.

The main contributions of this paper are the following:(i)The presented three-phase metaheuristic can successfully solve the most challenging SMPTSP benchmark instances(ii)The metaheuristic was able to solve 44 of the 47 instances to optimality(iii)The metaheuristic produced the best overall results compared to the previously published methods(iv)The metaheuristic generates comparable results to the methods using commercial MILP solvers as part of the solution process(v)The results can be reached when solving a more complicated General Task-based Shift Generation Problem

The paper is organized as follows: first, we describe the Shift Minimization Personnel Task Scheduling Problem and the General Task-based Shift Generation Problem. Then, we introduce the set of most challenging benchmark instances for SMPTSP. We describe the three-phase metaheuristic that we have used to solve a wide variety of combinatorial problems. Lastly, we present our computational results and compare them to the best-known results.

1.1. Related Work

The first major contribution for the employee-based shift generation problem was the study by Musliu et al. [3]. They introduced a problem in which the workforce requirements for a certain period were given, along with constraints about the possible start times and the length of shifts, and an upper limit for the average number of duties per week per employee. Di Gaspero et al. [4] proposed an employee-based problem in which the most important issue was to minimize the number of different kinds of shifts used. Application areas of employee-based shift generation include hospitals, retail stores, and call centers, where the number of employees required at certain timeslots can be predicted from the arrival times of customers.

In the task-based shift generation problem, the goal is to create shifts and assign tasks to these shifts so that the employees can be assigned to the shifts. Application areas include cleaning, home care, guarding, manufacturing, and delivery of goods. The first major contribution of the task-based problem was the study by Dowling et al. [5]. They created a master roster, a collection of working shifts, and off shifts, and then allocated a set of tasks to personnel with the requisite skills who are available for work on that day. Valls et al. [6] presented a model where they minimized the number of workers required to perform a machine load plan. Later, Krishnamoorthy and Ernst [7] introduced a similar group of problems, which they called Personnel Task Scheduling Problems (PTSPs). Given the staff that are rostered on a particular day, the PTSP is to allocate each individual task, with specified start and end times, to available staff who have skills to perform the task.

Subsequently, Krishnamoorthy et al. [8] introduced a special case referred to as Shift Minimization Personnel Task Scheduling Problem (SMPTSP) in which the only cost incurred is due to the number of personnel (shifts) that are used. Nurmi et al. [9] defined the General Task-based Shift Generation Problem (GTSGP) in which the task is to create anonymous shifts and assign tasks to these shifts so that employees can be assigned to the shifts.

2. Materials and Methods

2.1. Problem Description

The Shift Minimization Personnel Task Scheduling Problem (SMPTSP) consists of assigning a set of tasks with specific start and end times to employees who have specific skill sets and availability intervals. The objective is to find a feasible assignment of all the tasks that minimizes the number of employees used. The objective is motivated by situations where a large pool of casual employees is available and management would like to minimize the pool usage.

SMPTSP can be defined formally as follows. A set of tasks J = {t1, …, tn} needs to be allocated to a set of heterogeneous employees E = {e1, …, em} over a specified planning horizon. The processing time interval at which a task t has to be carried out is determined by a timetable with fixed start time st and finish time ft. Each employee e has a set of tasks Je ⊆ J that e can carry out. Each task t has a set of employees Et ⊆ E that can carry t. All sets Je and Et are defined based on skills of employees/skill requirements of tasks and availability of employees/time windows of tasks. The objective is to minimize the number of employees required to carry out the given set of tasks. The mathematical formulation of the problem was first given in [8].

Figure 1 shows a simplified instance of SMPTSP. The instance and the presented solution have the following characteristics:(i)The planning period is divided into 18 timeslots(ii)The number of tasks is fourteen and the number of employees is seven (indicated by letters from A to G)(iii)The duration of the tasks is given by the length of the corresponding rectangles(iv)The employees able to carry out a task are indicated in the rectangles(v)The colors indicate which tasks belong to the same shift(vi)The number of employees used to carry out the shifts is six(vii)An employee carrying out a task is denoted by parentheses (employee G has no tasks)

The following basic assumptions for SMPTSP hold:(A1) Preemption of tasks is not allowed(A2) There are no precedence constraints among the tasks(A3) Each task is processed exactly once without interruption(A4) Each employee can execute at most one task at a time

The General Task-based Shift Generation Problem (GTSGP) is to create anonymous shifts and assign tasks to these shifts so that employees can be assigned to the shifts. Instead of minimizing the number of employees required to carry out the given set of tasks, the objective is to maximize the number of feasible (shift, employee) pairs. A pair (s, e) is considered feasible if employee e can carry out shift s. The mathematical formulation of the problem was first given in [10].

The motivation for GTSGP derives from the workforce scheduling process. In the shift generation phase, we should create as versatile shifts as possible to ensure that the rostering of the staff can be completed. We should ensure that the resulting set of shifts can be carried out by the employees; that is, each shift can be assigned to an employee s.t. all shifts are assigned to someone and no employee is assigned to multiple shifts. In practical applications of GTSGP, the full-time permanent and temporary employees are expected to cover 100% of the total workload. This is opposite to the idea behind SMPTSP.

SMPTSP is a considerably simplified version of the GTSGP. Nonetheless, as Kroon et al. showed in [11], SMPTSP is NP-hard in the strong sense, even if preemption of tasks is allowed (A1). GTSGP has the same assumptions as SMPTSP besides (A2). GTSGP differs from SMPTSP in several important ways:(B1) Tasks are not explicitly assigned to employees(B2) Tasks are not fixed in time(B3) Tasks may have shift-local precedence constraints(B4) Transition times between tasks are considered(B5) Employees have total working time restrictions(B6) Employees have availability restrictions

To the best of our knowledge, eight significant and notable solution methods have been designed to tackle the problem. Krishnamoorthy et al. [8] used a Lagrangian approach to solve large instances of SMPSTP. They relaxed the task assignment constraints and used the deviations in the objective function with Lagrange coefficients, then solved each worker’s problem independently. Lin and Ying [12] developed a three-phase heuristic for SMPTSP. They obtained an initial solution using a simple but very effective construction heuristic, which is then improved using an iterated greedy heuristic, which in turn is used as an initial upper bound while solving the MIP model of the problem. The experimental results underlined the superiority of the proposed algorithm over the Krishnamoorthy et al. algorithm.

Smet et al. [13] used a 2-phase constructive metaheuristic approach. In phase one, they used three constructive heuristic methods, and in phase two, they used a hybrid search and optimization method. Their method was the first one to solve all the instances of the first SMPTSP benchmark set (see the next subsection) to optimality. The authors stated that their novel algorithm holds the state of the art for the SMPTSP. Fages and Lapegue [14] used constraint programming with both a top-down and a bottom-up approach. Their extensive experiments showed that their contribution significantly improved a straightforward SMPTSP model, so that it competed with the best-known approaches.

Baatar et al. [15] developed a branch and bound scheme, which was used in conjunction with two column generation-based approaches and a heuristic algorithm to create an efficient solution procedure. The authors demonstrated that their approach performs better than just using a standard commercial MILP solver (CPLEX). Hojati [16] proposed a novel greedy heuristic for SMPTSP. The results showed that the heuristic performs well relative to the current solution methods and has the added advantage of being able to solve very large instances fast. The author noted that his method requires no commercial MILP solver as the other methods described earlier. Niraj Ramesh et al. [17] solved the problem by decomposing it into separate subproblems and developed several exact and heuristic methods to solve the resulting subproblems. The authors stated that even though they proved interesting theoretical results, their methods fulfilled the expectations practically by easily surpassing other comparable exact methods.

Chirayil Chandrasekharan et al. [18] improved the metaheuristic introduced in [12] and presented a decomposition-based method where subproblems are solved to optimality using exact techniques. The optimal solutions of subproblems are subsequently utilized to construct a feasible solution for the entire problem. The new features not only improved solution quality, but they also played a pivotal role in improving the scalability of the algorithm. The authors state that the method is suitable for application in large real-world scenarios. Their method was the first one to solve all the instances of the second SMPTSP benchmark set (see the next subsection) to optimality. Furthermore, they were able to produce optimal solutions to almost all the instances from the third SMPTSP benchmark set (see the next subsection).

Currently, only one solution method has been published for the General Task-based Shift Generation Problem. The problem was only recently introduced in Nurmi et al. [9]. The authors used the PEAST metaheuristic to solve the GTSGP test instances introduced in [10]. We will give insight into the metaheuristic in The Three-Phase Metaheuristic section.

2.2. SMPTSP Benchmark Instances

To the best of our knowledge, no real-world benchmark instances for SMPTSP have been published. Thus far, three sets of artificial benchmark instances have been published. Krishnamoorthy et al. [8] presented the first data set of 137 instances for SMPTSP. The data set is referred to as KEB instances. The instances were generated by five characteristics: number of employees, number of tasks, lengths of the tasks, tightness level, and multiskilling level. The tightness level is defined as the total length of all tasks as a percentage of the total availability of all employees. When this level is 100%, the tasks could cover exactly the timeslots available for the employees. The multiskilling level is defined as the average percentage of the total number of tasks each employee is qualified for. When this level is 100%, each employee can carry out all tasks. From now on, we will call the multiskilling level as the task skill level. The number of employees in the KEB data set ranges from 22 to 422, while the number of tasks ranges from 40 to 2105. The shift lengths are fixed to 1440. The lengths of the tasks vary between 50 and 400. The tightness level is fixed to about 90%, and the skill level is either 33% or 66%.

Smet et al. [13] generated the second data set of ten instances, because they were able to solve all the KEB instances to optimality. The data set is referred to as SWMB instances. The number of employees ranges from 44 to 153, while the number of tasks ranges from 258 to 1577. The shift lengths are fixed to 1440. The length of the tasks is either 120 or 280. The tightness level is fixed to about 90%, and the skill level is either 20% or 30%. Fages and Lapègue [14] generated the third data set of 100 instances, since the KEB and SWMB instances are trivial with respect to finding good-quality lower bounds. This data set is referred to as FL instances. The number of employees ranges from 62 to 948, while the number of tasks ranges from 71 to 1583. The shift lengths are fixed to 1560. The length of the task is either 120 or 280. The tightness level is fixed to about 90%, and the skill level is fixed to about 25%. A good summary of the three data sets can be found in [16].

In the interest of brevity, we will not solve all the 247 instances from the three data sets. Some of the instances are easy to solve, and most of the eight methods described earlier in this section perform well on most of the instances. We next determine 47 of the most challenging instances of the three data sets as follows:KEB: the VAWA heuristic of Krishnamoorthy et al. [8] was not able to find the optimum solution, and the solution obtained by the constructive heuristic of Lin and Ying [12] was at least 5% inferior to the optimum solution totaling 14 of the 137 instances.SWMB: all the 10 instances.FL: the greedy heuristic of Hojati [16] was not able to find the optimum solution, or the solution obtained by the constraint programming approach of Fages and Lapègue [14] was at least 3% inferior to the optimum solution totaling 23 of the 100 instances.

Tables 13 show the characteristics of the selected instances. In addition to the tightness and skill levels, we define five other hardness indicators. The @AVG measure indicates the estimated average number of tasks per nonempty shift. The number of tasks is divided by a lower bound for the minimum number of shifts. For the lower bounds, we use the values published in [16] using the general lower bounding procedure by Solyali [19]. In addition to the task skill level, we define the shift skill level which describes how qualified an average employee is to carry out all the tasks of an average shift, i.e., tsln/lb, where tsl = task skill level, n = number of tasks, and lb = Solyali lower bound. The overlap level is the probability of two tasks to overlap. To calculate the probability, we need to iterate all task pairs once.

The %ICH measure indicates the percentage of the tasks feasibly assigned by the iterated constructive heuristic described in The Three-Phase Metaheuristic section. The heuristic maximizes the number of feasible (shift and employee) pairs, i.e., solving the GTSGP problem, as described earlier. The @PAIRS measure indicates an estimation of the average number of feasible pairs per nonempty shift. The number of feasible pairs generated by the iterated constructive heuristic is divided by the Solyali lower bound. Note that the GTSGP solution generated by the heuristic is quite far from the optimum. The values shown for %ICH and @PAIRS are the best of ten runs.

Recall that we have selected to solve the most challenging instances. The following observations can be drawn from the characteristics of the instances:(i) FL instances have fewer tasks per shift than in most of the other instances. Partly due to this, also the shift skill levels of FL instances are higher. These should make them easier to solve.(ii)The tightness levels of FL instances are lower than in the other instances. This should make them easier to solve.(iii)About half of KEB and SWMB instances have high overlap levels, which should make them easier to solve.(iv)The percentage of the tasks feasibly assigned by the iterated constructive heuristic (%ICH) is lower for SWMB instances, which could make them harder to solve. KEB and FL instances seem to be equally easy (or hard) to solve in this sense.(v)FL instances are very similar to each other. This should make them equally hard (or easy) to solve.

According to the @PAIRS value, barely one employee is skilled to carry out each shift in KEB and SWMB instances. For FL instances, the generated solution will be such that quite many employees may carry out several shifts. This could make FL instances easier to solve.

2.3. The Three-Phase Metaheuristic

We have created the PEAST metaheuristic (see, e.g., [20]) to solve real-world scheduling problems. The metaheuristic has been in commercial use for several years, for example, in staff rostering and in professional sports league scheduling. Furthermore, we have used it to solve more academic problems, such as balanced incomplete block design, single round robin tournaments with balanced home-away assignments and preassignments, days-off scheduling, and constraint minimum break problems. However, it should be clear that even though a metaheuristic can be powerful for several problem types, it is not guaranteed to work well for other problem types. The No Free Lunch Theorem [21] implies that there cannot exist a superior optimization method.

Recently, we have started to work with a new real-world problem, which is an application of GTSGP. As part of this process, we have created a solution method suitable for commercial use. The method is based on the PEAST metaheuristic. In all the previous adoptions of PEAST, we have used random initial solutions. We have found no evidence that a sophisticated initial solution improves results. On the contrary, random initial solutions seem to yield superior or at least as good results. However, due to the running time requirements, PEAST alone is too slow for the largest practical instances of GTSGP. We need a fast heuristic to generate very good initial solutions. Note that not reaching the best possible (academic) solution is not an issue in practical applications of GTSGP. After the optimization process, new tasks may arise and some of the tasks may need to be changed or removed.

We generate an initial solution using a simple ruin and recreate heuristic (RRH) similar to that described in [22]. The pseudocode is given in Figure 2. The ruin operator removes strings of adjacent tasks of random length from the solution. All tasks assigned to the incumbent solution are processed in random order. For each task t, a random string that contains t is removed unless the shift containing t has already been removed. When the total number of removed tasks exceeds the given parameter, the ruin operator quits. The recreate operator adds free tasks one by one to their respective best positions in the incumbent solution. First, all free tasks are sorted in random order. For each task t, all feasible addition positions in the incumbent solution are evaluated. Note that the concept of a position depends on the exact problem.

For GTSGP, a position is determined by an immediate predecessor, e.g., a task or an employee. Note that the tasks can have wide time windows in GTSGP; hence, the order of tasks within a shift is not predetermined. In SMPTSP, there are no time windows, which fixes the order of tasks within a shift. Task t is then added to the position that leads to the best objective function value, with a small chance to skip over to the next best position. Consecutive skipping is not constrained in any way, so t might not get assigned even if it has feasible addition positions. When all free tasks have been processed, the recreate operator quits.

We further speed up the overall running time by generating an initial solution to the ruin and recreate heuristic. Initial solutions are generated by using a very fast iterated constructive heuristic (ICH) based on the ideas presented in [12]. The pseudocode is given in Figures 3(a) and 3(b). This is the heuristic referred in The Three-Phase Metaheuristic section when calculating the %ICH measure. The heuristic assigns the unassigned tasks one by one to an assignable shift until all tasks have been considered (STEP 3). When trying to assign a task, we choose the shift with no employee and task conflicts and with the largest total cumulative processing time. A reassignment procedure (STEP 5) is repeatedly applied to allocate each unassigned task of to a shift until all tasks have been assigned or cycling exists. When trying to assign task t, we choose shift s with no employee conflicts and with the least total cumulative processing time with the tasks that are currently in conflict with task t. To reassign task t to shift s, we remove all tasks in shift s that conflict with task t.

The PEAST metaheuristic tries to improve the solution generated by the ruin and recreate heuristic (and the iterated constructive heuristic). In practical applications of GTSGP, the PEAST metaheuristic(i)generates as versatile shifts as possible to(ii)ensure that the rostering of the staff can be completed, so that(iii)the computation time is still acceptable considering the release time of the rosters.

Therefore, we do not seek the fastest possible solution method. For our purposes, it is advantageous to use more computation time in order to achieve shifts that are more versatile.

The pseudocode of the PEAST metaheuristic is given in Figure 4. We have created PEASTP (see, e.g., [20]) by combining features from six well-known metaheuristics: MH1, genetic algorithm; MH2, ejection chain method; MH3, tabu search; MH4, simulated annealing; MH5, variable neighbourhood search; MH6, ruin and recreate method.

The performance of these metaheuristics has been scientifically justified, and a significant number of experimental studies have been run at the algorithmic level. They have indisputably introduced true novelties to the repertoire of optimization methods. By combining and carefully tuning the most efficient operators of the scientifically valid metaheuristics, an experienced heuristic designer can solve real-life optimization problems efficiently. As stated at the beginning of the section, there is evidence that PEAST can successfully solve different problem domains. To the best of our knowledge, no other method combines metaheuristic features such as in PEAST.

PEASTP uses a population of solutions in each iteration (MH1). The reproduction operation is, to a certain extent, based on steady-state reproduction: the new solution replaces the old one if it has a better or equal objective function value. Furthermore, the worst solution is replaced with the best one at given intervals, i.e., elitism is used. The ejection chain search is the heart of PEASTP. It explores promising areas in the search space. The ejection chain search (MH2) extends a basic hill-climbing step to generate a sequence of moves in one step, leading from one solution candidate to another. The ejection operator actually implements several simple local search operators to work on a single solution (MH1).

The ejection chain search is improved by introducing a tabu list, which prevents reverse order moves in the same sequence of moves (MH3). A simulated annealing refinement is used to decide whether to commit to a sequence of moves in the ejection chain search (MH4). This refinement is different from the standard simulated annealing. It is used in a three-fold manner. The simulated annealing refinement and tabu search are used to avoid staying stuck in promising search areas too long. Shuffling operators assist in escaping from local optima. They are used to perturb a solution into a potentially worse solution in order to escape from local optima (MH5). A shuffling followed by several ejection chain searches obtains better solutions using the same idea as the ruin and recreate method (MH6).

PEAST uses a population of solutions in each iteration. A new solution instantly replaces an old one if it has a better or equal objective function value. Furthermore, the worst solution is replaced with the best one at given intervals, i.e., elitism is used. PEAST uses a traditional penalty method, which assigns positive weights (penalties) to the hard and soft constraints and sums the violation scores to get a single value to be optimized. The soft constraints are assigned fixed weights according to their significance. However, the hard constraints are assigned dynamic weights using a unique ADAGEN method described in [20].

Each of the five metaheuristic components in PEAST is crucial to produce good-quality solutions. This was verified for three different problem domains in [20]. Recently, the results in solving GTSGP instances [10] showed that when any one of the components was removed, the results were clearly worse. The same held even if PEAST was given twice as much time to run without one of the components.

The implementation of PEAST has changed so distinctly that we call the new version PEASTP. The data structures have been recoded, and the calculation of the cost function in ejection chain search has been renewed. The modifications have enabled us to parallelize PEAST (hence PEASTP), which in turn enables us to solve problem instances far larger than before. Nonetheless, we cannot use random initial solutions when solving practical GTSGP instances.

3. Results and Discussion

This section presents our computational results for SMPTSP benchmark instances introduced in the previous section. The results are juxtaposed against the results of all other recent SMPTSP solution methods known to us. Recall that we solve the instances as GTSGP instances and that SMPTSP is a considerably simplified version of GTSGP. As a result, we are not only optimizing the minimum number of shifts, but also at the same time generating as versatile shifts as possible. That is to say, we generate the SMPTSP results as a by-product when solving a more complicated General Task-based Shift Generation Problem. Due to the goal of this paper, we only present our SMPTSP results here, and not the actual GTSGP results. We used the same version of PEAST that was used when solving the GTSGP instances in [10]. We used no domain-specific knowledge to generate better solutions, nor did we do any parameter fine-tuning. In fact, we used the same version that is in commercial use for staff rostering and sports scheduling. The parameter values have been verified in several earlier implementations and applications of PEAST, see, e.g., the very detailed experiments in [20].

Table 4 shows the summary of the published results of the eight solution methods described in the previous section. Three of the solution methods are purely heuristic methods: K10 [8], LY14 [12], and H18 [16]. Our results are shown in column NK20. The five other methods use commercial MILP solvers as part of the solution process: S14 [19], FL13 [14], B15 [15], R18 [17], and C20 [18]. Table 5 shows the detailed results.

Chirayil Chandrasekharan et al. [18] noted that FL instances were unavailable when S14 method was published and that a summary of the S14 results was obtained upon request. The authors provided us the instance-specific results of the S14 method for FL instances.

As was explained earlier, we run our metaheuristic using as much computation time as practicably possible. For practical GTSGP applications, this could be several hours depending on the length of the planning period and on the number of processors and cores available for the computation. We ran our metaheuristic for eight hours for each of the benchmark instances. The calculation was not interrupted when the optimum SMPTSP value was reached since we were solving the GTSGP problem. To achieve the best possible GTSGP solution, we needed to use the entire calculation time for each instance. The test runs were carried out on a standard laptop, Intel Core i7-8550 at 1.8 GHz with 8 GB RAM running Windows 10.

The results show that our three-phase metaheuristic can successfully solve the most challenging SMPTSP benchmark instances. Another observation is that our metaheuristic generates comparable results to the methods using commercial MILP solvers as part of the solution process. We can also state that our metaheuristic produced the best overall results compared to the other methods. The metaheuristic was able to solve 44 of the 47 instances to optimality. For two instances, our solutions might be optimal. Only one instance was such that our solution was inferior.

Recall that we have selected to solve the most challenging instances. It is obvious that the number of tasks and the number of employees have a direct influence on the hardness of the instance, since they enlarge the search space. To be more exact, the combinatorial search space explodes when the average number of tasks for an employee increases. Accordingly, Lin and Ying [12] stated that instances with longer task lengths should be relatively easy to solve. Smet et al. [13] noted that shorter tasks and lower task skill levels make an instance harder to solve.

SWMB instances were generated based on these observations. Krishnamoorthy and Ernst [7] stated that the tightness of an instance should be close to 90% to make it challenging. FL instances were generated so that the maximal number of overlapping tasks does not provide a good lower bound for an instance [14]. Furthermore, the instances have an average more available employees per task than KEB instances, but they use a significantly smaller percentage of these employees in optimal solutions. This should make an instance more challenging. The SWMB instances were the only instance set where our method did not yield the best results out of all compared methods.

The above assumptions are in line with our reasoning given in the previous section. Based on our test runs, we can quite safely state that a low shift skill level, a low %ICH value, and a low @PAIRS value indicate a hard instance. Thus, the bold values in Tables 13 indicate a hard instance. Note that the bold values have been selected with respect to the other instances in the same data set.

In general, we argue that FL instances should be quite easy to solve, because quite many employees may carry out several shifts. For our metaheuristic, FL instances are quite easy to solve, excluding the instances FL#5 and FL#89. Our solution for these instances is one above the lower bound value. These solutions were reached very easily. Therefore, we speculate that these two instances might be such that the lower bound value is not the optimum value. Note that the instances were also unsolved by the Chirayil Chandrasekharan et al. method in [18].

The KEB instances are easy to solve. The instances have the highest task skill levels. Furthermore, as is the case for FL instances, KEB instances have high percentage of the tasks feasibly assigned by the iterated constructive heuristic.

We note that SWMB#7 instance is extremely hard for our metaheuristic. We have no conclusions for this at this time. Table 2 shows that SWMB#2 instance should be at least as hard to solve. However, it is not that hard for our metaheuristic. For SWMB#7 instance, in addition to the standard setup for our metaheuristic, we tried many different setups for iterated construction heuristic, ruin and recreate heuristic, and PEASTP. We even increased the running time. Nonetheless, our run-time measures showed that we are not even close enough to be able to solve the instance. As a final experiment, we tried to use Gurobi to improve our best solution. Unfortunately, we could not find a better solution or prove the solution optimal with seven days of computation time. Table 5 shows that as many as two solution methods have successfully solved the instance. Furthermore, the published results showed that the solutions were generated fairly fast.

Finally, we registered the running time elapsed to reach the first solution to SMPTSP while we were continuing to reach the best possible GTSGP solution within the given computation time. For the KEB instances, the median time to reach the SMPTSP solution was 0.9 minutes. The minimum time was 0.001 minutes, and the maximum time was 9 minutes. For the SWMB instances, the corresponding times were 86, 15, and 207 minutes, and for the FL instances 0.7, 0.1, and 19 minutes. This verifies our earlier discussions that SWMB instances should be the most challenging ones.

4. Conclusions

We presented a three-phase metaheuristic to Shift Minimization Personnel Task Scheduling Problem (SMPTSP). The metaheuristic actually solves a more complicated General Task-based Shift Generation Problem (GTSGP). The results for SMPTSP were generated as a by-product when solving GTSGP.

In the first phase of the method, we generated a population of initial solutions by using a very fast iterated constructive heuristic. In the second phase, a ruin and recreate heuristic was used to improve the solutions. Finally, the parallelized PEAST metaheuristic used the population of solutions to generate a final solution. In all the previous adoptions of PEAST, we have used random initial solutions. However, PEAST alone is too slow for the largest practical instances of GTSGP. The size of these instances is equal to the size of the largest SMPTSP instances solved in this paper.

The computational complexity of GTSGP and SMPTSP instances depends mainly on the number of tasks, the number of employees, and especially the average number of tasks per shift. These values set the limit for practical use of the PEAST method. Recall that we have only a couple of hours to generate solutions. Our test runs have shown that we cannot use PEAST when we have more than thousand tasks, especially when we have several hundred of employees or the average number of tasks per shift approaches ten. In these cases, we should only use ICH and RRH heuristics. However, the solutions could still be acceptable for practical applications, since after the shift generation and staff rostering have completed, new tasks will most certainly arise and some of the tasks need to be changed or removed.

We solved each instance using eight hours computation time on a standard laptop. Our test runs have shown that for practical instances on a high-computing environment, this corresponds to two hours. This requires using high-performance computers with high number of processors and cores and with very fast memory. Two hours is acceptable considering the entire workforce scheduling process and the release time of the final rosters. However, the running time is significantly higher than the running times of comparable methods.

We showed that the presented three-phase metaheuristic can successfully solve the most challenging SMPTSP benchmark instances. The metaheuristic produced the best overall results compared to the previously published methods. Furthermore, the metaheuristic generated comparable results to the methods using commercial MILP solvers as part of the solution process. The comparison is not entirely fair, as our method takes more time but also solves a more complex problem.

The metaheuristic was able to solve 44 of the 47 instances to optimality. For two instances, our solutions might be optimal. Only one instance was such that our solution was inferior. We are still working on finding out whether the two remaining FL instances can be solved to the lower bound. We also continue to examine why one of the SWMB instances is extremely difficult for our metaheuristic. Determining the cause should help us improve the generality of our solution method. We might encounter similar results on some of the easy instances, e.g., the ones not included in our experiments.

In the near future, we will publish a fourth data set for SMPTSP. We will also present an extension of SMPTSP.

Data Availability

The data for the SMPTSP instances are available online. The three data sets can be found in [23] (T. Lapègue, “Personnel Task Scheduling Problem Library” (online)), available at https://sites.google.com/site/ptsplib/smptsp/instances (last access 15 January 2021).

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.