`Mathematical Problems in EngineeringVolume 2012, Article ID 504713, 23 pageshttp://dx.doi.org/10.1155/2012/504713`
Research Article

## An Exact Algorithm for Bilevel 0-1 Knapsack Problems

1Centro de Investigação Algoritmi da Universidade do Minho, Escola de Engenharia, Universidade do Minho, 4710-057 Braga, Portugal
2Departamento de Produção e Sistemas, Universidade do Minho, 4710-057 Braga, Portugal
3LAMIH-SIADE, UMR 8530, Université de Valenciennes et du Hainaut-Cambrésis, Le Mont Houy, 59313 Valenciennes Cedex 9, France

Received 5 August 2011; Revised 26 October 2011; Accepted 31 October 2011

Academic Editor: Piermarco Cannarsa

Copyright © 2012 Raid Mansi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

We propose a new exact method for solving bilevel 0-1 knapsack problems. A bilevel problem models a hierarchical decision process that involves two decision makers called the leader and the follower. In these processes, the leader takes his decision by considering explicitly the reaction of the follower. From an optimization standpoint, these are problems in which a subset of the variables must be the optimal solution of another (parametric) optimization problem. These problems have various applications in the field of transportation and revenue management, for example. Our approach relies on different components. We describe a polynomial time procedure to solve the linear relaxation of the bilevel 0-1 knapsack problem. Using the information provided by the solutions generated by this procedure, we compute a feasible solution (and hence a lower bound) for the problem. This bound is used together with an upper bound to reduce the size of the original problem. The optimal integer solution of the original problem is computed using dynamic programming. We report on computational experiments which are compared with the results achieved with other state-of-the-art approaches. The results attest the performance of our approach.

#### 1. Introduction

Bilevel optimization problems were introduced for the first time in [1] in connection with the well-known Stackelberg game [2]. These problems are related to the decision making process conducted by two agents, each with his own individual objective, under a given hierarchical structure. The agent that is at the top of this hierarchy is called the leader. His distinctive feature in the process is that he knows which decision is taken by the other agent (called the follower). As a consequence, he can optimize his own objective by taking into account the decision of the follower.

The bilevel 0-1 knapsack problem (BKP) that is addressed in this paper is defined in this context. It is a hierarchical optimization problem in which the set of feasible solutions depends on the set of optimal solutions of a parametric 0-1 knapsack problem. The BKP can be formulated as follows: The decision variables related to the leader are denoted by . The follower has decision variables which are denoted by . The objective functions of the leader and the follower are denoted, respectively, by and . The weights of the variables and in the objective function of the leader are denoted by and , respectively, while the vector represents the coefficients of the follower variables in his own objective function. The vectors and are the set of coefficients related to the decision variables of the leader and the follower in the knapsack constraint of the follower, respectively. The capacity of this knapsack constraint is denoted by . All the coefficients of the problem are assumed to be positive. The standard 0-1 knapsack problem is a special case of BKP. It is obtained from (1.1) by setting and . As a consequence, the problem BKP is NP-hard.

Different methods have been proposed in the literature for bilevel programming problems with and without integer variables [3, 4]. A recent survey on the contributions for solving bilevel programming problems can be found in [5]. Many of these methods focus on problems with continuous variables.

The bilevel knapsack problem was addressed by Dempe and Richter [6], Plyasunov [7], and Brotcorne et al. [8], but only for the case where there is a single (continuous) variable for the leader and different binary variables for the follower. Note that, for these cases, there may be no optimal solution for the problem [4]. The branch-and-bound algorithm proposed by Moore and Bard in [9] and the method proposed by Brotcorne et al. in [10] can be adapted to solve the BKP addressed in this paper. At the end of the paper, we will compare our approach with the results obtained by these two algorithms.

Here, we consider the case where all the variables of the problem are binary variables. We propose a new exact approach for this problem based on several intermediate procedures. A lower bound for the problem is computed by applying first a polynomial time algorithm to solve the linear relaxation of BKP. The solutions generated by this algorithm are then used to compute feasible solutions to BKP and hence to obtain a valid lower bound for the problem. Using this lower bound and a given upper bound, the size of the problem is reduced by applying fixing rules. Dynamic programming rules are applied afterwards to obtain the optimal solution of BKP.

Bilevel programming problems have many applications on different fields including economics, engineering, the determination of pricing policies, production planning, transportation, and ecology. In [11], Dempe identified more than 80 references in the literature describing applications of bilevel problems. Other examples, namely, on the field of engineering are described in [12]. The BKP is a discrete bilevel problem that can be applied in many situations involving the interaction between two agents whose (binary) decisions are interrelated and with each one trying to optimize his own objective. Real applications of this problem can be found in revenue management, telecommunications, capacity allocation, and transportation, for example.

An application in revenue management may involve an individual searching for the best investment plan for his capital. The investor has the choice between placing his funds directly on financial applications with a guaranteed rate of return, letting an intermediary company (a broker, e.g.) decide how to invest these funds, or dividing the funds between these two possibilities. The intermediary company cannot invest more than the amount provided by the individual and it will do so in order to maximize its own profit. For this purpose, the intermediary will buy shares, bonds, or other financial assets that will provide it a revenue. Part of this revenue will be given back to the individual as a return on investment. In turn, the individual will decide on the amount to invest by itself and the amount to give to the intermediary with the objective of maximizing his own profit. In BKP, the individual will be the leader, while the intermediary will be considered as the follower. The value in (1.1) represents the capital of the individual. The coefficients of represent the amounts that the individual can invest by itself and which will provide him a guaranteed rate of return given by the vector . The alternative investment plans in which the intermediary company can invest, the revenue that these plans will provide to this company, and the revenue that will be paid back to the investor are represented in (1.1) by , , and , respectively. In BKP, the decision of the leader has a direct impact on the knapsack constraint of the follower. In fact, the decision of the individual (the leader) will set the capacity of the knapsack and determine the total amount of money that the intermediary (the follower) will be allowed to invest.

An alternative application of BKP occurs in telecommunications, and in particular in the problem of allocating bandwidth to different clients. An application in this area was addressed in [13] using an approach based on a bilevel programming problem with knapsack constraints. The BKP can be used in this context to model the interaction between a service provider and its competitors. The service provider can use its installed capacity to serve directly its clients, or it can grant capacity to another company that may use this capacity to route the demand of its own clients through the network of the service provider. The latter charges the company for this service, while the company will choose or not to reroute the traffic of its clients through the network of the service provider according to the offers of other competitors and so as to maximize its own profit. In this case, the leader is the service provider and the follower is the other company. The total capacity of the service provider is the coefficient in (1.1). The price that is charged by the service provider is represented by , while the amount of traffic required by the clients is given by and .

The remainder of the paper is organized as follows. In Section 2, we introduce different definitions and the notation that is used in the paper, and we describe the properties of BKP. In Section 3, we describe the details of our algorithm. We present our algorithm to solve the linear relaxation of BKP, the rules used to reduce the size of the problem, and the dynamic programming procedures developed to find the optimal solution of BKP. In Section 4, we report on computational results that illustrate the efficiency of our methods compared with the only available method from the literature [14]. Some final conclusions are drawn in Section 5.

#### 2. The Bilevel 0-1 Knapsack Problem

##### 2.1. Definitions

We introduce first the following standard definitions related to the bilevel 0-1 knapsack problem BKP described in the previous section:(i)the relaxed feasible set: (ii)the set of rational reactions of the follower for a fixed : (iii)the Inducible Region (IR), that is, the space over which the leader optimizes:

Using this notation, we can rewrite the BKP as follows.

When several optimal solutions exist for the follower problem, the previous model is not sufficient to define the optimal value of the problem because, for a leader decision, the follower can have several equivalent solutions. In this case, the solutions of the BKP can be defined either optimistically or pessimistically for each fixed leader variable [15]. These approaches can be described as follows.

(i) Optimistic
We assume that the leader can influence the decision of the follower in his favor. In this case, the problem to solve becomes and its optimal solution is called a weak solution.

(ii) Pessimistic
The follower takes his decision independently of the interests of the leader. In this case, the problem to solve becomes and its optimal solution is called a strong solution.

A detailed discussion of each approach can be found in [15]. The algorithms described in this paper can find both the strong and the weak solution of the problem. However, and for the sake of clearness, we will focus our presentation on the optimistic approach.

##### 2.2. An Upper Bound for BKP

The linear relaxation of BKP obtained by removing all the integrality constraints does not provide a valid upper bound for the problem. Hence, we resort to an upper bound for bilevel programming problems provided by a relaxation of (1.1) called the high-point problem [9, 16]. The high point problem is obtained by removing the objective function of the follower and the integrality constraints. It is defined formally as follows: The optimal solution of this relaxation can be computed using a classical procedure for the knapsack problem [17].

##### 2.3. Computing a Feasible Solution for BKP

A feasible solution to BKP can be computed by solving a different optimization problem related to the follower problem as shown in the following proposition.

Proposition 2.1. Let . An optimal solution to the following problem denoted by is also feasible for :

Proof. As long as admits a feasible solution, its optimal solution is feasible for the follower problem of BKP since the knapsack constraint of the follower is satisfied due to (2.9) and (2.10). This optimal solution is also optimal for the follower problem because it takes into account the follower objective function on the follower variables.

An optimal solution for BKP can then be defined using as follows: Clearly, finding an optimal solution for BKP by solving for each possible value of is computationally expensive. To obtain a good feasible solution for BKP, in our algorithm, we solve the problem for a set of good candidate values for , which are obtained by solving the linear relaxation of BKP with the polynomial time procedure described in Section 3.1.

#### 3. An Exact Algorithm for BKP

Before we describe our exact algorithm for BKP, we define and discuss first its different components. Our algorithm relies on the computation of an upper and lower bound for BKP. The upper bound is computed by solving exactly the problem HBKP defined previously. The lower bound is obtained by solving first a linear relaxation of BKP using the polynomial time procedure described in Section 3.1, and then by solving the problem for different values of the parameter . The values of are associated to feasible solutions of the linear relaxation of BKP which are obtained by applying the polynomial procedure mentioned previously. The upper and lower bounds are used to fix variables of BKP to their optimal values (Section 3.2) and to further enhance the definition of the original problem so as to improve its resolution in the remaining steps (Section 3.3). The optimal value for the resulting problem is computed using dynamic programming. This value is then used to generate an optimal solution for BKP. The two phases of this dynamic programming procedure are described in Section 3.4. The outline of our exact algorithm is given in Section 3.5.

##### 3.1. A Polynomial Time Solution Procedure for the Linear Relaxation of BKP

In this section, we show that the linear relaxation of BKP can be solved up to optimality in polynomial time, and we describe a procedure that computes this optimal solution. First, we recall the formal definition of this linear relaxation that will be denoted by CBKP (for continuous bilevel 0-1 knapsack problem):

Now, we show that solving CBKP is equivalent to the resolution of the linear relaxation of a standard knapsack problem.

Proposition 3.1. Assume that the follower variables are sorted in decreasing order of the relative value between their profit and their weight in the knapsack constraint, that is, such that . If , the order between the corresponding variables is determined according to the objective function of the leader, that is, (in the pessimistic case, one will consider ). Let be a decision of the leader. In this case, the total resource consumed by the leader in the knapsack constraint is given by . Furthermore, let be defined such that . The reaction of the follower related to the decision will be as follows:

Proof. Indeed, for a given decision of the leader, the problem CBKP becomes a standard 0-1 knapsack problem:

In Algorithm 1, we describe a polynomial time procedure that generates a weak solution for CBKP. The algorithm is based on the same idea that is used to solve the standard 0-1 knapsack problem. We start by solving the knapsack problem associated to the leader variables and objective function: with being the optimal solution of this problem. Then, we solve the follower knapsack problem that results from the leader decision : with being the corresponding optimal solution. The algorithm enumerates all the nondominated feasible linear programming basic solutions starting by the solution . At each iteration, we move from a feasible basic solution to another by transferring the resources consumed by the leader to the follower. To clarify the procedure, we illustrate its execution in Example 3.2.

Algorithm 1: A polynomial time solution procedure for CBKP.

Example 3.2. Consider the following continuous bilevel 0-1 knapsack problem denoted by CBKP1: We start by sorting the variables of the leader in decreasing order of the ratio , which results in the sequence (with ), (with ), and (with ). A similar ordering is applied to the variables of the follower resulting in the sequence (with ), (with ), (with ), and (with ). Note that and have the same ratio , but with outperforms with . In this example, we are considering the optimistic case. In the first phase of the procedure, we solve the following problem: The optimal solution of this problem is , , and . The optimal reaction of the follower for this decision of the leader is and . The solutions generated at each iteration of the Algorithm 1 are described as follows:(1), , and , , , with ;(2), , and , , , with ;(3), , and , , , with ;(4), , and , , , with .The value denotes the optimal value of the leader objective function as introduced in Algorithm 1. The optimal solution of CBKP1 is obtained at the third iteration. This solution is achieved after a polynomial number of steps.

In Algorithm 1, all the nondominated feasible basic solutions of CBKP are visited. For each one of these solutions, we can associate a value for the parameter in . In Example 3.2, the value of is equal to 4, 3, 2, and 0 at the iterations 1 to 4, respectively. These values are equal to , with being the value of the leader variables of a given basic solution generated in Algorithm 1. As shown in Section 2.3, we can obtain a feasible solution for BKP by solving the problem using these values of .

A basic solution of CBKP has at most two fractional variables (one for the leader, and another for the follower). If a basic solution of CBKP is integer for both the leader and the follower variables, then this solution is feasible for and for BKP too. If all the variables of the leader are integer, and only one variable of the follower is fractional, then we can fix the values of the leader variables in and solve the resulting problem which becomes a single knapsack problem. In these two cases, it is always possible to find a feasible solution for , and hence for BKP. However, when one of the leader variables is fractional, the problem may be infeasible. This is due to the fact that we are considering that . Since there is no guarantee that the equation in has a solution, the corresponding problem may be infeasible.

Solving the problem for a single value of the parameter can be done efficiently using branch-and-bound, for example. Clearly, solving this problem for all the values of in is much more expensive computationally. In our algorithm, our approach to generate a good feasible solution for BKP consists in inspecting a restricted set of good candidate values for . For this purpose, we choose the values of that are associated to the best solutions generated by Algorithm 1. In Example 3.2, if we set , then the values of associated to the two best solutions generated by Algorithm 1 (obtained at the iterations 2 and 3) will be used as a parameter in . The problems that will be solved in this case are FBKP2 and FBKP3.

The feasible solution (and corresponding lower bound) that is generated using this approach can be used together with the upper bound provided by HBKP to reduce the size of the original problem BKP. This can be done by fixing the values of some variables to their optimal values. The strategies used to reduce the size of the original BKP are described in the next section.

##### 3.2. Reducing the Size of BKP Using Fixing Rules

A strategy to improve the resolution of 0-1 mixed integer programming problems which has been used extensively in the literature consists in fixing the values of some variables to their optimal value. Many authors [1820] reported different procedures based on this idea to reduce the size of multidimensional knapsack problems. In this section, we show that it is also possible to apply fixing rules to BKP, and we describe the procedure that we used in our algorithm.

In many cases, fixing variables to their optimal value can be done via inexpensive operations. In the sequel, we show how variables can be fixed using information on the upper and lower bounds for the problem.

Proposition 3.3. Let and be a lower bound for . One will use the notation to indicate the optimal value of a given problem. The following fixing rules apply:(i)for any , if , then can be fixed to the value ;(ii)for any , if , then can be fixed to the value .

Proof. Let be the optimal value for BKP, and let denote both the variables and of the leader and follower, respectively. Note that . Therefore, if , then inevitably , and the optimal value can be fixed to .

These fixing rules depend only on an upper and a lower bound for the problem. The stronger the upper and lower bounds are, the more effective will be the rules for fixing the variables.

To introduce a new fixing rule, we rewrite the problem HBKP (used to derive an upper bound for BKP) in its standard form as follows: where corresponds to the vector of slack variables and is the vector with all elements equal to 1. Let and be the indices of the leader and follower variables and , respectively.

By reference to the LP basis that produces , we define is basic is basic} and is nonbasic is non-basic}. We subdivide to identify the four subsets , , , and .

Assume that is an optimal basic solution of HBKP. The problem HBKP can be written in the optimal basis related to in the following way: with being the optimal value of HBKP, and the vector of reduced costs corresponding to the variables of the optimal basis. For a given lower bound LB for BKP, we have The quantity is negative because of the negative reduced cost vector associated to the optimal basic solution, and the positive slack variables . Moreover, since for (resp., for ), and for (resp., for ), we can consider the following cut based on the reduced costs: This inequality can be used to derive the fixing rule introduced in the next proposition.

Proposition 3.4. If (resp., ) is a nonbasic variable, and (resp., ), then at optimality one has (resp., ).

Proof. The proof comes directly from the previous inequality (3.11).

Applying these fixing rules is useful for reducing the size of the original problem BKP, and hence to improve its resolution. However, because they do not take into account the objective function of the follower, these rules may result in problems whose solutions are infeasible for the original BKP. The problem occurs when the leader has solutions with the same value than the optimal solution, but which are infeasible for the original BKP because they are not optimal for the follower. To clarify this issue, we apply these rules on the case described in Example 3.2. The results are given in the following example.

Example 3.5. Consider the instance of BKP whose linear relaxation is given by CBKP1 in Example 3.2. We will denote this instance of BKP by BKP1. In Table 1, we describe an optimal solution for the corresponding problem HBKP, and we report on the values of the associated vectors of reduced costs . Furthermore, we specify whether a given variable is a basic variable or not by reference to the solution , and we identify the variables that can be fixed according to the fixing rules described previously.
Let UB and LB denote, respectively, the value of an upper and lower bound for this instance of BKP. The value of the solution given in Table 1 is 14, and hence we have UB = 14. By applying Algorithm 1, we obtain a lower bound of value LB = 14, as shown in Example 3.2. According to Proposition 3.4, since UB − LB = 0, all the nonbasic variables with an absolute reduced cost greater than 0 can be fixed, and hence we have , , , , and . The variable cannot be fixed because the absolute value of its reduced cost is not greater than UB − LB. Similarly, the variable cannot be fixed because it is a basic variable. Applying the fixing rules leads to the following problem: The resulting problem has two equivalent solutions. The first one consists in the leader action and the follower reaction . In this case, the complete solution (denoted by sol1) for the original problem is , and , and , , , and . The second solution consists in the leader action and the follower reaction . The complete solution for the original problem in this case (denoted by sol2) is , , and and , , , and . The value of both and is equal to 14. However, the optimal solution of original problem BKP is given by sol2, since for the leader action , , and the reaction of the follower , , , and is optimal for the follower problem. On the contrary, sol1 is not feasible for the problem because for the leader action , , and , the follower reaction should not be , , , and with the value 2 for the follower objective function. In this case, the follower reaction should be , , , and with a corresponding value for the follower objective function that is equal to 3.

Table 1: An optimal solution for HBKP (Example 3.5).

As shown in Example 3.5, the optimal solution of the problem can be found even when the fixing rules described in this section are applied. However, an additional treatment on the optimal solutions of the resulting problem is necessary to identify the solutions that are optimal for the follower problem (and hence feasible for the original problem BKP). To overcome this issue, in our algorithm, we fixed only the leader variables that are not directly influenced by the objective function of the follower.

##### 3.3. Reducing the Interval of Values for the Parameter in

In this section, we show how to decrease the knapsack capacity of the follower problem, and hence the size of the interval of possible values for in . Let and be the values of a lower and upper bound for in the problem . Initially, we have and , and hence . The smaller the size of the interval is, the easier the problem BKP will be to solve.

To improve the values of and , we solve the following two linear programming problems (denoted by and ) which relies on a lower bound LB for BKP. The optimal value of leads to a feasible value for , while leads to a feasible value for : Optimizing over the variable with the additional constraint in these two linear programs ensures that the resulting lower and upper bound for will not cut the optimal solution of the original BKP. In the next section, we show how this new interval helps in improving the performance of the dynamic programming component of our algorithm for BKP.

##### 3.4. Computing an Optimal Solution of BKP Using Dynamic Programming

In this section, we describe an approach based on dynamic programming to compute the optimal solution of BKP. The approach is divided into two phases. The first phase is a forward procedure whose objective is to find the value of an optimal solution for BKP. This forward phase divides in turn into two steps which are applied, respectively, to the leader variables and to the follower variables. The dynamic programming rules used for the follower variables are an extension of those used in [8]. In the second phase, a backtracking procedure is applied to generate a solution for the BKP with the value found in the forward phase. This dynamic programming algorithm has a pseudo-polynomial complexity, and it is able to solve both the optimistic and pessimistic cases mentioned previously. For the sake of brevity, we will focus our presentation on the optimistic case.

###### 3.4.1. Computing the Optimal Value of BKP: The Forward Phase

As alluded previously, the objective of the forward phase is to find the optimal value of BKP. This phase consists in two steps. The first step applies to the variables of the leader in BKP, and it considers only the objective function of the leader. The definition of this step relies on the interaction between the leader and the follower. For a given decision of the leader, the follower has to maximize his total profit using the corresponding residual capacity . For each value of , the best action for the leader has to be determined: . Hence, the dynamic programming subproblem for the leader states as follows: with .

The dynamic programming procedure for the leader in this first step of the forward phase is described in Algorithm 2. To illustrate the execution of this algorithm, we show in the following example how it applies to the instance of the BKP described in Example 3.2.

Algorithm 2: Forward procedure for the leader.

Example 3.6. Let us recall first the definition of the instance BKP1:
The results of the first step of the forward phase applied to the leader variables of BKP1 are given in Table 2. In this table, we report on the optimal values of the associated subproblems at this step.
Note that the value of is smaller than because there is no solution with a better value which consumes exactly 4 units of capacity.

Table 2: First step of the forward phase for the leader of BKP1.

In the second step of the forward phase, we focus on the variables of the follower. The problem that is solved at this stage is the following: Let denote the residual capacity associated with the leader action . In this second step, we consider both the leader and the follower objective functions, and we apply the forward procedure based on dynamic programming described in [8]. The objective is to determine all the reactions of the follower for a given action of the leader.

Two tables are generated in the second step of the forward phase: one that stores the optimal values of the follower (), and a second one that stores the optimal values of the leader values () with and . To illustrate the execution of the forward procedure for the follower, we applied it to the instance BKP1 used in the previous examples. The results are reported in Example 3.7.

Example 3.7. The results after the second step of the forward phase are reported in Table 3. This example shows that the values of the leader subproblems do not increase always because of the choice of the follower. For () and , the value for the leader decreases from 5 (for and ) to 2. This new value is associated with in order to satisfy the objective of the follower. Note that we applied the dynamic recurrence rules on the leader objective function for () and . The two values of the follower are equivalent: . In this case, the value for the leader is 5 because .

Table 3: Second step of the forward phase for the follower of BKP1.

This dynamic programming approach can be improved by fixing some variables of the problem BKP to their optimal value, and by reducing the size of the interval as discussed in the previous sections. Once this new interval has been computed, the first step of the forward phase can be applied with instead of . Since , this may reduce the number of steps of Algorithm 2. We do not apply this dynamic programming procedure up to the value of , because there is no solution with a value better than LB for . Similarly, in the second step of the forward phase, we use the following interval for : .

###### 3.4.2. Generating an Optimal Solution for BKP: The Backtracking Phase

Let be an optimal solution for BKP. The objective of the backtracking phase is to generate a solution with a value that is equal to the value computed in the forward phase. Before we introduce the backtracking procedure, we define first the optimal value that is determined in the forward phase. The optimal solution can be defined using the following rule: The main idea is based on the fact that for each leader decision with resources consumed, the follower reaction has to be optimum for the remaining resources.

From the value , we apply the backtracking procedure on the leader variables described in Algorithm 3. For the follower variables, we apply the backtracking procedure described in [8] by taking into account both the leader and the follower objective functions, and starting with the value .

Algorithm 3: Backtracking procedure for the leader.

For a given , if the follower has different equivalent choices, the value of is determined according to the profit of the leader. Note that the variable can take the value 0 or 1, if the two choices are equivalent for the leader and the follower. In Example 3.8, we illustrate the execution of the backtracking procedure on the instance BKP1 used in the previous examples.

Example 3.8. The optimal value for the problem BKP1 described in Example 3.2 is determined from Tables 2 and 3 as follows: with .
The results of the backtracking procedure for the follower and leader of BKP1 are given in Table 4. To determine the optimal action of the leader, we apply Algorithm 3 starting with . The optimal action for the leader is , , and . For the follower, we apply the backtracking procedure described in [8], starting from . The optimal reaction of the follower is , , , and .

Table 4: Backtracking procedure for the follower and leader of BKP1.
##### 3.5. Outline of the Algorithm

The outline of our exact algorithm for BKP is given in Figure 1. Each box in this figure corresponds to a step of our algorithm. The numbers identify the sequence by which the operations are performed.

Figure 1: Outline of the algorithm.

The algorithm starts by computing an upper bound for BKP through the exact resolution of HBKP. The next step consists in finding a good lower bound for BKP by computing a feasible solution for the problem. For this purpose, we solve first the problem CBKP using Algorithm 1. As referred to in Section 3.1, each solution generated by Algorithm 1 can be associated to a value of the parameter in . From the set of solutions found by Algorithm 1, we select the best solutions, and we solve the problems for the corresponding values of the parameter .

The upper and lower bounds (denoted, resp., by UB and LB in Figure 1) obtained in the previous steps are used to fix the variables of the leader to their optimal values. This is done by applying the fixing rules discussed in Section 3.2. The resulting problem is called the reduced problem in Figure 1. The lower bound LB is then used to reduce the size of the interval of possible values for . The new computed interval may help in reducing the number of steps of the dynamic programming procedures that are applied next. Similarly, the size of the reduced problem (i.e., solved with dynamic programming in the next step of the algorithm) is smaller than the original BKP, and hence, it is easier to solve using dynamic programming.

The next step of our algorithm consists in applying the forward phase of the dynamic programming procedure to the reduced problem in order to compute the value of an optimal solution for BKP. Finally, from this optimal value, an optimal solution for BKP is generated using the backtracking procedure described in Section 3.4.2.

#### 4. Computational Results

In this section, we report on the results of the computational study that was performed to evaluate the performance of our algorithm. Our approach is compared with other methods proposed in the literature. The limits of our algorithm are discussed, and a strategy to overcome these limits is presented and tested. All the algorithms analyzed in this section were coded in C++. We used the version 12.0 of the commercial optimization solver CPLEX. The experiments were conducted on a PC with 2.4 GHz with 4 GB of RAM.

Three sets of computational experiments were performed. In the first one, we compare our algorithm with the branch-and-bound algorithm proposed by Moore and Bard in [9]. In the second one, we perform a comparative study between our algorithm and the method of Brotcorne et al. described in [10]. The third set of experiments was conducted to analyze the impact of each component of our algorithm in the performance of our global approach. The limits of our algorithm are illustrated from a computational standpoint, and a variant that handles these issues is described and tested.

We used the generator proposed by Martello et al. [21] to generate instances of the knapsack problem. This generator gives us the data for the coefficients , , , and of BKP. The value of is computed as follows: , with . The input data for the leader () is generated randomly, such that all the coefficients are in the interval , with . We generated instances with uncorrelated coefficients (UC), and with correlated coefficients (C) [17].

In Table 5, we compare the performance of our algorithm (denoted by MACH1) with the branch-and-bound algorithm proposed in [9] (denoted by BM). For these experiments, we used a set of small instances with uncorrelated coefficients with and . We generated 5 instances for each set of instances characterized by the parameters and . In Table 5, we report on the average computing time (in seconds) required by BM and MACH1 to find an optimal solution for these instances. The computing times for BM and MACH1 are given, respectively, in the columns and .

Table 5: Computing time for BM and MACH1 for uncorrelated instances with and .

Table 5 shows the difficulty of the branch-and-bound algorithm of Moore and Bard in solving these instances, while our approach remains very fast. Note that the branch-and-bound algorithm is not able to find the optimal solution of medium instances with , , and and correlated coefficients in less than one hour. As we will see in the next experiments, our approach can solve these (and larger) instances very efficiently.

The results of our second set of experiments are reported in Table 6. We compare the performance of our algorithm with the method described in [10]. In the sequel, the latter will be denoted by BHM. The algorithm BHM is composed by two phases: the first phase is a dynamic programming procedure applied to the follower problem to determine all the possible reactions of the follower; in the second phase, a reformulated integer problem is solved by making the link between the actions of the leader and the reactions of the follower.

Table 6: Comparison between BHM and MACH1.

For these experiments, we used harder instances. The sets of instances are characterized by the parameters , , and . Again, we generated randomly 5 instances for each set. The parameters were chosen as follows: , , and . The coefficients were generated in the interval , and we considered both uncorrelated and correlated instances. For these experiments, we used a maximum time limit of 600 seconds.

In column , we give the number of times the algorithm BHM finds a proven optimal solution within the maximum time limit. Note that our algorithm always finds an optimal solution within this time limit. The average computing time required by BHM and MACH1 to find a proven optimal solution is given in the columns and , respectively. For BHM, the average time reported in Table 6 corresponds only to the cases where this algorithm finds a proven optimal solution within the time limit of 600 seconds.

From the results of Table 6, it is clear that our algorithm outperforms the approach of Brotcorne et al. [10]. Our approach remains very fast both for the uncorrelated and the correlated instances, while BHM is not able to find the optimum solution for most of the correlated instances. The performance of our algorithm is due in a large part to the strategies used for fixing the value of some variables, to our procedures for computing lower and upper bounds, and in particular to the strategy for reducing the interval of values for the parameter in . Note that our algorithm does not have any difficulty in proving the optimality of the solution found in the backtracking phase, since the optimal value is known at the end of the forward phase. The algorithm BHM spends more time precisely in its second phase when it solves the reformulated problem. At this stage, this algorithm has no information for the value of the optimal solution. Its computing time increases quickly with the correlation of the instances because in this case the size of the reformulated integer problem becomes larger.

In our final set of experiments, we focus on our algorithm. Despite its efficiency compared with other approaches, our algorithm may experience some difficulties with memory space for larger instances. These difficulties are illustrated in Table 7. The instances used in this case were generated as in the previous experiments, and with the parameters , , and given in Table 7. For each case, we generated 5 instances. Since these difficulties are due to the dynamic programming part of our algorithm, we used in these experiments a version of MACH1 in which the procedures described in Sections 3.1, 3.2, and 3.3 are disabled. We will denote this version by MACH1′. Table 7 reports the average computing time for the MACH1′ and for a variant that will be described hereinafter. The entry mem in Table 7 means that MACH1’ did not complete because of the memory space required for its execution. This problem arises for the largest instances with , , and . Recall that, for the coefficient , we have . In this case, the value of can be very large, and that is the main cause for this memory problem.

Table 7: Comparison between MACH1′ and MACH2′.

To overcome this issue, we propose a variant of the algorithm MACH1 (denoted by MACH2) that consists in replacing the backtracking phase based on dynamic programming in MACH1 by the exact resolution of the problem right after the forward phase. The forward phase gives us the optimal value for BKP. This optimal value is used for solving . Since we know this optimal value in advance, the resolution of becomes easier. In MACH2, we keep only two columns for dynamic programming at each iteration of the forward phase, and hence, the memory space that is necessary decreases.

In Table 7, we report on the average computing time required by a version of this variant without the procedures described in Sections 3.1, 3.2, and 3.3 (as in MACH1’). This version will be denoted by MACH2’. All the instances are solved up to optimality with a very small increase in the computing time compared to MACH1’. With this new variant, the problem with memory space does not occur anymore.

In Table 8, we compare the complete version of the algorithm MACH2 with the version MACH2′. In our implementation of MACH2, we solved the problem with the 10 best solutions generated by Algorithm 1 to find a valid lower bound. The objective of these experiments was to evaluate the impact of the additional components of our approach, namely, the polynomial procedure for solving CBKP described in Section 3.1 (Algorithm 1) and the reduction procedures described in Sections 3.2 and 3.3. We generated randomly 5 instances for each set of instances as in the previous experiments with the parameters , , and given in Table 8. The meaning of the entries in Table 8 is the following:(i): quality of the solution obtained with Algorithm 1 described in Section 3.1 (value of the best solution given by Algorithm 1 divided by the value of the optimal solution of the BKP);(ii): computing time (in seconds) required by Algorithm 1;(iii)fix: percentage of variables that were fixed;(iv): measure of the reduction achieved with the procedure described in Section 3.3; the values in this column are computed as follows: ;(v): computing time (in seconds) required by MACH2′;(vi): computing time (in seconds) required by MACH2.

Table 8: Comparison between MACH2′ and MACH2.

From the results of Table 8, we can observe that the additional components of the algorithm have a positive impact on the performance of our global approach. The average computing times for all the sets of instances decreased with MACH2. For the set of instances with , , and , the reduction is greater than 50%.

The lower bound given by CBKP is strong. Furthermore, it is computed very efficiently with Algorithm 1. The average computing time required by this algorithm is always smaller than 4 seconds. The fixing rules presented in Section 3.2 have a limited impact on the correlated instances. This can be explained by the quality of the upper bound that was considered (given by HBKP), and by the correlation between the coefficients of the instances. These rules perform better on the uncorrelated instances. While the lower bound on the optimal value of BKP does not seem to be very useful for fixing the values of the variables, it is for reducing the interval of feasible values for . Although the size of this interval decreases for all the instances, it is more significant for the uncorrelated instances. The reduction of the size of this interval has a strong influence on the resolution of the reduced problem with dynamic programming. Indeed, it implies reducing the capacity of the knapsack constraint at each step of the dynamic programming procedures. That explains in part the better performance of MACH2 compared with MACH2′.

#### 5. Conclusions

In this paper, we described a new exact algorithm for bilevel 0-1 knapsack problems (BKPs). We developed an original method for solving the linear relaxation of BKP, and we proposed a method for computing good feasible solutions for this problem using the information provided by the solutions of this linear relaxation. We described different strategies to enhance the resolution of BKP based on a valid upper and lower bound for the problem. Finally, we presented a dynamic programming procedure to find the integer optimal solution of the problem. To evaluate the performance of our approach, we conducted a set of computational experiments. Our results were compared with other algorithms proposed in the literature. The results that we obtained show that our algorithm clearly outperforms other state-of-the-art methods presented so far.

#### Acknowledgments

This work was partially supported by the Portuguese Science and Technology Foundation through the postdoctoral Gsssrant SFRH/BPD/64766/2009 for Raid Mansi, by the Algoritmi Research Center of the University of Minho for Cláudio Alves and José Valério de Carvalho, and by the International Campus on Safety and Intermodality in Transportation, the Nord-Pas-de-Calais Region, the European Community, the Regional Delegation for Research and Technology, the Ministry of Higher Education and Research, and the National Center for Scientific Research for Saïd Hanafi. The authors would like to thank the anonymous referees for their constructive comments, which helped improving the quality of the paper.

#### References

1. J. Bracken and J. T. McGill, “Mathematical programs with optimization problems in the constraints,” Operations Research, vol. 21, pp. 37–44, 1973.
2. H. Stackelberg, The Theory of the Market Economy, Oxford University Press, Oxford, UK, 1952.
3. K. Shimizu, Y. Ishizuka, and J. F. Bard, Nondifferentiable and Two-Level Mathematical Programming, Kluwer Academic, 1996.
4. S. Dempe, Foundations of Bilevel Programming, vol. 61, Kluwer Academic, Dordrecht, The Netherlands, 2002.
5. B. Colson, P. Marcotte, and G. Savard, “Bilevel programming: a survey,” Quarterly Journal of the Belgian, vol. 3, no. 2, pp. 87–107, 2005.
6. S. Dempe and K. Richter, “Bilevel programming with knapsack constraints,” Central European Journal of Operations Research, vol. 8, no. 2, pp. 93–107, 2000.
7. A. Plyasunov, “The bilevel optimization problem with multiple-choice knapsack problem in lower level,” Discrete Analysis and Research Operations, vol. 10, pp. 44–52, 2003.
8. L. Brotcorne, S. Hanafi, and R. Mansi, “A dynamic programming algorithm for the bilevel knapsack problem,” Operations Research Letters, vol. 37, no. 3, pp. 215–218, 2009.
9. J. T. Moore and J. F. Bard, “The mixed integer linear bilevel programming problem,” Operations Research, vol. 38, no. 5, pp. 911–921, 1990.
10. L. Brotcorne, S. Hanafi, and R. Mansi, “One-level reformulation of the bilevel knapsack problem using dynamic programming,” Tech. Rep., Université de Valenciennes et du Hainaut-Cambrésis, France, 2011.
11. S. Dempe, “Annotated bibliography on bilevel programming and mathematical programs with equilibrium constraints,” Optimization, vol. 52, no. 3, pp. 333–359, 2003.
12. O. Marcotte and G. Savard, Bilevel Programming: Applications, Kluwer Academic, Dordrecht, The Netherlands, 2001.
13. S. Kosuch, P. Le Bodic, J. Leung, and A. Lisser, “On a stochastic bilevel programming problem with knapsack constraints,” in Proceedings of the International Network Optimization Conference, 2009.
14. J. F. Bard and J. T. Moore, “An algorithm for the discrete bilevel programming problem,” Naval Research Logistics, vol. 39, no. 3, pp. 419–435, 1992.
15. P. Loridan and J. Morgan, “Weak via strong Stackelberg problem: new results,” Journal of Global Optimization, vol. 8, no. 3, pp. 263–287, 1996.
16. T. A. Edmunds and J. F. Bard, “Algorithms for nonlinear bilevel mathematical programs,” IEEE Transactions on Systems, Man and Cybernetics, vol. 21, no. 1, pp. 83–89, 1991.
17. S. Martello and P. Toth, Knapsack Problems: Algorithms and Computer Implementations, John Wiley and Sons, New York, NY, USA, 1990.
18. E. Balas and C. H. Martin, “Pivot and complement-a heuristic for 0-1 programming,” Management Science, vol. 26, no. 1, pp. 86–96, 1980.
19. D. Fayard and G. Plateau, “An algorithm for the solution of the 0-1 knapsack problem,” Computing, vol. 33, no. 5, pp. 1259–1273, 2006.
20. B. Gavish and H. Pirkul, “Efficient algorithms for solving multiconstraint zero-one knapsack problems to optimality,” Mathematical Programming, vol. 31, no. 1, pp. 78–105, 1985.
21. S. Martello, D. Pisinger, and P. Toth, “Dynamic programming and strong bounds for the 0-1 knapsack problem,” Management Science, vol. 45, no. 3, pp. 414–424, 1999.