Research Article  Open Access
Martin B. Bagaram, Sándor F. Tóth, Weikko S. Jaross, Andrés Weintraub, "A Parallelized Variable Fixing Process for Solving Multistage Stochastic Programs with Progressive Hedging", Advances in Operations Research, vol. 2020, Article ID 8965679, 17 pages, 2020. https://doi.org/10.1155/2020/8965679
A Parallelized Variable Fixing Process for Solving Multistage Stochastic Programs with Progressive Hedging
Abstract
Long time horizons, typical of forest management, make planning more difficult due to added exposure to climate uncertainty. Current methods for stochastic programming limit the incorporation of climate uncertainty in forest management planning. To account for climate uncertainty in forest harvest scheduling, we discretize the potential distribution of forest growth under different climate scenarios and solve the resulting stochastic mixed integer program. Increasing the number of scenarios allows for a better approximation of the entire probability space of future forest growth but at a computational expense. To address this shortcoming, we propose a new heuristic algorithm designed to work well with multistage stochastic harvestscheduling problems. Starting from the rootnode of the scenario tree that represents the discretized probability space, our progressive hedging algorithm sequentially fixes the values of decision variables associated with scenarios that share the same path up to a given node. Once all variables from a node are fixed, the problem can be decomposed into subproblems that can be solved independently. We tested the algorithm performance on six forests considering different numbers of scenarios. The results showed that our algorithm performed well when the number of scenarios was large.
1. Introduction
Uncertainty is common in all disciplines that involve decision making. In forestry, planners need to prescribe, several decades in advance, actions that should be taken in a forest in order to achieve a given management objective. Those actions include the segments of roads that should be built in a given period to allow hauling, the forest units or stands that should be treated to reduce the risk of fire, the stands that should be logged in each time period to produce timber and secure employment to the local community, etc. The most common objective is the maximization of the net present value subject to environmental, budgetary, and logistic restrictions [39]. The prescriptive models allowing to assign forest units to different actions in different time periods are known as harvestscheduling models.
To build harvestscheduling models, forest planners have traditionally used expected growth and yield coefficients to predict future merchantable timber volumes. However, uncertainties in longterm temperature and precipitation coupled with increasing wildfire, windstorms, or landslides due to climate change may affect forest development. The traditional approach fails to account for uncertainty in forest growth and leaves forest planners unprepared in case of occurrence of any of these uncertainties. The uncertainty in forest planning can be addressed by a particular class of mathematical programming known as stochastic programming. One can distinguish between twostage stochastic programming and multistage stochastic programming. In twostage stochastic programming, a decision is made, and then the uncertainty is revealed and a recourse action that depends on the revealed uncertainty is taken. However, in the case of multistage stochastic programming the sequence of decision and uncertainty revealing itself occurs more than once giving therefore more flexibility to the decision maker to take recourse actions as uncertainty unfolds progressively.
Multistage stochastic programs are mainly composed of scenarios and stages. Scenarios are the set of possible future outcomes of the uncertain parameter, and stages represent the level at which decisions can be made and/or recourse actions can be taken. We call “period,” the time that separates two consecutive stages. Uncertainty unfolds progressively in each period. Since at the beginning of the planning, we cannot anticipate which scenarios would unfold, we require that the decision made at the first stage must be the same for all scenarios. This requirement is known as nonanticipativity constraints. Nonanticipativity constraints (NACs) impose that if two scenarios cannot be distinguished up to any given stage, then the decision made in the two scenarios up to that stage must be the same. Multistage stochastic programming has been commonly used to model uncertainty in forest harvest scheduling because of the long planning horizon that characterizes forest harvest scheduling and the presence of interdependent relationships between periods such as the maximum contiguous area harvested from one period to another should not exceed a given limit. For instance, [41] solved a multistage stochastic harvestscheduling model with uncertainty in wood price, wood demand, and productivity while AlonsoAyuso et al. [1] focused on the uncertainty in the price and risk aversion. Finally, ÁlvarezMiranda et al. [2] assessed how key ecosystem services change in forest management if there is growth uncertainty stemming from climate change.
Although climate change might be one of the biggest challenges of forest management, it has received little attention in forest harvest planning in part because stochastic programs, especially multistage stochastic programs, are considered one of the most challenging classes of optimization problems to solve [5, 23, 40, 45]. For instance, the number of scenarios in [2] was limited to 32 although climate scientists forecast at least four climate pathways [31] which may translate into hundreds of possible forest growths.
Reallife applications of multistage stochastic problems lead to large models that are hard to solve directly [37]. The size of the models is directly related to the number of scenarios used to represent the uncertainty. Although uncertain parameters may be continuous, to make the problem suitable for stochastic programming, the discrete realization of the uncertainty is cast in a structure known as a scenario tree. In the scenario tree, each node represents a state decision and the branches of the tree represent the realization of the uncertainty. As more scenarios are included in the tree, the model will better approximate the entire probability space of the uncertain parameter [26]. However, increasing the number of scenarios makes the resulting stochastic programming models hard to solve. It is therefore necessary to develop some decomposition algorithms. The two most common decomposition methods are Benders decomposition [7] also known as stagewise decomposition or vertical decomposition and progressive hedging [36] also called scenariowise decomposition or horizontal decomposition. The study [38] provided an overview of the algorithms for stochastic programming decomposition. Additional summary of different solution methods for stochastic programming is available in [16].
Benders decomposition (BD) is a delayed constraint generation approach for solving mixed integer programs. It has been mainly applied to twostage stochastic programming problems with some assumptions on the nature of the firststage and secondstage variables. The authors in [27] give a summary of conditions for application of BD to twostage stochastic programs. According to [15], BD is well suitable when there is only a small set of constraints that prevent the decomposition of the problem into blocks. This is not the case in forest harvest scheduling characterized by long planning horizons and constraints such as even flow of wood and ending inventory linking variables from one decision stage to another of multistage stochastic programs. For instance, Egging [15] had limited success when extended their BD algorithm to multistage stochastic programs. We acknowledge, however, that there are multistage applications where BD had been used [17, 43, 44].
Progressive hedging (PH), on the other hand, was developed by [36] for convex twostage stochastic programs and is proven to produce a global optimal solution for continuous problems. For nonconvex problems such as stochastic mixed integer programs (SMIP), the algorithm is not proven to converge. It relaxes nonanticipativity constraints and iteratively solves the stochastic program by independently solving its scenarios and penalizing the violation of nonanticipativity constraints. PH is appealing because it is a method based on scenariowise decomposition and thus at each iteration the stochastic program solved is equivalent to a risk neutral problem that is the same as the deterministic problem which ignores uncertainty. Consequently, the algorithm can deal with a large number of scenarios.
Despite its benefits, PH performs poorly for nonconvex multistage stochastic programs. Several researchers have proposed PHheuristics for solving SMIP in different applications. The most promising PHbased heuristic explored is fixing variables that participate in defining nonanticipativity constraints as they meet consensus (see Section 2.2 for definition of nonanticipativity variables). For instance, Veliz et al. [41] fixed nonanticipativity variables during PH iterations as they meet consensus. After a given percentage of variables, 80%, for instance, consent on the value they should take, the reduced problem, which is the SMIP with some NAC variables fixed (named here and after reduced extensive form, REF), is then solved directly. The inconvenience of this approach is that the reduced extensive form may be infeasible because too many variables have been fixed (for proof, see Appendix). The infeasibility occurs when there is uncertainty in the yield of the forest such as in the case of climate change. In this case, the algorithm wasted considerable time iterating. On the other hand, solving the reduced extensive form problem can still be difficult if the number of variables fixed is too low. The third limitation stems from the fact that the reduced form problem is not separable and thus solving it cannot be parallelized. Similarly, for a twostage problem in resource management, the authors in [42] proposed a scheme for fixing variables. Their algorithm applicability was limited to a special class of resource management where constraints are one sided. They proposed “slamming” technique which forces nonanticipativity variables to converge, although they have not met consensus yet. They showed that this technique accelerated PH convergence. Recently, the authors in [30] proposed fixing variables as well. In their case, they had both binary and continuous decision variables and proposed therefore to fix the binary NAC variables, and the resulting linear stochastic program, since convex, can easily be solved using progressive hedging.
Aside from fixing variables, researchers have investigated other strategies for improving PH application to nonconvex problems. Some of the strategies are only applicable to twostage stochastic programs. For instance, Atakan and Sen [4] proposed a branchandbound algorithm for stochastic mixed integer programs and tested the algorithm on stochastic server location problem. Similarly, Barnett et al. [6] proposed a combination of branchandbound and PH using the former as a wrapper. In the same context, Gade et al. [19] proposed an algorithm for computing the lower bound of PH for twostage SMIP. Other researchers invested into reducing the duality gap that arises from relaxation of the nonanticipativity constraints [9, 19]. In the same spirit, Boland et al. [10] looked into the minimum number of nonanticipativity constraints that need to be reinforced to ensure the quality of SMIP. The authors in [8, 33] looked at reducing the number of scenarios by clustering them into bundles that can be solved independently. All these algorithms remain dependent on some assumptions of the nature of the firststage variables or the secondstage variables, or both. In addition, their applicability is limited to twostage SMIP and specific to certain disciplines. However, the main challenge that remains is that PH does not converge for nonconvex problems which are common in forest planning. In forest planning, there are some problems that possess only binary decision variables (there could be some accounting continuous variables which do not participate in decision making). Some of those problems are found in spatial explicit forest management planning where the decision variables are the forest units that should receive a specific treatment. This class of problems with pure binary decision variables is the focus of this research.
In this work, we aim at overcoming some limitations that are posed by using PH for multistage stochastic harvestscheduling problems with uncertainty in the yield. The objective is to have a heuristic that efficiently solves multistage stochastic harvest scheduling with little loss of optimality. (i) We propose decomposing SMIP into scenarios and solving them by a special heuristic form of progressive hedging that is completely parallelizable. We capitalize on the idea of fixing variables as they meet consensus and extend it so that the reduced extensive form problem is parallelizable. Our form of PH heuristic exploits the structure of the scenario tree by fixing variables starting from the rootnode. (ii) We assess the impact of fixing variables on the value of the optimal solution. (iii) We investigate as well some acceleration strategies that allow to accelerate the algorithm by extending the slamming technique proposed for twostage stochastic programs by [42] to multistage stochastic programming. (iv) We use climate change to show one source of uncertainty in the yield although there can be other sources of yield uncertainty such as the errors in measurement of forest growth or uncertainty of future yield due to prediction models. The algorithm is suitable as well to price uncertainty. Our aim is not to model perfectly the uncertainty in forest growth due to climate change but to provide an algorithm that can be used for reallife application in stochastic forest harvest scheduling.
The remainder of this paper is organized as follows. In Section 2, we provide a background on stochastic programming problem formulation and formally introduce the progressive hedging algorithm. In Section 3, the special form of PH for solving forest harvest planning is presented. We provide in Section 4 the computational experiment and the interpretation of the results. Finally, Section 5 concludes the paper and presents the limitations of the algorithm and future work.
2. Background
To illustrate variable fixation variant of the progressive hedging algorithm, it is necessary to represent the scenario tree that is the abstraction of the realization of uncertainty in growth and yield and formerly introduce progressive hedging (PH) algorithm.
2.1. Scenario Representation
The stochastic program representation can be visualized as a tree, the socalled “scenario tree.” It can be represented as follows. Let denote the set of periods in the planning horizon with being the number of periods. In the tree, a node represents the realization of the uncertain parameter and variable at a given time period. Let n and describe the node and the lexicographically numbered set of nodes in the tree, respectively. From each node , for , there is at least one branch leading to another node with a conditional probability with , that is, the set of nodes, immediate successors of the node . . We denote by the subset of nodes belonging to period such that and for . Let represent the finite set of representative scenarios in the tree. A scenario is a particular realization of the uncertain parameter represented as a path from the rootnode to a leafnode. Each scenario has an associated probability or weight denoted by . Note that . Similarly, let represent the set of nodes forming the scenario . In other words, is the set of nodes in a path from the rootnode to a leafnode. There is a number of scenarios that traverse each node. Let denote the set of scenarios that traverse the node . We have , and . Finally, let denote the set of successor nodes to the node , for . We have is a singleton and for (leafnodes). Furthermore, let denote the immediate ancestor node of node for , since is the rootnode of scenario tree. To introduce nonanticipativity, we need to define as the set of nodes having more than one leafnode as successor . For instance, in Figure 1, nodes 2 and 3 have as leafnode successors the set of nodes {5, 6} and {7}, respectively. Therefore, node but node 3 does not. Finally, Let represent the matrix of variables for all for the scenario . As the result, the stochastic program can be stated as follows:
Equation (1) maximizes the expectation from all scenarios. Equation (2) says that all the solutions should be feasible with respect to the constraints of the scenario. is the feasible set for scenario . Finally, equation (3) imposes the nonanticipativity constraints (NACs) which require that the solution up to a period from two scenarios should be the same if the two scenarios are indistinguishable up to period . We call this model the extensive form (EF).
Notice that we could write constraint (3) in a different form. Let be the vector of variables at node under scenario , and is the state variable of the forest unit or “stand” . Let represent a vector of binary variables at node . Imposing constraint (4) is equivalent to reinforcing NAC. Progressive hedging exploits the following formulation:
2.2. Progressive Hedging
Progressive hedging (PH) is a scenariobased decomposition algorithm proposed by [36] for stochastic programming models. The idea of progressive hedging is to relax the nonanticipativity constraints (NACs, equations (4)) in an augmented Lagrangian manner [18, 22, 35, 36] so that each scenario (subproblem) can be solved independently. This assumes that solving subproblems independently is much easier and faster. Nonanticipativity constraints require that values of variables that share the same ancestor nodes should be equal across scenarios up to that node. In other words, if two scenarios and are indistinguishable up to period , then the solutions of the two scenarios should be the same up to that period. For instance, in Figure 2, each variable at time should be the same across all the five scenarios and values of variables in period should be the same for scenarios 1 and 2. However, these constraints are not reinforced when scenarios are solved independently. Through progressive penalization of violations of NAC, the algorithm is proven to ultimately converge to the optimal solution for convex stochastic programs. In a nutshell, progressive hedging follows the following steps:(1)Solve each scenario without penalization(2)Compute the average for each variable(3)If solutions have sufficiently converged, then stop(4)Update penalization terms (5)Solve scenarios with penalization terms(6)Go to step 2
(a)
(b)
Algorithm 1 describes progressive hedging for multistage stochastic programming. The inputs of the algorithm are the penalty factor , the maximum number of iterations , and the termination criterion which indicates the level of consensus of nonanticipativity constraints that is acceptable. means that the algorithm stops if all the nonanticipativity constraints are satisfied. In the algorithm, line 2 initializes a Lagrangian multiplier for each NAC constraint. A Lagrangian multiplier is associated with each equation (4). To compute the average, of each variable, we need to have the conditional probability associated to each branch of the scenario tree in the detached form (Figure 2). Equation (5) computes the conditional probability from node to node which is an immediate successor. That probability is proportional to the number of leafnodes associated with node because the number of leafnodes informs on the number of scenarios passing by node . For instance, from Figure 2, , , and . Lines 6 and 17 compute the average of each variable at node . Lines 7 and 18 update the Lagrangian multiplier associated with each nonanticipativity constraint. Line 14 computes the NAC convergence euclidean distance. This distance is zero if all the NACs are satisfied:

3. Variable Fixation
Progressive hedging variable fixing (PHVF) algorithm is identical to the classic progressive hedging algorithm (Algorithm 1). However, instead of letting variables converge progressively, PHVF fixes variables as they converge. Line 10 in the progressive hedging algorithm (Algorithm 1) is replaced by Algorithm 2. The algorithm starts by fixing variables in the rootnode as they converge. For instance, for a given node , if a variable has a value of 1 across all scenarios , that variable will be fixed to 1 across all those scenarios. This process is better described in Algorithm 2. However, unless a variable belongs to the rootnode, it can only be fixed if all variables in the immediate ancestor have been entirely fixed. The first thing to notice is that once the scenario tree is decomposed, there are more scenarios traversing the rootnode (Figure 2). In addition, as we move away from the rootnode toward the leafnodes, there is a fewer number of scenarios passing by any given node. Hence, if for instance, there are three branches originating from the rootnode (as displayed in Figure 2), , then after fixing the rootnode, we get three distinct subextensive forms (SEFs) ( , and ) that can be solved independently (Figure 3). Furthermore, the three SEFs are much easier to solve compared to the original problem . The algorithm description exploits that property of the scenario tree.

3.1. Algorithm Description
During PHVF, at each iteration and for each scenario, the problem online 10 of Algorithm 1 needs to be solved. We refer to as the feasible set created by constraint (2) for each scenario. Note that if the uncertainty is only in the objective function (for instance, price uncertainty), then . In this case each solution of is feasible for . Since is the average of the values of the variables across scenarios (see line 6 of Algorithm 1), if its value is 1 (or 0), then it means all the scenarios met consensus that the variable should take a value of 1 (or 0), respectively. This realization is exploited in Algorithm 2 if it is supposed that the slamming factor . The algorithm moves to the node at the next stage and performs variable fixation if that node belongs to the set of nonanticipativity nodes and its predecessor is entirely fixed. The new feasible set created by constraint (2) and fixing some variables is denoted by . Note that this subproblem is much smaller compared to the original subproblem. In Algorithm 2, the inputs are , a function defining the value of the slamming factor depending on the stage , see Section 3.3 for more information on this parameter; which is the scenario; and which indicates the iteration.
Fixing variables impact the optimality since some variables may be fixed to values that they may not have taken in the optimal solution. Hence, the algorithm has an impact on the optimality. The objective of fixing variables is not to completely solve the model through the procedure but to obtain a reduced extensive form (REF) that is tractable. Let denote the percentage of NAC variables that are fixed before switching to solving the REF. is a proxy to the number of nodes fixed. The question is what is the appropriate that will make the REF tractable while not severely impacting the optimality? The answer to this question is investigated in Section 4.4.1.
3.2. Parallel Implementation
Algorithm 3 is designed for parallel implementation of PHVF. Hence, line 10 from Algorithm 1 has to be replaced by Algorithm 3. At each iteration of progressive hedging, the subproblems can be solved in parallel because they are independent. However, fixing variables and switching to solve the reduced extensive form (the extensive form with some variables fixed, REF) is not parallelized. Nonetheless, since our algorithm (PHVF) fixes variables from the rootnode to the leafnodes, after fixing each node, the problem can be separated into subextensive forms (SEFs) that can be solved independently. For example, as illustrated in Figure 3, after fixing the rootnode, we get three separate SEF models. Recursively, each subextensive form model can be solved through variable fixation leading to reduced subextensive form (RSEF). Each RSEF can then be solved directly if its size is deemed tractable. At the end, the solutions from all the REFs can be combined to get the solution of the stochastic program.

3.3. Acceleration Methods
Note from Algorithm 2 that if , then we fix a variable to 1 (or alternatively to 0) only if all scenarios of interest agree that the value of the variable should be 1 (or 0 alternatively), respectively. This requirement is hard to meet when dealing with hundreds of scenarios, especially for the rootnode variables which are replicated in all the scenarios. A variable may have the same value in all but one scenario for many iterations. Hence, that scenario slows the convergence. To avoid such a situation, “slamming” [42] forces variables to converge if the percentage of concordant scenarios for that variable reaches a threshold . However, instead of defining a scalar as in [42], we defined as a function that depends on the stage because if a low value of is acceptable for the rootnode, such a value is not acceptable when closer to the leafnodes (to avoid infeasibility). is the initial value of , whereas can be a linear or exponential function term that increases value as a function of the . Notice that if the uncertainty is in the objective function and we apply slamming, then the problem will always be feasible. In our case, preliminary tests allowed us to find the range of acceptable values of . Low values of led to infeasibility.
In the same line of thought, even with slamming, there is a chance that the algorithm is locked in a situation where there is no improvement of NAC convergence for many consecutive iterations. We define a “cascading” effect which lowers the value of to for one iteration if the algorithm does not improve (convergence of some nonanticipativity) for consecutive iterations. This behavior allows to avoid getting stuck, since lowering allows to fix some variables for which almost all the scenario already reached consensus. For those variables, the percentage of variables that agree on the value the variable should take is close to and the consensus could eventually be reached after several iterations. We have noticed that the cascading effect may lead to infeasibility because it may be forcing some variables to a value they would not take otherwise. When infeasibility arises, we roll back and eliminate the cascading effect. In general, there is a tradeoff between infeasibility and the value of . Low values of lead to a risk of infeasibility while raising that value may make the acceleration methods less efficient.
3.4. Stopping Criteria
In the case of classic progressive hedging, the algorithm stops because we have reached an acceptable level of consensus for NAC or because all the NACs are satisfied. The algorithm terminates as well if the maximum number of iterations is reached. In our algorithm, we keep these two stopping criteria, although we know that they may never be reached. We instead rely on the fact that at some points, many nodes will be fixed (NAC consensus is reached for some nodes) and the reduced subextensive forms can be solved directly. The number of NAC consensus reached is checked through the parameter which is the percentage of variables fixed. When the percentage of variables fixed is greater than or equal to , the algorithm switches to solve the REF or RSEF.
4. Numerical Experiment
We describe here an empirical performance analysis of our proposed algorithm. For easiness to follow our experiment and replicate the results, we formulate the stochastic version of the socalled “Model I” of forest harvest scheduling.
4.1. Problem Definition and Formulation
Indices : scenario : stand : time period Sets : set of stands : set of scenarios : set of scenarios passing by the node : set of nodes on which NAC should be reinforced : set of nodes at stage : set of time in the planning horizon Parameters : profit from selling wood in period ($/mbf). It is the discounted profit that includes selling cost : cost of harvesting and hauling wood from stand in period ($). It is the discounted cost : volume of wood harvestable per area from stand in period according to scenario (mbf/ac). This volume depends on the climate that materializes; therefore, it is a parameter that depends on the scenario : area of stand (acres) : age of stand at the end of the planning horizon if harvested in year : current age of stand : age of stand if not harvested during the planning horizon : allowable percentage of decrease of volume harvested from one period to another : allowable percentage of increase of volume harvested from one period to another : probability or weight of scenario Variables : binary variable taking a value of 1 if stand should be harvested in period according to scenario , and 0 otherwise : binary variable: 1 if management unit should not be harvested during the planning horizon under scenario , and 0 otherwise : accounting variable storing the volume of wood harvested in period according to scenario (mbf):which is subject to
Expression (6) maximizes the profit from timber harvest from all scenarios weighted by their respective probabilities. Constraints (7) require that a stand is at most harvested once during the planning horizon. We use the variable to capture stands that are not prescribed to be harvested during the whole planning horizon. We need that variable to compute the average ending age of the forest as shown in constraints (11). Constraints (8) compute the volume harvested in each period during the planning horizon. It uses the parameter which values depend on the forest growth scenario of interest. Note that the set of constraints (8) is not necessary. We could have written the same model without using that set of constraints. However, doing so would require rewriting constraints (9) and (10), and finally, it would negatively affect the readability of the model. Constraints (9) and (10) impose the even flow constraints so that the volume harvested from one period to another remain within an allowed fluctuation range. Constraints (11) require that on average, the forest at the end of the planning horizon is at least as old as the forest at the beginning of the planning horizon. Finally, constraints (12) impose the nonanticipativity constraints. It requires that for each stand , at time , if two scenarios are indistinguishable, then the decision should be the same for the two scenarios at that time. If , then there exists such that the two scenarios are indistinguishable at . Constraints (13) enforce that the decision variables are binary, and the accounting variables are continuous. We remind the reader that variables are not required for this model.
4.2. Climate Change Data
The potential mean annual increment, which is an indicator of forest growth in a year might change in the Pacific North West because of climate change. The change depends on temperature, precipitation, and air moisture content, all driven by human activities and economic development [28]. In addition, the change is not geographically uniform. Hence, the change tends to be negative in Oregon compared to Washington State. Similarly, the change tends to be negative in low altitudes in Oregon. We used the data from [28] which forecast the potential mean annual increment change (pMAI) by the year 2100. The pMAI is the potential change in forest growth that will be observed in a given year; hence, it is a volume given as a function of time and area. We assumed linear change of the growth from now until that year. For instance, the growth change for next two decades is the double of the growth change for the next decade. The climate paths defined in the Pacific Northwest are A2, A1B, B1, and Commit (or C). Tables 1 and 2 present the values of potential mean annual increment change for each one of the climate paths. In Table 2, climate paths D1 to D4 are artificial climate paths that suppose higher pMAI. We built forest growth scenarios by assuming it is possible to transit from one climate path to another because of mitigation or intensification of climate change due to human actions.


4.3. Experimental Design
The experiments were conducted under a DELL desktop computer running on Windows with Intel(R) Core(TM) 2 Quad CPU @ 3.70 GHz and 8 GB of memory. During PHVF iterations, each scenario is solved at optimality gap of 1%. This is a premature stop. Nevertheless, this criterion proved to accelerate solution time for each scenario. Furthermore, the last solutions of a mixed integer program are the most difficult ones with no to little improvement of the objective function value. All the models were solved using IBM ILOG CPLEX 12.6 (CPLEX, [12]). The memory allocated for storing the nodes was 3,000 MB, and the nodes were set to be stored in a compressed format on the hard drive. All other parameters were left to their default values. The code was implemented in Java 10 using Concert Technology of CPLEX.
For parallel computation, we used the paradigm of masterworkers. The master is in charge of coordinating the PHVF algorithm while distributing the task of solving individual submodels or REF to the workers. Each worker sends back its solution upon completion. The workers compete for access to the memory. Therefore, the choice of the number of workers must be judicious to avoid the overhead which is the amount of time required to coordinate parallel tasks, as opposed to doing useful work. Preliminary experiments showed that two workers was the optimal number given the configurations of the computer. The general framework of the model is presented in Figure 4. The input data contain the information on the forest, the climate change (growth change), and the regulations that ought to be met by the harvest planning. The input is fed to the PHVF module that is the master governing the optimization process. The workers are the ones interacting with the optimizer (CPLEX in this case). They send the model to the optimizer which returns the solution. The advantage of this framework is that we could change the optimizer without having to readapt our algorithm. At the end, the output module collects the harvest planning and informs the decision making.
Forest growth scenarios were generated using the information on the growth change reported in [28] (see Tables 1 and 2). We assumed that due to climate change and climate change mitigation efforts, it is possible to transition from one climate path to another in two consecutive periods. Hence, for instance, it is possible to transition from climate path A2 in year 2020 to climate path B1 in year 2030. We further assume that the probability of transiting from one climate path to the other is the same. As result, the scenarios are considered equally probable. Note that this assumption does not have much incidence on the performance of the algorithm. Based on the number of scenarios generated, we defined small, medium, and big instances. The small instances have 64 scenarios (1 × 4 × 4 × 4 × 1) which is four branching for the first three periods. The medium instances have 256 scenarios which correspond to four branching for the first four periods (1 × 4 × 4 × 4 × 4). For small and medium instances, pMAI corresponding to each climate path is reported in Table 1. The big instances have 512 scenarios made of eight branching for the first three periods (1 × 8 × 8 × 8 × 1). The indepth description of the scenario tree generation methods and its quality are beyond the scope of this research. The interested reader could refer to [11, 14, 21, 24, 25, 29, 33] and more importantly [34] who describe scenario generation in the forest management framework. We tested our algorithm on six different forests with three being real forests, and the data of which are publicly available^{1} (P1, P34, P36). The other three were computergenerated forests (P75, P83, P100). The six forests are set to be located in three different altitude classes. The forests are supposed to be located in Oregon and in Washington State. The values of pMAI for the six forests for large instances are reported in Table 2. Table 3 shows the characteristics of the forests and the scenario instances. In the tables, values show the potential mean annual increment change by 2100. Hence, negative values mean there will be a decrease in forest growth, whereas positive values mean that there will be an increase in forest growth.

The medium instances were used as reference to investigate how many variables should be fixed in order to have a problem that is computationally tractable. For that purpose, we solved the stochastic models, using the serial version of our algorithm (Algorithms 1 and 2), setting to just 20%, 40%, 60%, and 70% before switching to solving the reduced extensive form. means some variables are fixed but the rootnode is not completely fixed, and means that all the rootnode variables are completely fixed and some variables from the second period nodes are fixed (as stated before, is proxy to the number of nodes fixed. This scheme is possible because there are 25% of NAC variables at each stage and thus at the rootnode as well).
We made the experiment by allocating 15, 30, 45, and 60 minutes for each problem. The time is chosen to allow the effect of to manifest itself. The tests were run with five repetitions, and the average of the objective function value was reported. Notice that if the time is too long, then eventually all problems can be solved, even the extensive form. Conversely, if the time is too short, then PHVF may not have enough time to finish fixing variables before the time limit and hence low values of will be favored. During the whole experiment, we defined . As results, for the first stage , . Similarly, we defined . These values were defined from empirical experimentation.
For assessment of our algorithm performance, each problem was solved using our algorithm (PHVF) and comparing that solution to the one obtained by CPLEX solving directly the EF for the equivalent wall clock time. In addition, we report the optimality gap produced by CPLEX and the gain which is calculated according to where is the objective function value obtained by solving directly the extensive form model and is the solution obtained by PHVF algorithm for the same runtime. For medium and large instances, we report as well the EF value after solving the EF problems for 86,000 seconds (24 hours).
4.4. Results of the Experiment
In addition to the number of scenarios, SMIP size grows as well with the number of units (stands) because we need to define the nonanticipativity constraint for each stand at the time periods on which constraints (12) should be imposed. The problem size increases almost by tenfold when going from small instances to the medium ones. The smallest problem in this experiment has 600k + binary variables and over 69k constraints (Table 3).
4.4.1. Impact of on the Optimality and Solution Time
Tables 4–9 report the effect of on the value of the objective function with respect to the runtime for forests P1, P34, P36, P75, P83, and P100, respectively. We report as well, which is the percentage of increase (if positive) or decrease (if negative) of the objective function value from the objective function value for for the same runtime.
 
T: termination because PHVF ran out of time while iterating; NA: not applicable; : penalty factor; : percentage of variables fixed; : cascading factor; : initial slamming factor. 
 
N: terminated because could not find feasible solution; NA: not applicable; : penalty factor; : percentage of variables fixed; : cascading factor; : initial slamming factor. 
 
: penalty factor; : percentage of variables fixed; : cascading factor; : initial slamming factor. 
 
T: termination because run out of time iterating; NA: not applicable; : penalty factor; : percentage of variables fixed; : cascading factor; : initial slamming factor. 
 
: penalty factor; : percentage of variables fixed; : cascading factor; : initial slamming factor. 
 
T: termination because run out of time iterating; N: terminated because could not find feasible solution; NA: not applicable, : penalty factor; : percentage of variables fixed; : cascading factor; : initial slamming factor. 
In summary, as expected, everything else being equal, longer runtimes allow a higher objective function. As we can see for which corresponds to fixing entirely the rootnode and fixing some variables from the second stage, we have numerical issues. The numerical issues have two sources. On one hand, the algorithm may run out of time while fixing the variables or finished fixing the variables but does not have enough time to find a feasible solution to the REF. On the other hand, the basis of the REF may be disturbed in a way that the number of constraints and variables is not balanced so it is hard for branch and cut algorithm used by CPLEX to find an integral solution for the REF in the allocated time. This is the case for forest P34 with no solution when even when the time is 1 hour. However, for the same forest, there is an integral solution when is raised to 70%. This behavior occurred mainly for forest P34 and P100 and is not observed when .
The second aspect is the impact of on the optimality. The advantage that higher values of have over the small ones tends to vanish when the runtime is long. Similarly, the improvement in optimality from to higher values of is low. In fact, for some forest such as P83, over , the objective function value diminishes as increases which can be interpreted as during PHVF, many variables are fixed to values that are suboptimal. It clearly appears that fixing just the rootnode is the best choice for these problems. It allows sufficient time for REF to find an integral solution while avoiding to impact the optimality, and limiting the disturbance on the structure of the basis matrix.
4.4.2. PHVF versus EF Solved Directly
Table 10 presents the results from PHVF using which is equivalent to fixing the rootnode and then solving in parallel the SEF. As we can see for the small instances (64 scenarios), solving directly the extensive form outperformed using PHVF. This means the overhead of decomposing the problem into scenarios and solving each one overweighs the benefit. For the medium instances, the results are mixed. Out of the six forests, PHVF outperformed the EF in two cases. Even when PHVF under performed, it had results that were less than 1% away from the optimal solution. The benefit of using PHVF over the state of art commercial solver such as CPLEX is highlighted when dealing with big instances (512 scenarios). PHVF outperformed the EF for the six forests. In the case of P1 which is the biggest model with 1,363 units (Table 3), EF completely failed to solve the model because it could not fit it in the memory. For other cases for the same runtime, PHVF reached gains ranging from 0.84% to over 767% corresponding to forests P34 and P100, respectively. Even after leaving EF for 24 h, PHVF run for 15 min still outperformed in some cases. Comparing the gain to the gap from EF suggests that PHVF almost reached the optimal solution or its solution has a relative optimal gap less than 1%.
 
NA: no results because of memory limitation. 
5. Conclusions
In this paper, we have developed a method for solving multistage stochastic mixed integer programs that arise in natural resource management such as forest harvest planning. Although tested for growth uncertainty due to climate change, the method is valid as well for other sources of uncertainty such as errors in the model predicting forest growth, and uncertainty in the price of wood. This algorithm is also applicable to other disciplines where the decision variables are binary.
In the forest industry, managers solve multiple instances of the same forest model. It is therefore necessary to have an algorithm that allows one to quickly solve the SMIP. Furthermore, having an algorithm that is faster allows the practitioners to explore different management options. However, problems tested in this experiment are quite small compared to the models in the industrial standard. Nevertheless, this requirement may be compensated by the availability of more powerful computation resources in the industry.
The PHVF presented here overcomes some limitations listed in the literature regarding the choice of the penalty term . It is documented that large values of lead to the phenomenon of cycles [9]. However, because PHVF fixes variables, such a problem is avoided. Notwithstanding all these benefits, PHVF has more parameters that need to be set compared to the classic PH. Fortunately, most of those parameters are easy to set and the limitation remains on the choice of the penalty term . In the preliminary explorations, too low values of led to some numerical issues while too high ones, although accelerated the convergence, negatively impacted the objective function value.
The future extension of this work is to investigate the impact of different parameters on the solution quality. Similarly, having an algorithm that can dynamically determine the value of the penalty term would be an improvement for practitioners. On practical aspects, the question of the optimal number of scenarios necessary to approximate the growth change under climate change remains open. This can be done by computing the value of the stochastic solution, the expected value of perfect information for different scenario trees [3, 13, 20, 32].
Appendix
Proof of Infeasibility of Classic Progressive Hedging Coupled with Variables Fixation
Let us consider a hypothetical problem under studies for a structure with four scenarios and three stages as shown in Figure 5. Let us suppose the forest has four stands. The decision variable for each scenario is therefore
(a)
(b)
The variables in the matrix are all binary. means that the stand should be harvested in period . Since a stand cannot be harvested more than once, .
During progressive hedging, at iteration , we can have this solution. We use the superscript to refer to the solution from scenario , whenever necessary:
Since solutions of the second stage are identical for scenarios 1 and 2, we can fix entirely the variables of the second stage for the two scenarios. Similarly, we can fix entirely the second stage variables for scenarios 3 and 4:
Comparing variables of the first stage across the four scenarios, we can see that we can fix all variables except for stand 3 which is scheduled for harvest in period 1 for scenarios 1 and 2 but is scheduled for harvest in period 3 for the two other scenarios (equation (A.4)). Therefore, the reduced problem will have all variables fixed except for the third period variables as well as for all scenarios. For this problem to be feasible, either or for all scenarios. However, there is a constraint that links the volume harvested in one period to another (Constraints (9) and (10)). Hence, setting means there will be less than acceptable volume in period 1 for scenarios 1 and 2. Similarly, setting leads to volume of zero in period 3 for scenarios 3 and 4 which may not be acceptable as well. This situation occurs because the productivity of each stand is different with respect to the scenarios.
Data Availability
The code and the data used in this paper are available upon request or in Github link: https://github.com/MartinBagaram/proressivehedging.
Conflicts of Interest
The authors declare no conflicts of interest.
Acknowledgments
Martin B. Bagaram was funded by Precision Forestry Coop of the University of Washington. Andres Weintraub acknowledges the support of ANID PIA/Support AFB180003 and Fondecyt 1191531.
References
 A. AlonsoAyuso, L. F. Escudero, M. Guignard, and A. Weintraub, “Risk management for forestry planning under uncertainty in demand and prices,” European Journal of Operational Research, vol. 267, no. 3, pp. 1051–1074, 2018. View at: Publisher Site  Google Scholar
 E. ÁlvarezMiranda, J. GarciaGonzalo, F. UlloaFierro, A. Weintraub, and S. Barreiro, “A multicriteria optimization model for sustainable forest management under climate change uncertainty: an application in Portugal,” European Journal of Operational Research, vol. 269, no. 1, pp. 79–98, 2018. View at: Publisher Site  Google Scholar
 R. M. Apap and I. E. Grossmann, “Models and computational strategies for multistage stochastic programming under endogenous and exogenous uncertainties,” Computers & Chemical Engineering, vol. 103, pp. 233–274, 2017. View at: Publisher Site  Google Scholar
 S. Atakan and S. Sen, “A Progressive Hedging based branchandbound algorithm for mixedinteger stochastic programs,” Computational Management Science, vol. 15, no. 6, pp. 501–540, 2018. View at: Publisher Site  Google Scholar
 M. B. Bagaram and S. F. Tóth, “Multistage sample average approximation for harvest scheduling under climate uncertainty,” Forests, vol. 11, p. 1230, 2020. View at: Publisher Site  Google Scholar
 J. Barnett, J.P. Watson, and D. L. Woodruff, “BBPH: using progressive hedging within branch and bound to solve multistage stochastic mixed integer programs,” Operations Research Letters, vol. 45, no. 1, pp. 34–39, 2017. View at: Publisher Site  Google Scholar
 J. F. Benders, “Partitioning procedures for solving mixedvariables programming problems,” Numerische Mathematik, vol. 4, no. 1, pp. 238–252, 1962. View at: Publisher Site  Google Scholar
 P. Beraldi and M. E. Bruni, “A clustering approach for scenario tree reduction: an application to a stochastic programming portfolio optimization problem,” Top, vol. 22, no. 3, pp. 1–16, 2013. View at: Publisher Site  Google Scholar
 N. Boland, J. Christiansen, B. Dandurand et al., “Combining progressive hedging with a Frank–Wolfe method to compute Lagrangian dual bounds in stochastic mixedinteger programming,” SIAM Journal on Optimization, vol. 28, no. 2, pp. 1312–1336, 2018. View at: Publisher Site  Google Scholar
 N. Boland, I. Dumitrescu, G. Froyland, and T. Kalinowski, “Minimum cardinality nonanticipativity constraint sets for multistage stochastic programming,” Mathematical Programming, vol. 157, no. 1, pp. 69–93, 2016. View at: Publisher Site  Google Scholar
 M. S. Casey and S. Sen, “The scenario generation algorithm for multistage stochastic linear programming,” Mathematics of Operations Research, vol. 30, no. 3, pp. 615–631, 2005. View at: Publisher Site  Google Scholar
 IBM ILOG CPLEX, “12.6 User’s Manual IBM ILOG CPLEX Division,” Incline Village, NV, USA, 2019. View at: Google Scholar
 N. Di Domenica, G. Mitra, P. Valente, and G. Birbilis, “Stochastic programming and scenario generation within a simulation framework: an information systems perspective,” Decision Support Systems, vol. 42, no. 4, pp. 2197–2218, 2007. View at: Publisher Site  Google Scholar
 J. Dupacová, G. Consigli, and S. W. Wallace, “Scenarios for multistage stochastic programs,” Annals of Operations Research, vol. 100, pp. 25–53, 2000. View at: Google Scholar
 R. Egging, “Benders Decomposition for multistage stochastic mixed complementarity problems  applied to a global natural gas market model,” European Journal of Operational Research, vol. 226, no. 2, pp. 341–353, 2013. View at: Publisher Site  Google Scholar
 L. F. Escudero, M. A. Garín, J. F. Monge, and A. Unzueta, “On preparedness resource allocation planning for natural disaster relief under endogenous uncertainty with timeconsistent riskaverse management,” Computers & Operations Research, vol. 98, pp. 84–102, 2018. View at: Publisher Site  Google Scholar
 M. Fattahi, K. Govindan, and E. Keyvanshokooh, “A multistage stochastic program for supply chain network redesign problem with pricedependent uncertain demands,” Computers & Operations Research, vol. 100, pp. 314–332, 2018. View at: Publisher Site  Google Scholar
 M. J. Feizollahi, S. Ahmed, and A. Sun, “Exact augmented Lagrangian duality for mixed integer linear programming,” Mathematical Programming, vol. 161, no. 12, pp. 365–387, 2017. View at: Publisher Site  Google Scholar
 D. Gade, G. Hackebeil, S. M. Ryan, J.P. Watson, R. J.B. Wets, and D. L. Woodruff, “Obtaining lower bounds from the progressive hedging algorithm for stochastic mixedinteger programs,” Mathematical Programming, vol. 157, no. 1, pp. 47–67, 2016. View at: Publisher Site  Google Scholar
 N. Gupta, G. Dutta, and R. Fourer, “An expanded database structure for a class of multiperiod, stochastic mathematical programming models for process industries,” Decision Support Systems, vol. 64, pp. 43–56, 2014. View at: Publisher Site  Google Scholar
 H. Heitsch and W. Römisch, “Scenario tree reduction for multistage stochastic programs,” Computational Management Science, vol. 6, no. 2, pp. 117–133, 2009. View at: Publisher Site  Google Scholar
 M. R. Hestenes, “Multiplier and gradient methods,” Journal of Optimization Theory and Applications, vol. 4, no. 5, pp. 303–320, 1969. View at: Publisher Site  Google Scholar
 J. L. Higle, B. Rayco, and S. Sen, “Stochastic scenario decomposition for multistage stochastic programs,” IMA Journal of Management Mathematics, vol. 21, no. 1, pp. 39–66, 2010. View at: Publisher Site  Google Scholar
 K. Høyland, M. Kaut, and S. W. Wallace, “A heuristic for momentmatching,” Computational Optimization and Applications, vol. 24, pp. 169–185, 2003. View at: Google Scholar
 K. Høyland and S. W. Wallace, “Generating scenario trees for multistage decision problems,” Management Science, vol. 47, no. 2, pp. 295–307, 2001. View at: Google Scholar
 M. Kaut and S. W. Wallace, “Evaluation of scenariogeneration methods for stochastic programming,” Pacific Journal of Optimization, vol. 3, no. 2, pp. 257–271, 2007. View at: Google Scholar
 S. Küçükyavuz and S. Sen, “An introduction to twostage stochastic mixedinteger programming,” in The Operations Research Revolution, pp. 1–27, INFORMS, Catonsville, MD, USA, 2017. View at: Google Scholar
 G. Latta, H. Temesgen, D. Adams, and T. Barrett, “Analysis of potential impacts of climate change on forests of the United States Pacific Northwest,” Forest Ecology and Management, vol. 259, no. 4, pp. 720–729, 2010. View at: Publisher Site  Google Scholar
 N. Löhndorf, “An empirical analysis of scenario generation methods for stochastic optimization,” European Journal of Operational Research, vol. 255, no. 1, pp. 121–132, 2016. View at: Publisher Site  Google Scholar
 D. Manerba and G. Perboli, “New solution approaches for the capacitated supplier selection problem with total quantity discount and activation costs under demand uncertainty,” Computers & Operations Research, vol. 101, pp. 29–42, 2019. View at: Publisher Site  Google Scholar
 R. H. Moss, J. A. Edmonds, K. A. Hibbard et al., “The next generation of scenarios for climate change research and assessment,” Nature, vol. 463, no. 7282, pp. 747–756, 2010. View at: Publisher Site  Google Scholar
 G. Pantuso and T. K. Boomsma, “On the number of stages in multistage stochastic programs,” Annals of Operations Research, vol. 292, pp. 581–603, 2020. View at: Google Scholar
 H. Pranevicius and K. Sutiene, “Scenario tree generation by clustering the simulated data paths,” in Proceedings of 21st European Conference on Modeling and Simulation, pp. 203–208, Prague, Czech Republic, June 2007. View at: Google Scholar
 I. Rios, A. Weintraub, and R. J.B. Wets, “Building a stochastic programming model from scratch: a harvesting management example,” Quantitative Finance, vol. 16, no. 2, pp. 189–199, 2016. View at: Publisher Site  Google Scholar
 R. T. Rockafellar and R. J.B. Wets, “Measures as Lagrange multipliers in multistage stochastic programming,” Journal of Mathematical Analysis and Applications, vol. 60, no. 2, pp. 301–313, 1977. View at: Publisher Site  Google Scholar
 R. T. Rockafellar and R. J.B. Wets, “Scenarios and policy aggregation in optimization under uncertainty,” Mathematics of Operations Research, vol. 16, no. 1, pp. 119–147, 1991. View at: Publisher Site  Google Scholar
 O. Sanei Bajgiran, M. Kazemi Zanjani, and M. Nourelfath, “Forest harvesting planning under uncertainty: a cardinalityconstrained approach,” International Journal of Production Research, vol. 55, no. 7, pp. 1914–1929, 2017. View at: Publisher Site  Google Scholar
 S. Sen, “Algorithms for stochastic mixedinteger programming models,” Discrete Optimization, vol. 12, pp. 515–558, 2005. View at: Publisher Site  Google Scholar
 R. St. John and S. F. Tóth, “Spatially explicit forest harvest scheduling with difference equations,” Annals of Operations Research, vol. 232, pp. 235–257, 2015. View at: Google Scholar
 J. Thénié and J.P. Vial, “Step decision rules for multistage stochastic programming: a heuristic approach,” Automatica, vol. 44, no. 6, pp. 1569–1584, 2008. View at: Publisher Site  Google Scholar
 F. B. Veliz, J.P. Watson, A. Weintraub, R. J.B. Wets, and D. L. Woodruff, “Stochastic optimization models in forest planning: a progressive hedging solution approach,” Annals of Operations Research, vol. 232, pp. 259–274, 2014. View at: Google Scholar
 J.P. Watson and D. L. Woodruff, “Progressive hedging innovations for a class of stochastic mixedinteger resource allocation problems,” Computational Management Science, vol. 8, no. 4, pp. 355–370, 2011. View at: Publisher Site  Google Scholar
 A. Yossiri, J. F. Cordeau, and R. Jans, “Benders decomposition for production routing under demand uncertainty,” Operations Research, vol. 63, no. 4, pp. 851–867, 2015. View at: Google Scholar
 H. Zhou, J. H. Zheng, Z. Li, Q. H. Wu, and X. X. Zhou, “Multistage contingencyconstrained coplanning for electricitygas systems interconnected with gasfired units and powertogas plants using iterative Benders decomposition,” Energy, vol. 180, pp. 689–701, 2019. View at: Publisher Site  Google Scholar
 J. Zou, S. Ahmed, and X. A. Sun, “Stochastic dual dynamic integer programming,” Mathematical Programming, vol. 175, no. 12, pp. 461–502, 2019. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2020 Martin B. Bagaram et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.