Abstract

This study proposes an improved truncated Newton (ITN) method for the logit-based stochastic user equilibrium problem. The ITN method incorporates a preprocessing procedure to the traditional truncated Newton method so that a good initial point is generated, on the basis of which a useful principle is developed for the choice of the basic variables. We discuss the rationale of both improvements from a theoretical point of view and demonstrate that they can enhance the computational efficiency in the early and late iteration stages, respectively, when solving the logit-based stochastic user equilibrium problem. The ITN method is compared with other related methods in the literature. Numerical results show that the ITN method performs favorably over these methods.

1. Introduction

The main role of traffic assignment models is to forecast equilibrium link or path flows in a transportation network. These models are widely used in the field of transportation planning and network design. Traditionally, traffic assignment models are formulated as the user equilibrium (UE) or stochastic user equilibrium (SUE) problems, in which no traveler can reduce his/her actual or perceived travel time by unilaterally changing routes at equilibria [1, 2]. Among various types of traffic assignment models in the literature, the logit-based stochastic user equilibrium traffic assignment problem is most widely adopted and extensively studied [3]. This problem incorporates a random error term in the route cost function to simulate travelers’ imperfect perceptions, which follows Gumbel distribution [4]. The logit-based SUE problem can be equivalently formulated as a mathematical programming problem with a unique solution. This feature facilitates its usage in both theoretical and practical studies [58].

The widespread application of the logit-based SUE model makes its solution approach also receive considerable interest in recent years. In general, there are two classes of solution algorithms for the logit-based SUE problem. The first class is link-based algorithms. This type of algorithms uses link flow as its variable. Since link flow is an aggregate variable of different path flows, link-based algorithms do not require explicit path enumeration. It only assumes an implicit path choice set, such as the use of all efficient paths (Dial [9]; Maher [10]), or all cyclic and acyclic paths (Bell [11]; Akamatsu [12]). The most well-known link-based algorithm is the method of successive averages (MSAs) proposed in Sheffi and Powell’s study [4]. This algorithm uses a stochastic loading procedure to produce an auxiliary link flow pattern, and the search direction equals the difference between the auxiliary link flow and the current link flow. The step size for MSA is a predetermined sequence that is decreasing towards zero, such as where is the iteration index. Maher [10] made further modifications to the method of successive averages. In his study, the Davidon–Fletcher–Powell (DFP) method was used to generate a search direction and the cubic (or quadratic) interpolation was applied to estimate the optimal step sizes.

The other class is path-based algorithms. This kind of algorithm is built on path flow variables. It requires an explicit choice of a subset of feasible paths prior to or during the assignment. Unlike the link flow variable, path flow variable is a disaggregate variable which cannot be further decomposed. Therefore, different nonlinear programming methods can be utilized in a more flexible way. For example, Damberg et al. [13] extended the disaggregate simplicial decomposition method of Larsson and Patriksson [14] to solve the logit-based SUE problem. This path-based method iteratively solves subproblems that are generated through partial linearization of the objective function. The search direction is obtained by the difference between the solution of the subproblem and the current iteration point. Bekhor and Toledo [15] proposed using the gradient projection (GP) method to solve this problem. In their study, the gradient of the objective function was projected on a linear manifold of the equality constraints, with the scaling matrix being diagonal elements of the Hessian.

This study focuses on path-based algorithms for the logit SUE problem. To the best of our knowledge, almost all existing path-based algorithms have a linear or sublinear convergence rate, which is relatively slow when the iteration point is approaching the optimal solution. In order to improve the convergence, it is desirable to develop an algorithm with a superlinear convergence rate. Recently, Zhou et al. [16] proposed a modified truncated Newton (MTN) method to solve the logit-based SUE problem. This method consists of two phases. The major iteration phase is performed in the original space, while the minor iteration phase is performed in the reduced space. At each major iteration, a reduced Newton equation is approximately solved using the preconditioned conjugate gradient (PCG) method. The reduced variables in the reduced Newton equation can be changed dynamically, which facilitates the usage of the PCG method. Zhou et al. proved that the convergence rate of the MTN method is superlinear. It works very fast once the iteration point gets near to the optimal SUE solution. However, there are two important problems that are not resolved in Zhou et al.’s research. First, when the iteration point is far from the optimal SUE solution, the truncated Newton type methods are relatively slow. The reason for this phenomenon is not clear. Second, in Zhou et al.’s research, a dynamic principle on how to choose the basic route is proposed. This is only an intuitive principle. The rationale behind this principle is not explained.

With the aim of addressing the above two problems, in this study, we propose an improved truncated Newton (ITN) method for the logit-based SUE problem. The ITN method makes two improvements over the traditional truncated Newton method. First, a preprocessing procedure is introduced. This procedure utilizes the partial linearization method (Patriksson [17]) to generate a good initial point in the original space. It can largely replace the early iteration stage of the traditional truncated Newton method. Second, on the basis of the generated initial point, a static principle on how to partition the coefficient matrix and the variables is developed. With this principle, the computational efficiency of the truncated Newton method in the late iteration stage can be enhanced. Furthermore, the rationale behind these two improvements is analyzed theoretically, which broadens the theoretical significance of this study.

The remainder of the paper is organized as follows. Section 2 outlines the traditional truncated Newton method for a linear equality constrained optimization problem. Section 3 discusses some implementation issues when applying the traditional truncated Newton method to the logit-based SUE problem. Section 4 proposes a preprocessing procedure to determine a good initial point. Section 5 develops a maximal flow principle for the choice of the basic/nonbasic variable. Numerical experiments are conducted in Section 6. Section 7 wraps up the paper with conclusions and future research directions.

2. The Truncated Newton Method for a Linear Equality Constrained Optimization Problem

Consider the following convex problem:where is a strictly convex function that is twice continuously differentiable, is an matrix of full row rank, and . [P1] can be viewed as a general formulation of the logit-based stochastic user equilibrium problem that will be investigated in this study. We will first show how to solve [P1] by the truncated Newton method.

Since [P1] only involves linear equality constraints, it can be transformed into an unconstrained optimization problem using variable reduction technique. Specifically, the matrix and variable are partitioned as follows:where is a nonsingular matrix and , , and . is called the basic matrix and its columns correspond to the basic variables . is called the nonbasic matrix. The columns of correspond to the nonbasic variables .

Therefore, the constraints can be rewritten as

By rearranging the above equation, the basic variables can be expressed as follows:

Substituting equation (4) into [P1], we obtain the following reduced unconstrained problem:where is referred to as the reduced objective function.

Let be a feasible point for [P1-RED]. By approximating using a second-order Taylor series around , the following subproblem can be obtained:where and are the reduced gradient and reduced Hessian of , and is the difference between the nonbasic variable and the feasible point .

Clearly, [SUB-1] is a quadratic programming problem. A typical method for this problem is the preconditioned conjugate gradient (PCG) method. This method constructs a sequence of conjugate directions using the objective gradient and minimizes the objective function along each of the directions. Interested readers may refer to Chapter 5 in Nocedal and Wright [18] for a detailed description of this method. It is commonly known that for large-scale optimization problems, finding the exact solution of [SUB-1] is computationally intensive. The truncated Newton method is thus designed to alleviate this drawback by solving [SUB-1] approximately if is far from the optimal solution of [P1] and solving [SUB-1] more accurately when the optimal solution is approached.

Let be an approximate solution of [SUB-1] generated by the PCG method. According to Lemma A.2 in Dembo and Steihaug [19], defines a descent direction with respect to the reduced objective function . Hence, by finding an appropriate step size in this direction, the new solution point for the next iteration can be obtained.

In what follows, we will give a detailed description of the truncated Newton method for the linear equality constrained optimization problem [P1]. As elaborated above, this method consists of two phases. The major iteration phase transforms the original problem into an unconstrained one and applies the truncated Newton framework to solve it. The minor iteration phase uses the PCG method to solve a quadratic programming subproblem approximately.

The detailed steps of major iteration are described in Algorithm 1 below. It is performed in the reduced variable space.

Step 0: partition the matrix and vector as , , respectively. Transform [P1] into [P1-RED] according to equation (5).
Step 1: let be an initial point in the reduced space. Set .
Step 2: if is a sufficiently accurate approximation to the minimizer of [P1-RED], terminate the algorithm.
Step 3: solve [SUB-1] approximately to generate a search direction in the reduced space.
Step 4: compute a step size along , for which produces a sufficient decrease in the reduced function .
Step 5: set , . Go to Step 2.

The minor iteration is elaborated in Algorithm 2 below. In each minor iteration, is the sequence of iterations, is the gradient of the objective of [SUB-1] evaluated at , is the conjugate search direction, and and are scalars that are used to determine and , respectively.

Given a preconditioner and a forcing term ;
Step 1: set
        Solve for
        Set
        Set .
Step 2: set
        Set
        Set
        If the termination criterion is satisfied, return , and terminate.
        Else continue with Step 3.
Step 3: solve for
        Set
        Set
        Set
        Go to step 2.

From the above description of the truncated Newton method, we have the following two remarks:(1)The forcing term in Algorithm 2 is usually chosen to be (Nocedal and Wright [18] p168)where is a given positive parameter. Clearly, plays the role of controlling the solution accuracy of [SUB-1] for each major iteration. When the incumbent solution is far from the optimal solution, we have , which means that only a few inner iterations are sufficient to satisfy the termination criterion in Step 2. When the incumbent solution is near to the optimal solution, we have , which implies that more inner iterations should be performed.(2)In Step 2 of Algorithm 2, the reduced Hessian matrix need not be formed explicitly. Algorithm 2 only requires matrix-vector products, i.e., the value of for some vector . In the next section, we will use this feature to simplify the calculation process of the logit-based SUE problem.

Furthermore, we would like to emphasize that the truncated Newton method employed in this study is performed in a fixed reduced space. In other words, once the partitions of the matrix and vector are made, they remain unchanged in all major iterations. This is different from the modified truncated Newton method (MTN) proposed in Zhou et al. [16], for which the partitions can be changed from one iteration to another. Using a fixed partition simplifies both theoretical analysis and practical implementation of the algorithm. However, it puts forward higher requirements for the selection of the initial points and the basic/nonbasic variables, which is the reason why the improved truncated Newton (ITN) method is proposed. In what follows, we will first apply the traditional truncated Newton method to the logit-based SUE problem (Section 3) and then propose improvements made by the ITN method (Sections 4 and 5).

3. Solving the Stochastic User Equilibrium Problem Using Truncated Newton Method

3.1. Stochastic User Equilibrium Problem

As discussed in Section 1, stochastic user equilibrium problem is fundamental to the analysis of transportation systems. It concerns the distribution of travel demands to routes in a transportation network under the assumption that travelers have different perception errors when selecting routes. This problem is defined over a transportation network , where is the set of nodes and is the set of directed links in the network. Let be the set of all origin-destination (OD) pairs in the network, be the set of simple (loop-free) routes between OD pair , and be the travel demand between OD pair . For a route connecting OD pair , the route flow is denoted as . Let be the travel time on link , which is assumed to be a continuous and differentiable function of the flow on that link only. The logit-based SUE problem can be expressed as the following minimization problem (Fisk [3]):

In the above formulation, equation (8) is the objective function. It consists of an integral term and an entropy term. The parameter reflects an aggregate measure of people’s perception of travel costs. Equation (9) defines the demand/route flow conservation conditions Equation (10) indicates the nonnegativity constraints. The link-route flow relationship is characterized by equation (11), in which the indicator if route between OD pair uses link and otherwise.

By substituting equation (11) into equation (8), we obtain a minimization problem in terms of route flow variable only. Fisk [3] has proved that the objective function is strictly convex with respect to , which ensures the uniqueness of equilibrium route flows. On the other hand, it is well known that the logit model assigns strictly positive flows to all paths in the choice set. Therefore, the nonnegative constraints (10) are not binding at the optimal solution. Consequently, constraints (10) and (11) can be ignored and problem [SUE] is essentially equivalent to the equality constrained minimization problem [P1].

The optimal solution to the logit-based SUE problem yields an equilibrium flow distribution, which is fundamental to many transportation planning and network design problems. For practical transportation networks, in order to obtain the equilibrium flow distribution in a reasonably short time, a fast method that can cope with problem size should be employed. As is known, the truncated Newton method is one of the most accurate and fast methods for large-scale problems. In what follows, we will discuss how to use this method to solve [SUE].

3.2. Implementing the Truncated Newton Method to Solve the Stochastic User Equilibrium Problem

From a practical point of view, when applying the truncated Newton method to solve the logit-based SUE problem, some implementation issues should be addressed. Next, we will discuss them, respectively.

3.2.1. Application of the Variable Reduction Technique

The coefficient matrix for equation (9) can be written as follows:

This is a block diagonal matrix whose diagonal consists of vectors of all ones (in different length). Clearly, for each row of matrix , any variable whose coefficient is “1” can be chosen as the basic variable, and the rest variables in this row are nonbasic variables. The basic and nonbasic matrices are then formed by combining columns that correspond to the basic and nonbasic variables. It is easy to see that the basic matrix is an identity matrix, which is obviously invertible.

As is known, each variable in the logit-based SUE problem represents a specific route flow. For OD pair , let denote the basic route flow variable where is the index of the basic route, and denote the nonbasic route flow variable where is the index of the nonbasic route. Therefore, we can express the basic flow variable in terms of the nonbasic flow variables, i.e.,where is the set of indices of nonbasic routes between OD pair .

Next, we present an example to show how the coefficient matrix and the variable are formed and partitioned for a specific network.

Example 1. Consider the network shown in Figure 1, consisting of 4 nodes and 5 links. There are two OD pairs in the network, one is from node 1 to node 4 (indexed by OD pair 1) and the other from node 2 to node 4 (indexed by OD pair 2). The demand between each OD pair is , . There are three routes connecting OD pair 1, which are indexed by route (1, 1), route (1, 2), and route (1, 3), respectively. The two routes that connect OD pair 2 are numbered as route (2, 1) and route (2, 2).
The node sequence for each route is presented as follows:route (1, 1): node sequence 1-2-4,route (1, 2): node sequence 1-3-4,route (1, 3): node sequence 1-2-3-4,route (2, 1): node sequence 2-4,route (2, 2): node sequence 2-3-4.Let , , , , and denote the flows through the five routes; then, the demand/route flow conservation conditions for equation (9) isThe coefficient matrix for equations (14) and (15) isIf we choose as the basic flow variable for OD pair 1, and as the basic flow variable for OD pair 2, then the basic flow variables in equations (14) and (15) can be expressed in terms of the nonbasic flow variables, i.e.,Equations (17) and (18) can be written in the following matrix form:from which we can observe that the basic and nonbasic for matrix (c.f. equation (16)) are

3.2.2. Computation of a Search Direction

By substituting equation (13) into equation (8), the reduced objective function can be rewritten as

Elements of the reduced gradient are given by

Elements of the reduced Hessian matrix arewhere indicator is equal to 1 if , , and 0 otherwise, and indicator is equal to 1 if , and 0 otherwise.

As mentioned before, when solving [SUB-1], we only need to compute the following matrix-vector product for some vector :

In equation (24), is any vector whose dimension equals the total number of the nonbasic route flow variables.

The coordinates of the vector can be calculated by

The diagonal elements of the reduced Hessian matrix are

Following the suggestion in Nash [20], the preconditioner matrix is chosen as a diagonal matrix whose elements are equal to the diagonal elements of the reduced Hessian. Obviously, with defined in this way, it can be easily inverted.

Up to now, all the ingredients required in the computation of the search direction are available. The search direction can be obtained by iteratively performing Algorithm 2.

3.2.3. Determination of the Step Size

For the logit-based SUE problem, although the optimal solution satisfies the nonnegativity constraints (10), some iteration points may violate these constraints during the iterative process. When this happens, the term that appears in equations (21) and (22) becomes undefined. To avoid such a circumstance, at each iteration, a restriction on the step size should be imposed to ensure strictly positive path flows, i.e., the step size should satisfy the following two constraints:

By incorporating the above restriction into the well-known Armijo line search rule, the step size can be determined bywhere is the smallest nonnegative integer which satisfies equations (27) and (28) and the following inequality:

4. A Preprocessing Procedure for the Determination of a Good Initial Point

The first important feature of the improved truncated Newton method is the incorporation of a preprocessing procedure into the traditional truncated Newton method. This procedure overcomes the drawback of the truncated Newton method and generates a good initial point to start with. It in essence replaces the early iteration stage of the truncated Newton method.

4.1. A Drawback of the Truncated Newton Method

The restriction in equations (27) and (28) is indispensable when applying the truncated Newton method, since it makes the reduced objective function (21) and the reduced gradient (22) well defined in all iterations. However, when the iteration point is far from the optimal SUE solution (this usually occurs in the early iteration stage), this restriction may deteriorate the performance of the algorithm. The reason is as follows, By the procedure of the Armijo rule (equations (29) and (30)), if anyone of the nonnegative constraints (10) is violated, the step size should be reduced so that each iteration is strictly feasible. Hence, in the early iteration stage, the actual step size is usually much smaller than needed, which will result in a very slow rate of convergence.

However, when the iteration point is near the optimal solution (this usually occurs in the late iteration stage), the restriction in equations (27) and (28) does not work. In addition, by Theorem 2.3 in Dembo and Steihaug [19], the truncated Newton method converges superlinearly when the iteration point is within a neighbourhood of the optimal SUE solution. As a result, the performance of the truncated Newton is quite fast at the late iteration stage.

The following proposition rigorously validates the above phenomenon.

Proposition 1. Let be the sequence of iterations generated by Algorithm 1. Suppose that is chosen as equation (7), and the step size is chosen according to equations (27)—(30); then there exists a positive integer such that the restriction in equations (27) and (28) can be automatically satisfied for all .

Proof. By Theorem 2.1 in Dembo and Steihaug [19], we havewhich, together with step 2 in Algorithm 2 and (7), impliesSince is the gradient of the objective of [SUB-1] evaluated at , we haveSubstituting equation(33) into equation (29) and noting that is positive definite, we obtainDefineThen equation (34) can be rewritten asSince is positive definite and , it follows that is a bounded sequence. In view of equations (35) and (31), we haveHence, for every , there exist a positive integer , such that for all ,Equation(38) implies that for every component of ,Assume that at iteration (), both and are bounded away from zero. Then there exist and such thatSince can be chosen sufficiently small and , it follows from Equations (27), (28), and (40) thatThe above two inequalities suggest that and are also bounded away from zero. Therefore, the conclusion of Proposition 1 holds true.
Furthermore, we would like to point out that in the early iteration stage of the truncated Newton method, although the subproblem (6) is only solved approximately, usually it still requires more than one inner iteration for each major iteration. This consumes more CPU times when compared with algorithms that do not need any inner iteration.
From all discussions above, if we find an initial point such that it is close to the optimal SUE solution, the drawback of the truncated Newton method can be avoided. This can be achieved by replacing the early iteration stage of the truncated Newton method with a preprocessing procedure, which will be discussed in the next subsection.

4.2. Preprocessing Procedure

The preprocessing procedure is proposed to find a good initial point to start with. It can largely replace the early iteration stage of the truncated Newton method. This procedure is based on the partial linearization method in Patriksson [17]. When performing this procedure, all iteration points will strictly satisfy the nonnegativity constraints, such that the step size restriction, (27) and (28), is not needed. Furthermore, at each iteration of the procedure, the generated subproblem has a closed form solution. Therefore, it does not require any inner iteration to solve this subproblem. Next, we elaborate the preprocessing procedure.

From equation (8), we know that the objective function for the logit-based SUE problem is composed of two terms. Without loss of generality, we reconsider the problem [P1] in the following equivalent form:

Suppose that at iteration , a feasible point is given. By approximating the first term of the objective function in [P2] with a first-order Taylor series around , the following subproblem can be obtained:

Let be an exact solution to [SUB-2]. By Theorem 2.1 in [17], it is known that if the vector is nonzero, it is a descent direction with respect to the original objective function . The next iteration point is then obtained through a line search along this descent direction.

Details of the preprocessing procedure are described in Algorithm 3.

Step 0: let be an initial feasible point in the original space. Set .
Step 1: if is close to the minimizer of [P1], terminate this procedure.
Step 2: solve [SUB-2] to obtain a solution , then the search direction in the original space is given by
Step 3: compute a step size where is the smallest nonnegative integer such that
Step 4: set , . Go to Step 1.

Clearly, by letting in [P2], we obtain the logit-based SUE problem. The resulting subproblem that corresponds to [SUB-2] iswhere is the travel cost on path based on the vector of path flows at iteration .

Unlike the complicated solution approach to [SUB-1], the solution to [SUB-2-SUE] in the preprocessing process can be given in closed form. The next proposition presents a detailed derivation of this solution.

Proposition 2. The subproblem [SUB-2-SUE] has a closed form solution, which can be explicitly expressed as

Proof. Consider the following Lagrange function:where is the Lagrange multiplier associated with equation (45). Then, [SUB-2-SUE] can be transformed into the following minimization problem:The first-order conditions for the above problem state thatHence, it follows from equation (49) thatSolving the above equation yieldsInserting into equation (45), we haveThis completes the proof.
In conclusion, the preprocessing procedure is an application of the partial linearization method to the logit-based SUE problem. Compared with the truncated Newton method, the preprocessing procedure has two advantages. On one hand, it is strictly feasible for all iteration points. Therefore, the step size restriction, (27) and (28), is not necessary so that the step size will not be forced to be reduced. On the other hand, it utilizes the special structure of the logit-based SUE model so that the resulting subproblem has a closed form solution. As a result, inner iterations are not needed, which can save a lot of CPU times. However, since the preprocessing procedure only uses a first-order approximation of the objective function, its convergence rate is sublinear. It may become very slow during the late iteration stage. This is why this procedure is only used to the take place of the truncated Newton method in the early iteration stage.
Convergence of the partial linearization method is established in Patriksson [17], which ensures that it is easy for the preprocessing procedure to find a good initial point. Since the preprocessing procedure is performed in the original space, it does not need to determine the basic and nonbasic variables. However, the outcome of this procedure lays a good foundation for what variable should be chosen as the basic/nonbasic variables.

5. A Maximal Flow Principle for the Choice of the Basic/Nonbasic Variables

The second feature of the improved truncated Newton method is the development of a practical principle for choosing the basic variables in the logit-based SUE problem. This principle is applied after the preprocessing procedure. It relies on the information contained in the initial point that belongs to the original space. Such a principle accelerates the PCG method used in the minor iteration, and it in essence improves the computation efficiency of the truncated Newton method in the late iteration stage.

5.1. The Principle and Its Rationale

The preprocessing procedure generates an initial point that is near to the optimal SUE solution. Since this point is in the original space, it can provide us valuable information on how to optimally partition the variables and the coefficient matrix. In view of the special structure of matrix A, it is known that any variable whose coefficient is nonzero can be potentially used as the basic route flow variable. However, in practice, different choice of the basic route flow variables affects the performance of the PCG method differently and thus yields different convergence rates of the inner iteration.

As discussed in Nocedal and Wright [18], the convergence behavior of the PCG method is strongly dependent on the condition number of the quadratic optimization problem (6), which is defined aswhere and are the largest and smallest eigenvalues of the matrix . Nocedal and Wright [18] showed that the larger the condition number, the slower the likely convergence of the PCG method. For the logit-based SUE problem, it is obvious that a different choice of the basic route flow variables yields a different . Therefore, among all possible choices, the most suitable way is to select a group of basic route variables such that the condition number is as small as possible. However, in practice, the value of is difficult to evaluate, which makes it hard to sort different condition numbers. Fortunately, based on the information contained in the initial point, we can at least avoid the case in which the condition number is very large. The next principle presents a strategy to avoid such a case.

5.1.1. The Maximum Flow Principle

Let be the initial point generated by the preprocessing procedure. The basic route can be set as the one that corresponds to the maximum flow route of each OD pair at the initial point. In other words, the basic route index for OD pair should satisfy

It is worth to note that the maximum flow principle proposed in this study is static. Once the basic routes are selected, they are unchanged in all remaining iterations. In [16], a dynamic principle with a similar idea is also discussed. However, that is only an intuitive principle inspired by a simple example. The rationale behind that principle is not clear. In this study, we will fill this gap by rigorously explaining the rationale behind the maximum flow principle. Details of the explanation are given in the following two propositions.

Proposition 3. At a certain initial point, if the basic route for each OD pair is chosen as the one whose flow is nearly zero, then the condition number of the reduced Hessian matrix defined in equation (53) will be very large.

Proof. Elements of the reduced Hessian matrix are given in equation (23). This equation can be decomposed into three parts.
The first part isThe second part isThe third part isAssume that for each OD pair, we choose the basic route as the one whose flow is nearly zero. Under this assumption, the values of equations(55) and (56) are limited, whereas the value of equation (57) will be quite large. The reason is that is the flow on the basic route. By the assumption that is nearly zero, will tend to infinity. Therefore, we can omit equations (55) and (56) and only consider equation (57). As a result, equation (23) approximately equalsFrom equation (58), we can observe that if the flow on the basic route is nearly zero, the reduced Hessian matrix can be approximately viewed as a block diagonal matrix. Elements in each block are nearly equal. Hence, there exist nearly linearly dependent columns in . This means is almost singular, and it has eigenvalues that are nearly zero. Since the reduced Hessian matrix is positive definite, all its eigenvalues are positive. As a result, the smallest eigenvalue presented in equation (53) is very small, which means that is quite large.
From Proposition 3, we know that at the initial point, if the basic route for each OD pair is chosen as the one whose flow is away from zero, we can avoid the case of very large condition numbers. At the remaining iteration points, this conclusion still holds. Proposition 4 below establishes it theoretically.

Proposition 4. Let be an initial point that is close to the optimal SUE solution. At , if the basic route is chosen according to the maximum flow principle, then for all remaining iteration points (), the flow on the basic route is always bounded away from zero.

Proof. At , let be the flow on the basic route. By the maximum flow principle, is the maximum variable for OD pair . Therefore, is bounded away from zero, and there exists such thatSince is close to the optimal SUE solution , we can assume thatwhere a very small positive number.
The above inequality implies thatBy Theorem 2.1 in Dembo and Steihaug [19], converges to the SUE solution as , and accordingly, the basic route flow variable also converges to . i.e.,In view of equation (61) and the definition of limit, we have that for all ,Adding equation (61) in equation (63) yieldsUsing the triangle inequality, we haveSince is very small, as a result of equations (59) and (65), we obtain that for all ,which implies that the flow on the basic route is away from zero for all .
The above two propositions rigorously explain the rationality of the maximum flow principle. In Proposition 3, the special structure of is crucial to prove the resultant assertion. In order to illuminate more clearly, we introduce an example below.

5.2. An Illustrative Example

Example 2. Consider the grid network in Figure 2. There are 9 nodes and 12 links in the network with only one OD pair between node 1 and node 9. The OD demand is 100. There are six different routes connecting the origin and destination.
The link travel time function is defined as follows:where , , , respectively, are link a’s free-flow travel time, capacity, and travel time with flow , and , are deterministic parameters. In this example, we set , . We assume that for all links; for links (4, 5), (5, 6), and (7, 8), ; and for the remaining links .
In this example, there are 6 route flow variables. The initial point is given by . At this point, if route () is chosen as the basic route, the reduced Hessian matrix corresponding to this route can be formed, and its condition number can be calculated. For abbreviation, we use “the reduced Hessian matrix for route ” and “the condition number for route ” to express the above meanings.
Tables 16 present the reduced Hessian matrices for routes 1–6. Substituting these matrices into equation (53), we can calculate the condition number for each route, which is shown in the lower left corner below each table.
By comparing these tables above, we can observe that the column vectors in Table 2 are nearly linearly dependent. This coincides with the fact that route 2 is the route whose flow is nearest to 0 (). Clearly, the condition number for route 2 is 397.28, which is much larger than the condition number for any other routes.

6. Numerical Results

In order to numerically justify the theoretical analysis conducted in this research, this section presents some performance comparisons between the ITN method and other typical methods in the literature. These methods include the MSA method (Sheffi and Powell [4]), the MTN method (Zhou et al. [16]), and the GP method (Bekhor and Toledo [15]). Among the above four methods, the MSA method is an implicit enumeration method. It uses Dial’s STOCH procedure for network loading, which implicitly considers all efficient paths of the network. The other three methods are explicit enumeration methods, which require an explicit choice of a subset of feasible paths prior to the traffic assignment. These methods are tested on the Sioux Falls network and Winnipeg network. The Sioux Falls network is a medium-size network. It is composed of 76 links, 24 nodes, and 528 OD pairs. The Winnipeg network is a real-size network. It consists of 2836 links, 1052 nodes, and 4344 OD pairs. Both networks are taken from Bar-Gera [21].

To provide a common basis for the comparison of the above three explicit enumeration methods, each of them is performed on the same working route set. We use a combination of the link elimination method (Azevedo et al. [22]) and link penalty method (De La Barra et al. [23]) to generate this working route set. For the Sioux Falls network, the average number of generated routes is 7.3 per OD pair, and the maximum number of generated routes is 11 for any OD pair. For the Winnipeg network, the average and maximum number of generated routes per OD pair are 20.3 and 29, respectively.

The performances of the three explicit enumeration methods are compared in terms of iteration number and CPU times under different circumstances. For the MTN method, we only concern its major iterations, because the number of minor iterations can be inherently reflected by its CPU times. Similarly, for the ITN method, its iteration number is calculated by the sum of the iteration number of the preprocessing procedure (Algorithm 3) and the major iteration number of the truncated Newton method (Algorithm 1). The convergence criterion for the ITN method is based on the root mean square error (RMSE) of the reduced gradient (Zhou et al. [24]), i.e.,where is the reduced gradient at the kth iteration and is the total number of routes.

In Step 2 of Algorithm 1, we use to terminate the ITN method. In Step 1 of Algorithm 3, we terminate the preprocessing procedure when RMSE is smaller than one-tenth of its initial value for the Sioux Falls network, and one-half of its initial value for the Winnipeg network. The parameter in the forcing term (7) in Algorithm 2 is set to be 0.5.

For all of the computation instances in this section, the start point is obtained by averagely assigning the demand to each route in the working route set between each OD pair. Our computer programs are coded in MATLAB and executed on a notebook computer.

6.1. Algorithm Performance

The algorithm performance test is carried out to compare the convergence progress of different methods at different iteration stages. In this test case, we use as the convergence progress measure. The dispersion parameter is assumed to be 0.5.

Figure 3 displays the convergence performance of the explicit and implicit methods in terms of iteration numbers on the Sioux Falls and Winnipeg network. This figure in essence illustrates the convergence rate of the four methods. We find that in the early iteration stage, the convergences of the MTN and GP method are relatively slow in comparison with the ITN and MSA method on both networks. Since the MTN method belongs to the category of the truncated Newton method, this phenomenon in essence confirms the analysis developed in Section 4.1, which indicates that the step size restriction may deteriorate the convergence rate of the truncated Newton type methods in the early iteration stage. As for the GP method, it also requires a maximum step size restriction to ensure strictly positive path flows in each iteration (c.f. Bekhor and Toledo [15]). With a similar reasoning, it is also slower than the ITN method in the early iteration stage.

We can also observe from Figure 3 that in the late iteration stage, the slope of the curve for the MSA method is much more gentle than the other three methods. The reason is that the predetermined step size employed in the MSA method diminishes to zero as approaches infinity, which deteriorates its rate of convergence. Furthermore, if we compare the three explicit methods, we find that the slopes of the curve for the ITN method and the MTN method are similar, both of which are steeper than the GP method. This observation is consistent with the fact that the ITN and MTN methods are locally superlinearly convergent, while the convergence rate for GP method is only linear. Furthermore, the fast convergence rate of the ITN method in the late iteration stage also validates Proposition 1, which suggests that the restriction on the step size does not work when the iteration point is close to the optimal SUE solution.

Figure 4 shows the convergence performance of the explicit and implicit methods in terms of CPU times on the Sioux and Winnipeg network. This figure in essence illustrates the computational efficiency of the four methods. From this figure, we can see that the computational efficiency of the ITN method is very high, which validates the effectiveness of the maximum flow principle proposed in Section 5.1. Compared with the MTN and GP methods, the ITN method can reduce the CPU times by roughly 50%–80%. The reason for the greater efficiency of the ITN method is two-fold. First, the number of iterations required by the ITN method is smaller than that of the MTN and GP method. Secondly, the preprocessing procedure of the ITN method does not entail any inner iteration, thus saving a lot of CPU times in the early iteration stage.

6.2. Sensitivity Analysis

By varying the value of or multiplying the model demand with different factors, we can examine the impact of the dispersion parameter or demand factor on the performance of different methods. As we learned from Section 6.1, the MSA method (i.e., the implicit method) is much inferior to the other three methods, especially when computing exact solutions. In order for the curves of different methods contained the relevant figures to be comparable, in the sensitivity analysis, we will omit the MSA method and only examine the performance of the ITN method with the MTN and GP method under various conditions.

6.2.1. Sensitivity Tests for Different Dispersion Parameters

The iteration numbers and CPU times by different for the three methods on the Sioux Falls and Winnipeg network are illustrated in Figures 5 and 6. As can be observed from Figure 5, the ITN method is superior to the MTN and GP method in terms of iteration number for all values of on both networks. This suggests that the average convergence rate of the ITN method is higher than the other two methods. As for the required CPU times shown in Figure 6, we can observe the following fact. For the Sioux Falls network, the ITN method consumes less CPU times than the MTN and GP method in all . For the Winnipeg network, the CPU times for the three methods are similar when is very small (i.e.,  = 0.1), but the ITN method consumes much less CPU times than the other two methods when is relatively large. This indicates that the ITN method is more efficient than the MTN and GP method for most practical cases.

Furthermore, we can find that the larger the dispersion parameter is, the more advantages the ITN method gains. The reason for this phenomenon is that if is large, the start point is much farther from the optimal solution than the case when the is small. Since the ITN method contains a preprocessing procedure which is tailored to avoid the step size restriction in the early iteration stage, the effects of this procedure will become more prominent if the start point is farther from the optimal solution.

6.2.2. Sensitivity Tests for Different Levels of Demand

Figures 7 and 8 present the iteration numbers and CPU times required by different demand levels for the three methods on the Sioux Falls and Winnipeg network. As illustrated in Figure 7, for the Sioux Falls network, the number of iterations of the ITN method is smaller than the MTN and GP method for all demand factors. For the Winnipeg network, when the demand level is relatively small, the ITN method requires fewer iterations in comparison with the other two methods. When the demand level is high, the performance of the ITN and that of MTN method are comparable, both of which are superior to the GP method. From Figure 8, we can observe similar results for the curves of the three methods in terms of CPU times.

In view of all discussions above, we conclude that the ITN method performs better than the MTN and GP method for most practical cases.

7. Conclusions and Future Research

This study investigated truncated Newton type algorithms for the logit-based SUE problem. We showed that using the traditional truncated Newton method to solve this problem may be very slow in the early iteration stage, and we thus proposed an improved truncated Newton method to overcome this drawback. We compared the ITN method with GP and MTN method on the Sioux Falls network. Numerical results validate that the ITN method can indeed improve the performance of these methods.

Further research work can be undertaken in the following aspects. First, the current research concentrates on solving the logit-based SUE model. Extending the ITN method to the more general traffic assignment models, such as the combined distribution and assignment model [25], the cross-nested logit model [26], and the generalized extreme value (GEV) family of model [27] is highly anticipated. Second, when implementing the ITN method, routes are generated prior to the traffic assignment. It is desirable to incorporate a column generation scheme to the ITN method, which can generate the path set dynamically during the assignment.

Data Availability

The data used to support the findings of this study are available from http://www.bgu.ac.il/∼bargera/tntp/.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this article.

Acknowledgments

This research is supported by the National Natural Science Foundation of China (nos. 71601046 and 51778141), Natural Science Foundation of Jiangsu Province (BK20160686), National Key R&D Program of China (2018YFB1600900), and Hong Kong Polytechnic University (1-BE1V).