Abstract

A new linearizing method is presented for globally solving sum of linear ratios problem with coefficients. By using the linearizing method, linear relaxation programming (LRP) of the sum of linear ratios problem with coefficients is established, which can provide the reliable lower bound of the optimal value of the initial problem. Thus, a branch and bound algorithm for solving the sum of linear ratios problem with coefficients is put forward. By successively partitioning the linear relaxation of the feasible region and solving a series of the LRP, the proposed algorithm is convergent to the global optimal solution of the initial problem. Compared with the known methods, numerical experimental results show that the proposed method has the higher computational efficiency in finding the global optimum of the sum of linear ratios problem with coefficients.

1. Introduction

In this paper, we consider the following sum of linear ratios problem with coefficients: where and are all linear functions, are all arbitrary real coefficients, , , , and is a nonempty compact set. Here, symbols in the problem (SLRC) can be considered the weight of each linear ratio, since are all arbitrary real numbers; thus the sum of linear ratios with coefficients can be called the weighted sum and difference of linear ratios. Therefore, the problem (SLRC) considered in this paper popularizes the mathematical model of sum of linear ratios problem investigated in many literatures.

The sum of linear ratios problem with coefficients (SLRC) has attracted the interest of researchers and practitioners for many years. In part, this is because the problem (SLRC) and its special case have broad applications in many practical problems, for example, profit rates and pricing decisions problem [1], cluster analysis problem [2], bond portfolio optimization problem [3], and so forth. Another reason for the strong interest in the problem (SLRC) is that from a research point of view the problem (SLRC) poses many important theoretical and computational difficulties. This is mainly because the problem (SLRC) is a global optimization problem; that is, it shows multiple local optimal solutions which are not global optimal solutions.

During the past 20 years, many algorithms have been presented for globally solving special cases of the problem (SLRC), which are only developed for solving the sum of linear ratios problem with the assumption that all coefficients , . For example, when feasible region is a polyhedral set and , Konno et al. [4] proposed a parametric simplex method for sum of linear ratios problem without coefficients. When all numerators and all denominators , Wang et al. [5] proposed a linear relaxation method for the sum of linear ratios problem without coefficients. When the numerators and the denominators , Wang and Shen [6] and Ji et al. [7] presented two different branch and bound algorithms for globally solving the sum of linear ratios problem; Shen and Wang [8] first proposed a global optimization algorithm for the sum of linear ratios problem with coefficients, which is investigated in this paper. Recently, by using the linear characteristics of convex envelope and concave envelope of double variables product function, Gao and Jin [9] proposed an effective branch and bound algorithm for solving sum of linear ratios problem; by using subgradient and convex envelope, Pei and Zhu [10] presented a global optimization method for maximizing the sum of difference of convex functions ratios over nonconvex region, which can be used to globally solve the sum of linear ratios problem. However, to our knowledge, although there exist many algorithms to solve special cases of the problem (SLRC), due to its intrinsic difficulty, less theoretical research and algorithm design have been still expanded to globally solving the sum of linear ratios problem with coefficients (SLRC).

The purpose of this paper is to develop a new linearizing method for globally solving the sum of linear ratios problem with coefficients (SLRC). The proposed new linearizing method uses more information of the objective function of the problem (SLRC); by using the method, the linear relaxation programming (LRP) of the problem (SLRC) is established, which can provide a tighter lower bound of the global optimal value of the problem (SLRC) than the previous linearizing method in branch and bound algorithm and which can be used to reduce rapidly the growth of branching tree in the branch and bound algorithm for solving the problem (SLRC); therefore it can improve the computational efficiency of the algorithm. By successively subdividing feasible region of the problem (SLRC) and solving a series of the LRP, the proposed algorithm is convergent to the global optimum of the problem (SLRC). Finally, compared with the known methods, numerical experimental results imply that the proposed new linearizing method can be used to globally solve the problem (SLRC) with the higher computational efficiency.

The remainder sections of this paper are organized as follows. In Section 2, a new linearizing method is proposed; by utilizing the method, the linear relaxation programming (LRP) of the problem (SLRC) is established. In Section 3, a branch and bound algorithm for globally solving the problem (SLRC) is presented. Several test examples in the recent literatures and their numerical results obtained are listed in Section 4. At last, some concluding remarks are described in Section 5.

2. New Linearizing Method

To globally solve the problem (SLRC), by solving simple linear programming problems, we obtain easily the initial partitioned rectangle

By the characteristic of fractional function , , we have the denominator ; that is, or . If , by letting , we have . Thus, we can always assume that all denominators . If is an arbitrary function, then there always exists an enough large positive number such that ; thus we can assume . Therefore, we can always suppose that and , for all .

The important structure in the establishment of a branch and bound procedure to globally solve the problem (SLRC) is the calculation of lower bounds for this problem and for its subproblems. A lower bound of the global optimal value of the problem (SLRC) and its subproblems can be computed by solving a sequence of linear relaxation programming problems. The proposed new linearizing method for establishing linear relaxation programming of the problem (SLRC) is to underestimate or overestimate each function with linear functions or , . By the former assumption, let

Firstly, for all , for each , let

Let and be the linear underestimating function and the linear overestimating function of the over the interval . By the linear characteristic of the concave function , we can get the following inequalities: where

By the above inequality (5), it follows that

By the above inequalities (7), for each , then we can establish the underestimating function and the overestimating function of the function as follows: It is obvious that we have

Secondly, for all , we let

By the characteristic of the convex function, we can derive its linear underestimating function of over the interval as follows: such that

Based on the above discussion, for each , by (11)-(12), then finally we can derive the linear underestimating function of , which underestimates the value of the function as follows: such that

Similarly, for each convex function over the interval , we can derive its linear overestimating function as follows: such that

According to the above discussion, for each , by (15)-(16), then finally we derive the linear overestimating function of , which overestimates the optimal value of the function as follows: such that

By (9), (14), and (18), we get

Let Then, by (19), we have

According to the above linearizing method, for , we can construct the linear relaxation programming (LRP) of the problem (SLRC) over as follows:

According to the construction method of the linear relaxation programming (LRP), for , the problem provides a valid lower bound for the global optimal value of the problem .

The following theorem ensures that the linear function will approximate infinitely the corresponding function as .

Theorem 1. For all , then one can follow that the error as .

Proof. Let Then, we have
Firstly, in the following, we will prove that as . Since then we only need to prove that and as .
Let
Then, we have where
Let
Since is a concave function about , we know that can attain the maximum at the point . Let ; then through computing, we can derive Since as , then we have as .
Since is a convex function about , it can attain the maximum at the points or . Let ; then through computing, we can derive Since as , then we have as .
Therefore, we have
By the above proof, we have Since is a continuous and bounded function about variable and by the above (27) and (33), we have
Next, we consider the difference , and it follows that Since is a convex function about , for any defined in the former, then it follows that can obtain the maximum at the points or . Let Then, through computing, we can follow that Since as and as , it is obvious that as . Therefore, we have as .
Secondly, we will prove that as . Since then we only need to prove that and as . Thus, Since is a concave function about , for any defined in the former, then it follows that can obtain the maximum at the points . Let Then, through computing, we can get Since as and as , it is obvious that as . Therefore, we have as .
Thus, where
Let
Since is a convex function about , it can attain the maximum at the points or . Then, through computing, we derive Since as , then we have as .
Since is a concave function about , we can get the fact that can attain the maximum at the point . Then, through computing, we derive Since as , then we have as .
Therefore, we have By the above proof, we have
Since is a continuous and bounded function about variable and by (42) and (48), we have
By the above proof, we can follow that
Therefore,
By the above discussion, it is obvious that the conclusion is followed.

3. Algorithm and Its Convergence

In this section, based on the former new linearizing method, we present an effective branch and bound algorithm for globally solving the sum of linear ratios problem with coefficients (SLRC). The critical construction in ensuring that the proposed branch and bound algorithm is convergent to the global optimum of the problem (SLRC) is the selection of a reasonable branching rule. In this paper, we use a standard branching rule which is called bisection. The selected branching rule is described as follows.

Assume that the hyperrectangle is a node subproblem identified, select , and subdivide by partitioning the maximum edge of the interval into the two subintervals and .

Assume that is the optimal value of the LRP() and is the corresponding optimal solution of the LRP(). The steps of the proposed algorithm are as follows.

Step 1. Set the initial convergence tolerance , the initial number of iteration , and the set of initial active node . Let the initial upper bound and the initial feasible point set .
Calculate and by solving the over ; if is feasible to the problem (SLRC), update the feasible point set and the upper bound , if necessary. If , then the algorithm stops with being the global optimal solution of the problem (SLRC); otherwise, continue the following Step 2.

Step 2. Subdivide hyperrectangle into two new sub-hyper-rectangles by utilizing the proposed branching rule and denote the set of new partitioned subrectangles as .

Step 3. For every , calculate and by solving the over . If , then let ; if is feasible to the problem (SLRC), then update feasible point set , the upper bound , and the best known feasible point , if necessary, and let and update the lower bound .

Step 4. Let . If , then the algorithm stops with being the global -optimal value for the problem (SLRC) and being a global optimization solution for the problem (SLRC). Otherwise, let , select subrectangle satisfying , and return to Step 2.

Theorem 2. The above proposed algorithm either stops finitely with the global optimal solution for the problem (SLRC) or generates an infinite sequence of iterations , the limitation point of which will be the global optimal solution for the problem (SLRC).

Proof. If the proposed algorithm stops finitely at iteration , then, when the algorithm stops, we obtain , where is the global optimal value of the problem (SLRC). Thus, by the structure of the proposed algorithm, we know that must be a global optimal solution for the problem (SLRC).
If the proposed algorithm generates an infinite sequence of partitioned subrectangles , then, by exhaustiveness of the branching method, we know that the sequence must shrink to a singleton. By the design of the proposed algorithm we can follow that the sequence is nonincreasing, and the sequence is nondecreasing; thus, the sequence is nonincreasing. In conclusion, we can get the fact that the bounding operation is consistent and that the selection operation is bound improving; thus, by Theorem IV.3. in [11], we can follow easily that the sequence must converge to zero. Since , this shows that . Since , any cluster point of the sequence is a feasible point for the problem (SLRC) with the objective function value . Therefore, the conclusion is followed.

4. Numerical Examples

To verify the reliability and effectiveness of the proposed new linearizing method, several test examples that appeared in the recent literatures are implemented on an Intel(R) Core(TM)2 Duo CPU (1.58ā€‰GHZ) microcomputer. The proposed algorithm using the new linearizing method is coded in C++ and every linear relaxation programming problem is solved by simplex method. These test examples and their numerical results are described as follows.

Example 1 (see [9]). We have the following:
With , using the method proposed in this paper, the optimal solution is , the global -optimal value is 4.912587413, the number of algorithm iterations is 29, the maximal number of algorithm active nodes necessary is 9, and the computational time is 0.12715 seconds.
Using the method in [9], with , the optimal solution is , the global -optimal value is 4.9125, the number of algorithm iterations is 113, and the computational time is 201.626020 seconds.

Example 2 (see [6]). We have the following:
With , using the method proposed in this paper, the optimal solution is , the global -optimal value is 5.000000000, the number of algorithm iterations is 59, the maximal number of algorithm active nodes necessary is 19, and the computational time is 0.25843 seconds.
But, using the method in [6], with , the optimal solution is , the global -optimal value is 5, the number of algorithm iterations is 32, the maximal number of algorithm active nodes necessary is 32, and the computational time is 1.089285 seconds.

Example 3 (see [10]). We have the following:
With , using the proposed method in this paper, the optimal solution is , the global -optimal value is 3.000000000, the number of algorithm iterations is 37, the maximal number of algorithm active nodes necessary is 10, and the computational time is 0.109402 seconds.

Using the method in [10], with , the optimal solution is , the global -optimal value is 3.0009, the number of algorithm iterations is 1033, the maximal number of algorithm active nodes necessary is 200, and the computational time is 99.3570 seconds.

Example 4 (see [10]). We have the following:
With , using the proposed method in this paper, the optimal solution is , the global -optimal value is 4.090702948, the number of algorithm iterations is 21, the maximal number of algorithm active nodes necessary is 3, and the computational time is 0.0831093 seconds.
Using the method in [10], with , the optimal solution is , the global -optimal value is 4.0001, the number of algorithm iterations is 1640, the maximal number of algorithm active nodes necessary is 233, and the computational time is 120.6355 seconds.
By substituting verification, we know that the global optimal solutions of Examples 3 and 4 obtained using our algorithm are feasible and correct and the global optimal values of Examples 3 and 4 obtained using our algorithm are correct. From the above comparison, we know that the optimal solutions and optimal values for Examples 3 and 4 using our method are much better than those in the literature [10]; that is, using our algorithm, we can obtain the more accurate global optimal solution.
From numerical results for Examples 1ā€“4, the proposed new linearizing method can be used to globally solve the problem (SLRC) with the higher computational efficiency.

5. Some Extensions

The new linearizing method used for solving the problem (SLRC) can be extended to seek a global optimal solution of the sum of linear ratios problem with coefficients whose domain is not linear. Some extensions are given as follows.

When domain is convex set, by solving simple convex programming problems, we obtain easily the initial rectangle . For any , utilize the presented linearizing method to construct linear underestimating function of the sum of linear fractional function, so that we can establish a convex relaxation programming problem of the problem (SLRC) over . Use the algorithm proposed in Section 3, at each iteration, and solve convex relaxation programming subproblems rather than linear relaxation programming subproblems; we can globally solve the problem (SLRC) over convex set.

When constraint functions of the problem (SLRC) are also sum of linear fractional functions with coefficients, whose mathematical modeling can be reformulated as follows: where and are all linear affine functions, are all arbitrary real coefficients, , , , , and is a nonempty compact set.

Using the linearizing method proposed in Section 2 to construct the linear underestimating function of each function , , we can establish the linear relaxation programming (LRP1) of the problem (SLRC1) over as follows:

Using the same algorithm step in Section 3, at each iteration, and solving linear relaxation programming subproblem (LRP1), we can globally solve the sum of linear ratios problem (SLRC1) with sum of linear ratios constraints. To verify the feasibility of the proposed new linearizing method, two test examples are implemented on a microcomputer; test examples and their computational results are given as follows.

Example 5. We have the following: where .

With , using the proposed method in this paper, the global optimal solution is , the global -optimal value is āˆ’4.849404762, the number of algorithm iterations is 181, and the maximal number of algorithm active nodes necessary is 64.

Example 6. We have the following:
With , using the proposed method in this paper, the global optimal solution is , the global -optimal value is āˆ’0.888192268, the number of algorithm iterations is 98, and the maximal number of algorithm active nodes necessary is 18.
From numerical results for Examples 5ā€“6, the proposed new linearizing method can be extended to globally solve the problem (SLRC) with the weighted sum of linear fractional functions constraints.
It should also be noted that our approach could be extended to solve more general generalized linear fractional programming problems; this will constitute a subject for future research.

6. Concluding Remarks

In this paper, by utilizing the linear approximation of exponential and logarithmic functions, a new linearizing method is presented. Combining the linearizing method within the branch and bound scheme, a branch and bound algorithm is constructed for solving the problem (SLRC). By subsequently partitioning linear relaxation of the feasible region and solving a series of linear programming problems, the proposed algorithm is convergent to a global optimal solution of the problem (SLRC). Compared with the known methods, numerical experimental results show that the proposed new linearizing method can be used to globally solve the problem (SLRC) with the higher computational efficiency.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This paper is supported by the National Natural Science Foundation of China under Grant no. 11171094 and the Science and Technology Key Project of Education Department of Henan Province no. 14A110024.