Computational Intelligence and Neuroscience

Volume 2017, Article ID 1853131, 14 pages

https://doi.org/10.1155/2017/1853131

## The Artificial Neural Networks Based on Scalarization Method for a Class of Bilevel Biobjective Programming Problem

^{1}School of Information and Mathematics, Yangtze University, Jingzhou 434023, China^{2}School of Management, Huaibei Normal University, Huaibei 235000, China^{3}School of Mathematical Sciences, Beijing Normal University, Beijing 100875, China

Correspondence should be addressed to Tao Zhang; moc.621@189htam_tz

Received 6 January 2017; Revised 19 May 2017; Accepted 7 August 2017; Published 14 September 2017

Academic Editor: Leonardo Franco

Copyright © 2017 Tao Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

A two-stage artificial neural network (ANN) based on scalarization method is proposed for bilevel biobjective programming problem (BLBOP). The induced set of the BLBOP is firstly expressed as the set of minimal solutions of a biobjective optimization problem by using scalar approach, and then the whole efficient set of the BLBOP is derived by the proposed two-stage ANN for exploring the induced set. In order to illustrate the proposed method, seven numerical examples are tested and compared with results in the classical literature. Finally, a practical problem is solved by the proposed algorithm.

#### 1. Introduction

The bilevel programming problem (BLP) is a nested optimizations problem with two levels in a hierarchy: the upper and lower level decision-makers. The upper level maker makes his decision firstly, followed by the lower level decision-maker. The objective function and constraint of the upper level problem not only rely on their own decision variables but also depend on the optimal solution of the lower level problem. The decision-maker at the lower level has to optimize his own objective function under the given parameters from the upper level decision-maker. Since many practical problems, such as engineering design, management, economic policy, and traffic problems, can be formulated as hierarchical problems, BLP has been studied and received increasing attention in the literatures. During the past decades, some surveys and bibliographic reviews were given by several authors [1–4]. Reference books on bilevel programming and related issues have emerged [5–8].

The bilevel programming problem is a nonconvex problem, which is extremely difficult to solve. As we know, BLP is a NP-Hard problem [9–11]. Vicente et al. [12] also showed that even the search for the local optima to the bilevel linear programming is NP-Hard. Even so, many researchers are devoted to develop the algorithms for solving BLP and propose many efficient algorithms. To date a few algorithms exist to solve BLP; it can be classified into four types: Karus-Kuhn-Tucker approach (KKT) [13–16], Branch-and-bound method [17], penalty function approach [18–21], and descent approach [22, 23].

Unfortunately, the bilevel programming problem is nonconvex and the properties such as differentiation and continuity are necessary when proposing the traditional algorithms. Thus, many researchers tend to propose the heuristic algorithms for solving BLP because of their key characteristics of minimal problem restrictions such as differentiation. Mathieu et al. [24] firstly developed a genetic algorithm (GA) for bilevel linear programming problem because of its good characteristics such as simplicity, minimal problem restrictions, global perspective, and implicit parallelism. Motivated by the same reason, other kinds of genetic algorithm for solving bilevel programming were also proposed in [25–28]. Because of the prominent advantage that neural computing can converge to the equilibrium point (optimal solution) rapidly, the neural network approach was used to solve bilevel programming problem in [29–31]. In additional, since McCulloch and Pitts [32] and Pyne [33] utilized logical calculus to emulate nervous activities, there have been various types of analogue neural networks proposed for computation. Sheng et al. [34] firstly proposed a neural network approach based on Frank-Wolfe method for a class of BLP problems. Shih et al. [35] and Lan et al. [29] presented neural network for solving linear BLP problem, respectively. Recently, Lv et al. [30, 36] investigated the nonlinear bilevel programming and the convex quadratic bilevel programming by neural network. However, there are few results on application of neural networks to the BLBMPP in the literature.

Particle swarm optimization (PSO) is a relatively novel heuristic algorithm inspired by the choreography of a bird flock. Due to its high speed of convergence and relative simplicity, the PSO algorithm has been employed for solving BLP problems. For example, Li et al. [37] proposed a hierarchical PSO for solving BLP problem. Kuo and Huang [38] applied the PSO algorithm for solving bilevel linear programming problem. Jiang et al. [39] presented the PSO based on CHKS smoothing function for solving nonlinear bilevel programming problem. Gao et al. [40] presented a method to solve bilevel pricing problems in supply chains using PSO. Zhang et al. [41] presented a new strategic bidding optimization technique which applies bilevel programming and swarm intelligence. The hybrid algorithms based on PSO are also proposed to solve the bilevel programming problems [42–44]. Besides, Tabu search [45–47], simulated annealing [48], ant colony optimization [49], and -cut and goal-programming-based algorithm [50] are also typical intelligent algorithms for solving bilevel programming problem.

However, the algorithms mentioned above are only for the simple single objective bilevel programming problems. In fact, the multiobjective characteristics widely existing in the BLPP and the bilevel multiobjective programming problem (BLMPP) have attracted many researchers’ interesting. For example, Shi and Xia [51, 52], Abo-Sinna and Baky [53], and Nishizaki and Sakawa [54] presented an interactive algorithm for BLMPP. Eichfelder [55] developed a numerical method for solving nonlinear nonconvex bilevel multiobjective optimization problems. In recent years, the metaheuristic has attracted considerable attention as an alternative method for BLMPP. For example, Deb and Sinha [56–58] as well as Sinha and Deb [59] discussed BLMPP based on evolutionary multiobjective optimization principles. Based on those studies, Deb and Sinha [60] proposed a viable and hybrid evolutionary-local-search based algorithm and presented challenging test problems. Sinha [61] presented a progressively interactive evolutionary multiobjective optimization method for BLMPP. Lately, Zhang et al. [62] proposed an improved PSO for BLMPP and a framework of PSO for solving BLMPP is established. Subsequently, Zhang et al. [63] proposed an elite quantum behaved PSO for relatively complex BLMPP. In 2013, Zhang et al. [64] proposed a hybrid particle swarm optimization algorithm with crossover operator to solve high-dimensional BLMPP. Almost all the research object of the BLMPPs is the BLBOP, so we mainly consider the BLBOPs in this paper.

As we known, the authenticity of the lower level Pareto optimal solution is very important for the BLBOP. If the obtained optimal Pareto solutions possess the fraudulence, it can lead to the failure to solve the whole problem. In this paper, the induced set of the BLBOP is firstly expressed as the set of minimal solutions of a biobjective optimization problem by a scalar approach which can greatly improve the accuracy of the lower level Pareto optimal solutions. Based on the efficient set of the BLBOP, a two-stage ANN is presented for solving whole problem which can reduce the computation burden.

The remaining of this paper is organized as follows. In Section 2, we give the formulation of the model and related definitions. In Section 3, we will introduce the scalar approach for the induced set of the BLBOP and the two-stage ANN algorithm for the whole problem. In Section 4, some numerical examples and practical problem are given to demonstrate the feasibility and efficiency of the proposed algorithm, while the conclusion is reached in Section 5.

#### 2. Problem Formulation and Main Theory Results

Let be a nonempty subset of be a nonempty subset of , and , , , and be vector-valued mappings. We consider the following bilevel multiobjective programming problem (BLMPP): where and are the upper level and the lower level objective functions, respectively. and denote the upper level and the lower level constraints, respectively. Let and . For the fixed , let denote the Pareto optimal solutions to the lower level problem; the induced set of problem (1) is denoted as . Note that the constraint is uncoupled from the lower level variable . Particularly, if , we also call the BLMPP as bilevel biobjective programming problem (BLBOP). In the following, we shall focus on the BLBOP and assume that and .

*Definition 1. *Let be a closed pointed convex cone. A point is called a Pareto optimal solution of the lower level problem with respect to if .

*Definition 2. *For a fixed , if is a Pareto optimal solution to the lower level problem, then is a feasible solution to problem (1).

*Definition 3. *If is a feasible solution to problem (1) and there are no , such that , then is a Pareto optimal solution to problem (1), where “” denotes Pareto preference.

*Definition 4. *If is the optimistic solution for problem (1), then is given by .

*Remark 5. *The optimistic solution to the BLBOP is the one that optimizes the leader’s objective function over the set of efficient solutions to the follower, assuming that the follower has no preferences among the efficient solutions obtained for each leader’s decision or that the follower will choose the solution that most benefits the leader. In this paper, we only consider the optimistic BLBOP.

For problem (1), it is noted that a solution is feasible for the upper level problem if and only if is an optimal solution for the lower level problem with . In practice, we often make the approximate optimal solutions of the lower level problem as the optimal response feedback to the upper level problem, and this point of view is accepted usually. On the other hand, the authenticity of the lower level Pareto optimal solution is very important for the BLMP. If the obtained optimal Pareto solutions possess the fraudulence, it can lead to the failure to solve the whole problem. In this paper, we proposed the scalar approach for the lower level Pareto optimal solutions in order to improve the accuracy of the lower level Pareto optimal solutions.

#### 3. Algorithm

##### 3.1. The Scalar Approach for the Induced Set of BLBOP

In bilevel optimization, the constraint set of the upper level problem is given by the solution set of the lower level optimization problem. According to the Theorem 4.1 of the literature [55], the induced set of problem (1) is equivalent to the Pareto optimal set of the multiobjective optimization problem

Thus, to solve the induced set of BLBOP is transformed to solve the Pareto optimal solution set of problem (2). Inspired by the scalar approach adopted in [55], the approximation for the induced set of problem (1) can be solved by the following algorithm.

*Algorithm 1. **Step** 1*. Discretize the upper level variable.*Step 1.1*. For one-dimension upper level decision variable, choose and discretize by , .*Step 1.2*. For -dimension upper level decision variable, choose and discretize by .*Step 2*. Initialize the loop variables and predefine an accuracy measure .*Step 3*. Execute the following steps for and let .*Step 3.1*. Solve and . Then, determine and . Denote the approximate Pareto optimal solution set by . Let with a small and .*Step 3.2*. If , solve problem (2) using the scalarization method in [56] with , for the minimal solution . Set .*Step 3.3*. Calculate the according to Theorem 4.3 and determine according to (5.2) in [56]. Set and let ; go to Step 4.*Step 4*. Output the approximation Pareto optimal solution set ; that is, .

##### 3.2. The Two-Stage ANN

If in Step 2 is small, then set will consist of many points. Then, the determination of all nondominated points for the upper level problem can be very expensive. In this subsection, the two-stage ANN is presented for determining the Pareto optimal solution of the BLBOP on the induced set . The first stage is to map the vectors from to and the second stage is to determine the Pareto optimal solutions of problem (1). The detail of each stage is described as follows.

The first stage of the ANN is a feed-forward artificial neural network (FFANN) which is composed by two subnetworks and with the same structure and neuron output function. For each subnetwork, the nodes are organized into two layers and the weighted arcs only link nodes in lower layers to nodes in higher layers. The first stage is to compute the objective function value for the upper lever.

The second stage is a quasineural artificial network; namely, the network has no connectivity weight and the output value can be computed directly by software. The input layer of the quasineural artificial network is the output layer of the first stage network. For the hidden layer, the input of and the output are defined by (3) and (4), respectively.

For the output layer, the input of and the output are defined by (5) and (6), respectively.

Based on the set obtained by Algorithm 1, the first approximation Pareto optimal solutions of problem (1) can be achieved by the following algorithm.

*Algorithm 2. **Step 1*. Discrete induced set , denote the discrete set by , and let .*Step 2*. Divide the set into two subsets and randomly; that is .*Step 3*. Input and into subnetwork and subnetwork , respectively.*Step 4*. Select Pareto optimal solutions.*Step 4.1*. If , then . Let , go to Step 4.*Step 4.2*. If , then . Let , go to Step 4.*Step 4.3*. If , then . Go to Step 4.*Step 5*. If , then .*Step 6*. If , then .*Step 7*. If the stopping criterion is met, then stop. Otherwise, go to Step 2.

In Step 2, the two subsets and have approximately equal number of feasible solutions. The output of Algorithm 2 is the first approximation Pareto optimal solution set of problem (1); we denote the set by .

##### 3.3. The Algorithm for BLBOP

Based on Algorithms 1 and 2, as well as the refinement strategy used by the literature [9] which is employed in this paper, the algorithm for solving the BLBOP can be described as follows.

*Algorithm 3. **Step 1*. Based on the induced set obtained by Algorithm 1, we determine the approximation Pareto optimal solution set of problem (1) by Algorithm 2.*Step 2*. Let and choose the distance .*Step 3*. For any point , determine the refinement induced set around this point by the refinement strategy according to (5.4) in [55].*Step 4*. Set ; update the approximation Pareto optimal solution set by Algorithm 2.*Step 5*. If approximation of solution set of problem (1) is sufficient, then stop. Otherwise set , choose , and go to Step 3.

#### 4. Results

In this section, we considered seven numerical examples and a practical problem to illustrate the feasibility of the proposed algorithm for problem (1). In order to evaluate the closeness between the obtained Pareto optimal front and the theoretical Pareto optimal front, as well as the diversity of the obtained Pareto optimal solutions along the theoretical Pareto optimal front, we adopted the following evaluation metrics.

##### 4.1. Performance Evaluation Metrics

###### 4.1.1. Generational Distance (GD)

This metric used by Deb [65] is employed in this paper as a way of evaluating the closeness between the obtained Pareto optimal front and the theoretical Pareto optimal front. The GD metric denotes the average distance between the obtained Pareto optimal front and the theoretical Pareto optimal front:where is the number of the obtained Pareto optimal solutions by the proposed algorithm and is the Euclidean distance between each obtained Pareto optimal solution and the nearest member of the theoretical Pareto optimal set.

###### 4.1.2. Spacing (SP)

This metric is used to evaluate the diversity of the obtained Pareto optimal solutions by comparing the uniform distribution and the deviation of solutions as described by Deb [65]:where , , is the mean of all , is the Euclidean distance between the extreme solutions in obtained Pareto optimal solution set and the theoretical Pareto optimal solution set on the th objective, is the number of the upper level objective function, and is the number of the obtained solutions by the proposed algorithm.

All results presented in this paper have been obtained on a personal computer (CPU: AMD 2.80 GHz; RAM: 3.25 GB) using a c# implementation of the proposed algorithm.

##### 4.2. Numerical Experiment

In this section, we will present seven BLBOPS to illustrate the proposed algorithm for the bilevel biobjective programming. Problem 1 and problem 2 are low-dimensional problems. Problems 3–6 are high-dimensional problems. For problem 7, the theoretical optimal front is unknown. In this paper, we refined every upper problem’s feasible solutions three times and the obtained results are compared with the classical literature.

*Example 1. *Example 1 is taken from [57]. Here . In this example, the parameters of Algorithm 3 are 0.23, 0.11, and 0.03.Figure 1(a) shows the obtained Pareto front by the proposed algorithm and the method in [57]. Table 1 shows the comparison results between the two algorithms considering the metrics previously described. It can be seen that the performance of the proposed algorithm is better with respect to the generational distance, although it places slightly below the method in [57] with respect to spacing. By looking at the obtained Pareto fronts, some nondominated vectors produced by the method in [57] are not part of the true Pareto front of the problem; however, the proposed algorithm is able to cover the full Pareto front. Figure 2(b) shows that all the obtained solutions by the proposed algorithm, which follow the relationship; that is, , , and . However, some solutions obtained by the method in [57] do not meet the relationship.