- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Advances in Operations Research

Volume 2013 (2013), Article ID 279030, 10 pages

http://dx.doi.org/10.1155/2013/279030

## Well-Posedness and Primal-Dual Analysis of Some Convex Separable Optimization Problems

Department of Informatics, South-West University “Neofit Rilski”, 2700 Blagoevgrad, Bulgaria

Received 9 January 2013; Accepted 18 February 2013

Academic Editor: Hsien-Chung Wu

Copyright © 2013 Stefan M. Stefanov. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

We focus on some convex separable optimization problems, considered by the author in previous papers, for which problems, necessary and sufficient conditions or sufficient conditions have been proved, and convergent algorithms of polynomial computational complexity have been proposed for solving these problems. The concepts of well-posedness of optimization problems in the sense of Tychonov, Hadamard, and in a generalized sense, as well as calmness in the sense of Clarke, are discussed. It is shown that the convex separable optimization problems under consideration are calm in the sense of Clarke. The concept of stability of the set of saddle points of the Lagrangian in the sense of Gol'shtein is also discussed, and it is shown that this set is not stable for the “classical” Lagrangian. However, it turns out that despite this instability, due to the specificity of the approach, suggested by the author for solving problems under consideration, it is not necessary to use modified Lagrangians but only the “classical” Lagrangians. Also, a primal-dual analysis for problems under consideration in view of methods for solving them is presented.

#### 1. Introduction

##### 1.1. Statement of Problems under Consideration: Preliminary Results

In this paper, we study well-posedness and present primal-dual analysis of some convex separable optimization problems, considered by the author in previous papers. For the sake of convenience, in this subsection we recall main results of earlier papers that are used in this study.

In paper [1], the following convex separable optimization problem was considered:

where are twice differentiable strictly convex functions, and are twice differentiable convex functions, defined on the open convex sets in , respectively, for every , , and .

Assumptions for problem are as follows. for all . If for some , then the value is determined *a priori*.(). Otherwise, the constraints (2) and (3) are inconsistent and the feasible set , defined by (2)-(3), is empty. In addition to this assumption, we suppose that in some cases which are specified below.() (Slater’s constraint qualification) There exists a point such that .

Under these assumptions, the following characterization theorem (necessary and sufficient condition) for problem was proved in [1].

Denote by the values of , for which for the three problems under consideration in this paper, respectively.

Theorem 1 (characterization of the optimal solution to problem ). *Under the above assumptions, a feasible solution is an optimal solution to problem if and only if there exists a such that
*

The following polynomial algorithm for solving problem with strictly convex differentiable functions , was suggested in [1].

*Algorithm 2. *(0) Initialization: , , , ,, , , initialize , . If go to Step (1), else go to Step (9). (1) Construct the sets , , by using (4), (5), (6) with . Calculate
If then , go to Step (8) else if , then if go to Step (2) else if go to Step (9) (there does not exist such that ). (2). Calculate by using the explicit expression of , calculated from the equality , where , are given by (6). Go to Step (3). (3) Construct the sets , , through (4), (5), (6) (with instead of ) and find their cardinal numbers, , , respectively. Go to Step (4). (4) Calculate
where , are calculated from (6) with . Go to Step (5). (5) If or then , ,, , go to Step (8) else if go to Step (6) else if go to Step (7). (6) for , , , , , . Go to Step (2). (7) for , , , , , . Go to Step (2). (8) for ; for ; assign the value calculated from (6) for . Go to Step (10). (9) Problem has no optimal solution because or there does not exist satisfying Theorem 1. (10) End.

It is proved in [1] that this algorithm is convergent.

Theorem 3 (convergence of Algorithm 2). *Let be the sequence generated by Algorithm 2. Then, *(i)*if , then , *(ii)*if , then . *

In paper [2], the following two convex separable optimization problems were considered:

and

where for both problems, are twice differentiable convex functions, defined on the open convex sets in , , respectively, , for every , , and .

Assumptions for problem are as follows.() for each .(). Otherwise the constraints (10), (11) are inconsistent and , where is defined by (10)-(11).

Under these assumptions, the following characterization theorem (necessary and sufficient condition) for problem is proved in [2].

Theorem 4 (characterization of the optimal solution to problem ). * A feasible solution is an optimal solution to problem if and only if there exists a such that
*

Assumptions for problem are as follows.() for all .(). Otherwise the constraints (13), (14) are inconsistent and , where is defined by (13)-(14).

Under these assumptions, the following theorem (sufficient condition) for problem is proved in [2].

Theorem 5 (sufficient condition for optimal solution to problem ). *Let, be components of the optimal solution to problem . Then:*(i)* If , then ,, solve problem as well. *(ii)* If , then , defined as follows: * *,
* *,
* *for all such that ,* * for all , such that **solve problem .*

The following polynomial algorithm for solving problem with strictly convex differentiable functions was suggested in [2].

*Algorithm 6. *(1) Initialization: , , , , , , initialize . If go to Step (2), else go to Step (9). (2) . Calculate by using the explicit expression of , calculated from the equality constraint , where ,, are given by (17). Go to Step (3). (3) Construct the sets ,, through (15), (16), (17) (with instead of ) and find their cardinalities ,,, respectively. Go to Step (4). (4) Calculate
where ,, are calculated from (17) with . Go to Step (5). (5) If or then , ,, , go to Step (8) else if go to Step (6) else if go to Step (7). (6) for , , , , , . Go to Step (2). (7) for , , , , , . Go to Step (2). (8) for for ; assign the value, calculated from (17) for . Go to Step (10). (9) The problem has no optimal solution because .(10) End.

It is proved in [2] that this algorithm is convergent.

Theorem 7 (convergence of Algorithm 6). *Let be the sequence generated by Algorithm 6. Then,*(i)*if , then ,*(ii)*if , then .*

The following algorithm for solving problem with strictly convex differentiable functions is suggested in [2].

*Algorithm 8. *(1) Initialization: , , , , , , initialize . If then go to Step (2), else go to Step (9). Steps (2)–(7) are the same as Steps (2)–(7) of Algorithm 6, respectively. (8) If then for ; for ; assign the value, calculated through (17) for , go to Step (10) else if then for , for , if and then else if and then , go to Step (10). (9) Problem has no optimal solution because or there do not exist , such that . (10) End.

Since Algorithm 8 is based on Theorem 5 and Algorithm 6, and since the “iterative” Steps (2)–(7) of Algorithms 6 and 8 are the same, then the “convergence” of Algorithm 8 follows from Theorem 7 as well.

##### 1.2. Organization of the Paper

The rest of the paper is organized as follows. In Section 2, the concepts of well-posedness of optimization problems in the sense of Tychonov, Hadamard, and in a generalized sense, as well as calmness in the sense of Clarke, are discussed. It is shown in Section 2.3 that the convex separable optimization problems under consideration are calm in the sense of Clarke. In Section 3, the concept of stability of the set of saddle points of the Lagrangian in the sense of Gol’shtein is also discussed and it is shown that this set is not stable for the “classical” Lagrangian. However, it is explained that despite this instability, due to the specificity of the approach, suggested by the author in previous papers for solving problems under consideration, it is not necessary to use modified Lagrangians but only the “classical” Lagrangians. In Section 4, primal-dual analysis of the problems under consideration in view of methods for solving them is presented. Main results of well-posedness and primal-dual analysis are included in Section 2.3 and in Sections 3 and 4.

#### 2. Well-Posedness of Optimization Problems

Questions of existence of solutions and how they depend on problem’s parameters are usually important for many problems of mathematics, not only in optimization. The term *well-posedness* refers to the existence and uniqueness of a solution and its continuous behavior with respect to data perturbations, which is referred to as *stability*. In general, a problem is said to be *stable* if
where is a given tolerance of the problem’s data, is the accuracy with which the solution can be determined, and is a continuous function of . Besides these conditions, accompanying robustness properties in the convergence of sequence of approximate solutions are also required.

Problems which are not well-posed are called *ill-posed*, or, sometimes, *improperly posed*.

##### 2.1. Tychonov and Hadamard Well-Posedness: Well-Posedness in the Generalized Sense

Recall that is a *proper function* if for at least one and for all , or, in other words, if
is a nonempty set on which , where dom is the *effective domain* of . Otherwise, is *improper*.

*Definition 9. *Let be a space with either a topology or a convergence structure associated and let be a proper extended real-valued function. Consider the problem
The problem (21) is *Tychonov well-posed* if and only if has a unique global minimum point on towards which every minimizing sequence converges.

An equivalent definition is as follows: problem (21) is *Tychonov well-posed* if and only if there exists a unique such that for all and

There are two ways to cope with ill-posedness.

The first one is to change the statement of the problem.

The second one is the so-called *Tychonov regularization*. A parametric functional is constructed such that if it approaches 0, the solution of the “regularized” problem converges to the exact solution of the original problem.

Consider the problem

Associate the following problem with (23): where is perturbation in the input data and is an optimal solution to the perturbed problem.

Let

If when , then problem (23) is stable with respect to perturbation .

A parametric function with a parameter is called a *regularizing function* for problem (23) with respect to perturbation if the following conditions are satisfied.

(1) is defined for all and .

(2) If is an optimal solution to problem then there exists a function such that when .

Following Tychonov, an ill-posed problem is said to be *regularizable* if there exists at least one regularizing function for it.

The concept of Tychonov well-posedness can be extended to problems without the uniqueness of the optimal solution.

*Definition 10. *Let be a space with either a topology or a convergence structure associated, and be a proper real-valued function. Problem (21) is said to be *well-posed in the generalized sense* if and only if and every sequence such that has some subsequence with arg .

Problem (21) is Tychonov well-posed if and only if it is well-posed in the generalized sense and arg is a singleton.

* Hadamard well-posedness* is primarily connected with problems of mathematical physics (boundary value problems for partial differential equations) and can be extended to mathematical programming problems. We do not discuss this topic here.

As recent studies in the calculus of variations, optimal control, and numerical methods of optimization show, uniqueness and continuity are often too restrictive to be adopted as the standards of well-posedness. It turns out that practical concepts concerning well-posedness are some forms of semicontinuity in the problem’s data and solution mapping, along with potential multivaluedness in this mapping.

##### 2.2. Calmness in the Sense of Clarke

Let be a Banach space.

*Definition 11. *Let be a subset of . A function is said to satisfy a *Lipschitz condition* on provided that, for some nonnegative scalar , the following inequality holds true:
for all points ; this is also referred to as a Lipschitz condition of *rank *. We say that is *Lipschitz* (of rank ) *near * if for some , satisfies a Lipschitz condition (of rank ) on the set (i.e., within an -neighborhood of ), where is the open unit ball around .

A function , which satisfies a Lipschitz condition, sometimes is said to be *Lipschitz continuous*.

Consider the following general mathematical programming problem:

where , are real-valued functions on .

Let and be the functions , .

Let be imbedded in a parametrized family of mathematical programs, where , :

Denote by the feasible region of problem .

*Definition 12 (Clarke [3]). *The *value function * is defined via (i.e., the value of the problem ). If there are no feasible points for , then the infimum is over the empty set and is assigned the value .

*Definition 13 (Clarke [3]). *Let solve . The problem is *calm* at provided that there exist positive and such that for all , for all which are feasible for , one has
where is the open unit ball in and is the Euclidean norm of .

Let be an open convex subset of .

Theorem 14 (Roberts and Varberg [4], Clarke [3]; Lipschitz condition from boundedness of a convex function). *Let be a convex function, bounded above on a neighborhood of some point of . Then, for any in , is Lipschitz near . *

Recall that *limit superior* of a bounded sequence in , denoted or , equals the infimum of all numbers for which at most a finite number of elements of (strictly) exceed . Similarly, *limit inferior* of is given by at most a finite number of elements of are (strictly) less than .

A bounded sequence always has a unique limit superior and limit inferior.

Theorem 15 (Clarke [3], Calmness). *Let be finite and suppose that
**(this is true in particular if is Lipschitz near ). Then, for any solution to , problem is calm at . *

Sometimes problem is said to be *calm* provided satisfies the hypothesis of Theorem 15.

*Slater-Type Conditions.* Suppose that has no equality constraints (i.e., ), that the functions , , are convex, and that is a convex set. Recall that *Slater’s condition* (*Slater’s constraint qualification*), then is: there exists a point in such that , ( is called a *strictly feasible point*).

For , let be the infimum in the problem in which the constraints of problem are replaced by .

Theorem 16 (Clarke [3]; Lipschitz property of the value function from Slater’s condition). *If is bounded and is Lipschitz on , then Slater’s condition (i.e., the existence of a strictly feasible point) implies that is Lipschitz near . *

Theorems 15 and 16 mean that Slater’s constraint qualification implies calmness of problem in this case.

Theorem 17 (Clarke [3]; Calmness of a problem subject to inequality constraints). *Let incorporate only inequality constraints and the abstract constraint and suppose that the value function is finite for near . Then, for almost all in a neighborhood of , the problem is calm. *

*Remark 18. *In the case of problem , in which equality constraints exist, it is a consequence of Ekeland’s theorem that is calm for all in a dense subset of any open set upon which is bounded and lower semicontinuous.

Consider the following way of perturbing problem :

where is a vector of real components. The value function then would be a function of .

This is a special case of problem with , , , ,,. At least when the dependence of , and on is locally Lipschitz, we can consider problem with , ,,,, and rather than problem . Hence, the methods and results, considered above, can be applied to perturbed family as well.

Constraint qualifications (regularity conditions) can be classified into two categories: on the one hand, Mangasarian-Fromowitz and Slater-type conditions and their extensions, and, on the other hand, constraint qualifications called *calmness*. It turns out that calmness is the weakest of these conditions, since it is implied by all the others (see, e.g., Theorem 16).

##### 2.3. Well-Posedness of Problems , , and

###### 2.3.1. Existence of Solutions

The question of existence of solutions to problems , , and has been discussed in Theorems 1, 4, and 5, respectively. Steps (0), (1), and (9) of Algorithm 2 and Steps (1) and (9) of Algorithms 6 and 8, respectively, refer to these results.

###### 2.3.2. Uniqueness of Solution

The question of uniqueness of the optimal solution to problems under consideration is also important.

If defined by (1) (by (9), (12), resp.) is a strictly convex function, then problem (problem or problem , resp.) has a unique optimal solution in the feasible region , resp.) in case problem (problem or problem , resp.) has feasible solutions; that is, ,, are uniquely determined from (6) [(17)] in the interval in this case. If the parameters , and so forth of particular problems of the form ( and , resp.) are generated in intervals where the functions are strictly convex, then problem (problem or problem , resp.), if it has feasible solutions, has a *unique* optimal solution.

In the general case, if functions are convex but not necessarily strictly convex, then, as it is known, a convex programming problem has more than one optimal solution and the set of optimal solutions to such a problem is convex. Further, the optimal value of the objective function is the same for all optimal solutions to problem (problem or problem , resp.) if it has more than one optimal solution. If, for example, (6) ((17), resp.) is a linear equation of , then , are also uniquely determined from (6) (from (17), resp.).

###### 2.3.3. Calmness of the Problems (of the Optimal Solutions)

Let , and be the parametrized families of mathematical programs associated with problems ,, and , respectively.

Feasible regions of problems and are nonempty by the assumption; this is satisfied when and , respectively. Without loss of generality, feasible regions

: and : of problems and , respectively, are also nonempty in a neighborhood of .

Since the value function , associated with problems and , is finite near (according to Definition 12 and the assumption that the corresponding feasible set is nonempty) then both problems are calm according to Theorem 17.

*An alternative proof* of calmness of problem is the following.

The objective function of problem (1)–(3) is convex (and, therefore, Lipschitz in accordance with Theorem 14), and Slater’s constraint qualification is satisfied by the assumption. From Theorem 16, it follows that the value function is Lipschitz, and problem is calm at any solution of problem according to Theorem 15.

Consider the parametrized family in which problem is imbedded as follows:

where is defined as follows:

As it has been pointed out, if whereas if

Without loss of generality, assume that there exists a such that . This is satisfied, for example, when in addition to the requirement . Then the value function , associated with , is finite by Definition 12.

Theorem 19 (Convexity of the infimum of a convex function subject to linear equality constraints). *Let be a convex function and be a convex set in . Then, function
**
is convex. *

*Proof. *Let ,,,. Therefore as a convex combination of elements of the convex set . Then,
We have used that is a convex function, the property that
and the fact that implies .

Therefore, is a convex function by definition.

For problem , matrix of Theorem 19 consists of a single row, that is, , and convex set is the -dimensional parallelepiped The value function associated with problem is

From Theorem 19 and the assumption that , it follows that is convex and finite, respectively, and from Theorem 14 it follows that it is Lipschitz. Then, problem is calm according to Theorem 15.

In the general case, if the mathematical program is not convex and equality constraints exist, we can use the approach of the Remark 18.

#### 3. On the Stability of the Set of Saddle Points of the Lagrangian

##### 3.1. The Concept of Stability of Saddle Points of the Lagrangian

Besides well-posedness of the optimization problems, stability of methods for solving these problems is also important.

Let be a convex function of and a concave function of , where and are convex and closed sets.

Recall the definition of a saddle point.

A point is said to be a *saddle point* of function , , , if the following inequalities hold:
for all ,, that is, if

This means that attains at the saddle point its maximum with respect to for fixed and attains at its minimum with respect to for fixed .

Set

Denote by and the sets of optimal solutions to the optimization problems respectively, that is, Let , be bounded sets. Then,

that is, This means that is the set of saddle points of and Consider the sets that is, and denote the sets of arguments of with and , respectively, for which the value of is equal to its value at the saddle point.

In the general case, , ; that is, the sets , contain sets , , respectively.

*Definition 20. *If and , then the set of saddle points of is said to be *stable*.

If the set of saddle points of is stable, then from it follows that and from it follows that where is the distance from to the set . The implications written above mean that convergence of to with respect to and convergence of to with respect to implies convergence of sequence to the set of saddle points of .

The concept of stability, introduced by Definition 20, is important for constructing iterative gradient algorithms for finding saddle points of the Lagrangian associated with an optimization problem.

The set of saddle points of the Lagrangian associated with the problem
*is not stable* according to Definition 20. Concerning the dual variables this can be proved as follows.

Let th constraint of (60) be satisfied as an equality at , that is, for some , . Then, the Lagrangian of problem (59)-(60) does not depend on and therefore is satisfied for every . Hence, it is impossible to determine by using the relation .

In order to avoid this difficulty, so-called *modified Lagrangians* are used instead of the “classical” Lagrangian. Modified Lagrangians are usually *nonlinear* functions of and the set of their saddle points is stable and it coincides, under some assumptions, with the set of saddle points of the “classical” Lagrangian for the same problem. This is important to ensure convergence of iterative gradient algorithms (see, e.g., Gol’shtein [5]).

##### 3.2. About the Stability of the Set of Saddle Points for the Approach Considered in this Paper

Consider problem (problem and problem , resp.). Obviously, the Lagrange multiplier , associated with the constraint (2) ((10) and (13), resp.), is not involved in the equality
when , that is, when . For problem , (>0) is either determined *uniquely* from when or we set when (Algorithm 2). Although the set of saddle points of the Lagrangian , associated with problem (problem and problem , resp.) is not stable in the sense of Gol’shtein, the specificity of the approach suggested (the algorithms are not of gradient type and is determined *uniquely* in all cases for the three problems under consideration) overcomes this “weakness” of the classical Lagrangian. That is why it is not necessary to use modified Lagrangians for problems , , and .

On the one hand, we need a closed form expression of at Step (2) of the algorithms suggested. However, it is this feature of the algorithms that allows us to use classical Lagrangians instead of modified Lagrangians in the approach suggested. Moreover, the method for finding , and, therefore, for finding , in the corresponding problem , , and , is *exact* although it is an iterative method.

As it usually happens, the disadvantage in one aspect turns out to be an advantage in another aspect and vice versa.

All conclusions in this section have been drawn under the assumption that the objective function and the constraint function(s) of the three problems under consideration , , are *nondegenerate*, that is, , ; otherwise, the application of the Karush-Kuhn-Tucker theorem with differentiability is void of meaning.

Some optimality criteria for degenerate mathematical programs are given, for example, in the book of Karmanov [6].

#### 4. Primal-Dual Analysis

Some of the main characteristics of the approach, suggested for solving problems , , and , are following.

Since the method, proposed for problem , uses values of the first derivatives of functions ,, we can consider it as a *first-order method*. Also, this method is a *saddle point method* or, more precisely, a *dual variables saddle point method* because it is based on convergence with respect to the Lagrange multiplier (dual variable) associated with the single constraint (2).

At Step (2) of Algorithm 2, we use the expression of , calculated from the equality , where are determined from (6), . As it was proved, under the assumptions for problem , we can always determine from as an implicit function of . For example, when ,, are linear functions, the explicit expression of is always available for Algorithm 2. There are also many other examples of functions, for which it is possible to obtain closed form expressions of , and therefore, the suggested approach is applicable and gives good results.

Analogous commentary is valid for the method suggested for solving problem .

When the (optimal) Lagrange multiplier associated with (2) is known, then problem (1)–(3) can be replaced by the following separable convex optimization problem

The problem, dual to problem , is where

Problem can be considered similarly; and for it.

Thus, using the Lagrangian duality and Theorem 1 for problem (Theorem 4 for problem ) we have replaced the *multivariate* problem (problem ) of by the *single-variable* optimization problem for finding (, resp.).

Since Algorithm 8 is based on Theorem 5 and Algorithm 6, and since the “iterative” Steps (2)–(7) of Algorithms 6 and 8 are the same, then primal-dual analysis for problem is similar to that for problems and .

#### 5. Bibliographical Notes

The definition of Tychonov well-posedness is given by Tychonov in [7]. Tychonov and Hadamard well-posedness and well-posedness in the generalized sense are considered, for example, in the work of Cavazzuti and Morgan [8], in the book of Dontchev and Zolezzi [9], in works of Hadamard [10, 11], and so forth. Other questions regarding stability, ill-posed problems, and Tychonov regularization can be found in [12–19], and so forth. Sometimes Tychonov is written as Tykhonov, Tikhonov or Tychonoff in references. Well-posedness is also discussed in the book of Rockafellar and Wets [20].

Calmness in the sense of Clarke is proposed and studied in the works of Clarke [3, 21, 22]. Concerning Theorem 14, see also the paper of Roberts and Varberg [4].

Stability of the set of saddle points of Lagrangians is considered in the papers of Gol’shtein [5], Gol’shtein and Tret’iakov [23], and so forth.

Various convex separable optimization problems are considered and convergent polynomial algorithms for solving them are proposed in Stefanov [1, 2, 24–29].

#### References

- S. M. Stefanov, “Convex separable minimization subject to bounded variables,”
*Computational Optimization and Applications*, vol. 18, no. 1, pp. 27–48, 2001. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - S. M. Stefanov, “Convex separable minimization problems with a linear constraint and bounded variables,”
*International Journal of Mathematics and Mathematical Sciences*, no. 9, pp. 1339–1363, 2005. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - F. H. Clarke,
*Optimization and Nonsmooth Analysis*, vol. 5 of*Classics in Applied Mathematics*, SIAM, Philadelphia, Pa, USA, 1990. View at Publisher · View at Google Scholar · View at MathSciNet - A. W. Roberts and D. E. Varberg, “Another proof that convex functions are locally Lipschitz,”
*The American Mathematical Monthly*, vol. 81, pp. 1014–1016, 1974. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - E. G. Gol'shtein, “Generalized gradient method for finding saddle points,”
*Economics and Mathematical Methods*, vol. 8, no. 4, pp. 569–579, 1972 (Russian). View at Google Scholar - V. G. Karmanov,
*Mathematical Programming*, Mir, Moscow, Russia, 1989. View at MathSciNet - A. N. Tychonov, “On the stability of the functional optimization problems,”
*USSR Computational Mathematics and Mathematical Physics*, vol. 6, no. 4, pp. 631–634, 1966 (Russian). View at Google Scholar - E. Cavazzuti and J. Morgan, “Well-posed saddle point problems,” in
*Optimization: Theory and Algorithms*, J. B. Hiriart-Urruty, W. Oettli, and J. Stoer, Eds., vol. 86 of*Lecture Notes in Pure and Applied Mathematics*, pp. 61–76, Marcel Dekker, New York, NY, USA, 1983. View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - A. L. Dontchev and T. Zolezzi,
*Well-Posed Optimization Problems*, vol. 1543 of*Lecture Notes in Mathematics*, Springer, Berlin, Germany, 1993. View at MathSciNet - J. Hadamard, “Sur les problemes aux dérives partielles et leur signification physique,”
*Bulletin Princeton University*, vol. 13, pp. 49–52, 1902. View at Google Scholar - J. Hadamard,
*Lectures on Cauchy's Problem in Linear Partial Differential Equations*, Dover, New York, NY, USA, 1953. View at MathSciNet - A. N. Tychonov, “Stability of inverse problems,”
*Doklady Akademii Nauk SSSR*, vol. 39, no. 5, pp. 195–198, 1943 (Russian). View at Google Scholar - A. N. Tihonov, “On the solution of ill-posed problems and the method of regularization,”
*Doklady Akademii Nauk SSSR*, vol. 151, no. 3, pp. 501–504, 1963 (Russian). View at Google Scholar · View at MathSciNet - A. N. Tihonov, “On the regularization of ill-posed problems,”
*Doklady Akademii Nauk SSSR*, vol. 153, no. 1, pp. 49–52, 1963 (Russian). View at Google Scholar · View at MathSciNet - A. N. Tihonov, “Regularisation methods for optimal control problems,”
*Doklady Akademii Nauk SSSR*, vol. 162, no. 4, pp. 763–765, 1965 (Russian). View at Google Scholar · View at MathSciNet - A. N. Tihonov, “ll-posed problems of optimal planning and stable methods for solving them,”
*Doklady Akademii Nauk SSSR*, vol. 164, no. 3, pp. 507–510, 1965 (Russian). View at Google Scholar · View at MathSciNet - A. N. Tychonov, “Ill-posed optimal planning problems,”
*USSR Computational Mathematics and Mathematical Physics*, vol. 6, no. 1, pp. 81–89, 1966 (Russian). View at Google Scholar - A. N. Tychonov and V. Y. Arsenin,
*Solutions of Ill-Posed Problems*, John Wiley & Sons, New York, NY, USA, 1977. - A. N. Tychonov and V. Y. Arsenin,
*Methods for Solving Ill-Posed Problems*, Nauka, Moscow, Russia, 1986. - R. T. Rockafellar and R. J. B. Wets,
*Variational Analysis*, vol. 317 of*Fundamental Principles of Mathematical Sciences*, Springer, Berlin, Germany, 1998. View at Publisher · View at Google Scholar · View at MathSciNet - F. H. Clarke,
*Necessary conditions for nonsmooth problems in optimal control and the calculus of variations [Ph.D. thesis]*, University of Washington, Seattle, Wash, USA, 1973. - F. H. Clarke, “A new approach to Lagrange multipliers,”
*Mathematics of Operations Research*, vol. 1, no. 2, pp. 165–174, 1976. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - E. G. Gol'shtein and N. V. Tret'iakov, “Modified Lagrangians,”
*Economics and Mathematical Methods*, vol. 10, no. 3, pp. 568–591, 1974 (Russian). View at Google Scholar - S. M. Stefanov, “Method for solving a convex integer programming problem,”
*International Journal of Mathematics and Mathematical Sciences*, no. 44, pp. 2829–2834, 2003. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - S. M. Stefanov, “Convex quadratic minimization subject to a linear constraint and box constraints,”
*Applied Mathematics Research eXpress*, no. 1, pp. 17–42, 2004. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - S. M. Stefanov, “Polynomial algorithms for projecting a point onto a region defined by a linear constraint and box constraints in ${\mathbb{R}}^{n}$,”
*Journal of Applied Mathematics*, no. 5, pp. 409–431, 2004. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - S. M. Stefanov, “An efficient method for minimizing a convex separable logarithmic function subject to a convex inequality constraint or linear equality constraint,”
*Journal of Applied Mathematics and Decision Sciences*, vol. 2006, Article ID 89307, 19 pages, 2006. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - S. M. Stefanov, “Minimization of a convex linear-fractional separable function subject to a convex inequality constraint or linear equality constraint and bounds on the
variables,”
*Applied Mathematics Research eXpress*, vol. 2006, no. 4, Article ID 36581, 24 pages, 2006. View at Google Scholar - S. M. Stefanov, “Solution of some convex separable resource allocation and production planning problems with bounds on the variables,”
*Journal of Interdisciplinary Mathematics*, vol. 13, no. 5, pp. 541–569, 2010. View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet