A Modified Lagrange Method for Solving Convex Quadratic Optimization Problems ()
1. Introduction
Whenever a viable decision is made by selecting from a variety of alternatives which can be evaluated by some performance measure for quality’s sake, and there are some restrictions involving the alternatives, there is an optimization problem. An optimization problem is therefore characterized by a set of decision alternatives called decision variables, at least a performance measure often called objective function (or objective functions), and often restrictions involving the decision variables called constraints [1] . The occurrence of optimization problems is an everyday phenomenon and cuts across diverse disciplines, including engineering, computer science, applied mathematics, economics, and management, to name a few. Interest in the study of optimization has grown immensely from its early days going back to those of Sir Isaac Newton, and given new impetus in the 1940s by the seminal work of George B. Dantzig [2] in linear optimization. Today it attracts many researchers and practitioners from diverse fields and backgrounds, including engineers, systems analysts, operations researchers, numerical analysts, and management scientists.
Optimization problems are broadly classified as either linear or nonlinear, constrained or unconstrained [3] . Among constrained nonlinear optimization problems is a special subclass known as convex optimization with convex quadratic optimization problems being special subset of this subclass. Convex optimization problems in general have gained a lot of interest over the years, with applications discovered in many fields including computer science, robotics and signal processing [4] . Convex quadratic optimization problems also called convex quadratic programming problems (which are the focus in this study) are characterized by either linear or convex quadratic objective function and linear or convex quadratic constraints. A convex quadratic programming problem is a type of nonlinear programming. A typical form of the problem (see [5] ) is:
(1.1)
where x is a vector of n decision variables, c is n dimensional constant vector, Q and A are respectively n × n symmetric positive semi-definite and m × n constant matrices, and b is m dimensional constant vector [6] . If the objective function is linear and constraints a mix of linear, nonlinear, or convex quadratic equations, the problem becomes:
(1.2)
where,
is linear or nonlinear, convex quadratic, for some
.
1.1. The Classical Lagrange Method
The Lagrange Multiplier Method, referred to also as Classical Lagrange Method (CLM) is well-known for solving equality constrained nonlinear optimization problems (Hoffman et al., 1992). The main idea of the method is to convert an equality constrained optimization problem into an unconstrained one, by aggregating the objective function and constraint functions into a composite function called the Lagrangian function, to which the first derivative test for optimality is applied [7] . The Lagrangian function can be seen as constituting a relationship between the objective function and the equality constraints which simply gives a reformulation of the original problem as unconstrained one [8] . For convex quadratic programming problems, the objective function and constraint functions are convex and so the Lagrangian function is convex [9] . Therefore, for the sake of this work, we can limit the discussion of the CLM to convex problems only.
Consider a convex optimization problem with equality constraints as:
where f and
are respectively convex quadratic objective function and convex ith constraint functions. By letting
, the problem may be posed as:
(1.3)
The problem (1.3) is transformed into an unconstrained Lagrangian function L as:
(1.4)
where
is an n vector of constraint functions and λ is an m vector of multipliers, which are scalars (called Lagrange Multipliers).
The first order necessary optimality conditions for the Lagrangian function, which are also sufficient for a minimum solution for convex unconstrained optimization problems [10] are:
(1.5a)
(1.5b)
Equations (1.5a) and (1.5b) imply that the gradients of the Lagrangian function with respect to x and λ independently vanish at the minimum points (solutions)
and
. Furthermore, it means that if minimum solutions
and
exist to (1.4), they satisfy (1.5a) and (1.5b). However, for non-convex problems not all minimum solutions satisfying the first order necessary optimality conditions would be minimum solutions [10] ; such problems require that sufficient optimality conditions are also satisfied [10] . Observe also that (1.5a) and (1.5b) have n + m unknown variables (i.e., n variables in x and m variables in λ) whose values have to be determined.
To find the minimum values
,
the m + n system of equations resulting from (1.5a) and (1.5b) are solved simultaneously in the CLM. The larger m and n are, the more computational effort and time are required to solve for the minimum values. Therefore, finding a method that reduces the computational effort and time demands for finding the solutions is interesting and important, since that can have consequent positive impact on solving many real-world convex (quadratic) problems which occur in many disciplines [4] . This, indeed, is the motivation behind the pursuit of a Modified Lagrange Method (MLM).
1.2. Some Previous Works
A few related previous works in the research area are by [11] who developed a Lagrange relaxation-based decomposition algorithm for an integrated off-shore oil production planning optimization problem. The algorithm was used to provide compact bounds and thus replace the large-scale problem with moderate scale ones and therefore solve several single-batch subproblems simultaneously. An Adaptive Augmented Lagrangian method by [12] was developed for large scale constrained optimization problems. They used a novel adaptive updating technique for the penalty parameters which was motivated by techniques for exact penalty methods. A modified augmented Lagrangian multiplier method was proposed by [13] who used an outer iteration approach in which the Lagrangian multipliers and various penalty parameters were updated and used in the iteration involving constrained problems with simple bounds. Earlier, a class of Augmented Lagrangian methods for constrained optimization problems had been developed (see [14] [15] [16] ) which are similar to the Penalty methods and replace a constrained problem with a series of unconstrained ones and add a penalty term to the objective function. Komzsik and Chiang [17] investigated the effects of the Lagrange multiplier method in solving large-scale problems in a parallel environment.
In the next section, the philosophy and methodology behind the MLM is presented and discussed. The subsequent section assesses the MLM, in this first instance, in terms of its ability to produce solutions for convex quadratic optimization problems as does the CLM, using selected test problems. The paper is concluded with a section that elucidates the findings of the study and outlines prospects and consequences of the MLM for further research.
2. The Modified Lagrange Method
The Classical Lagrange Method (CLM) is modified in this section with the intention to improve it. The modification centers on the evolution of a new approach premised on departure from the classical and conventional thinking to finding solution of a convex quadratic nonlinear equality constrained optimization problem. While the CLM solves directly the two systems of Equations (1.5a) and (1.5b) resulting from the application of the first order necessary optimality conditions to (1.4), the Modified Lagrange Method (MLM) avoids this path, and rather uses only one of the two sets of Equations (i.e., (1.5a)) to develop a new approach to solving the problem. This is considered a major innovation that simplifies the solution process. The rest of the section develops the concept behind the MLM.
Consider the convex quadratic optimization problem given by (1.3). A scalar form of the Lagrangian function (1.4) which converts the problem (1.3) into an equivalent unconstrained problem and which we seek to minimize over x only, is:
(2.1)
It is clear that if there exist some
for which the second term of Equation (2.1) vanishes for all
,
, then the constraint equations of (1.3) are satisfied. At such a point, the function
and
become coincident, regardless of λ, and so
is a critical vector value that minimizes the function
and therefore
. In other words, the Lagrangian function is always equal to the objective function for some critical point
for which
, for all i. It turns out, under the CLM, that such a point,
, is the desired optimal (or minimum) solution of (2.1) or (1.3). This observation provides four (4) important facts that serve as bases for the evolution of the MLM. They are as follows: 1) At
,
for all i, in spite of λ; 2) This (i.e., (1)) indicate that we may seek to find
independently of λ; 3) Since
identically for all i and does not vary from zero at
, it follows that
for all j; and 4) By independently finding
we may independently find
the optimal value of λ. These four observations derived from the CLM means that we can instead find the optimal solutions
of problem (1.3) and problem (2.1) by two independent processes which first finds
and subsequently finds
. This, in essence, is the philosophy or conceptual framework of the new method called MLM.
By the CLM, the first order necessary optimality conditions (in scalar forms) of (2.1) which are also sufficient for the minimum solutions
are (2.2a)
(2.2a)
(2.2b)
It is noted that (2.2a) involves n equations in n unknown variables of x and m unknown variables of λ. Therefore, if we are able to eliminate
from (2.2a), then we can solve (2.2a) independently of (2.2b). secondly, we observe on the other hand, that (2.2b) cannot be solved independently of (2.2a), since it has m equations in the n unknown variables of x, and for nontrivial cases of the problem (1.3),
. Thirdly, it is important to note that the solution
which satisfies (2.2a) would not violate (2.2b), since the first and second terms of (2.2b) vanish for all
at
.
2.1. A Novel Process for Finding
With the observations made, therefore, we proceed to eliminate
(
) by some means in (2.2b), so that we can obtain equations involving only the n unknown variables of x. Consider the system (2.2a) in the expanded form given by the Equation (2.3).
(2.3)
As was noted earlier,
and
both vanish at
for any i and j, in spite of
. therefore
can be viewed as arbitrary for all i, as far as finding
from (2.2a) and therefore from (2.3) is concerned. This means we can choose
arbitrarily in our quest to solve (2.2a) independently of (2.2b). By choosing
for some
and
for all
,
, we can obtain the result in (2.4).
(2.4)
From (2.4), we can eliminate
by taking ratios of the jth and the
equation
, leading to:
which is generalized as (2.5):
(2.5)
The result in (2.5) leads to (2.6):
(2.6)
The results in (2.6) produces for each i, n − 1 equations, which we refer to as Subsidiary Constraint Equations (SCE). The SCE become additional available equations to the m original equations given by
of the problem (1.3) for finding
. Together, they provide
equations in x only and therefore simpler to solve. They are generally of sufficient number for finding
. Occasionally, through numerical experimental work, the authors have observed that some of the n − 1 number of SCE produced from (2.6) may be redundant equations, and depending on the number which are redundant, there may not be sufficient number of equations for finding
. To deal with this phenomenon, the authors have further observed that apart from taking ratios consecutively as in (2.5) for any
, the ratios can instead be taken in any arbitrary order, which means that there can be a large number (innumerable) possible forms of the SCE that can be constructed. For the purpose of this work, we produce another form of ordering of the ratios that leads to other forms of the SCE that appears to avoid the redundancy phenomenon, at least for the current work. The ratios are given by (2.7):
(2.7)
where
,
. The result in (2.7) is generalized as (2.8).
(2.8)
It follows from (2.8) that:
(2.9)
The authors have observed further (through numerical experimental work) that taking aggregates of the set of SCE obtained from (2.9) is even more effective at avoiding the redundancy phenomenon. Therefore, summing from the kth SCE of (2.9) (
), involving the equations
, we obtain:
Which is the same as:
(2.10)
It is noted that various sums, involving two (2) or more equations, may be obtained from (2.9) or (2.10), indicating further the variety of the forms of aggregates of the SCE that may be created from (2.9).
The following observations are made about the SCE in general: 1) They are devoid of the unknown parameter
and produce equations in the unknown variables of x only; 2) They are derived from the first order necessary optimality condition given by (2.2a) only, which involves derivatives that are easy to compute, due to the convex nature of the functions: and 3) They are derived under the implicit assumption of the existence of the minimum solution. Therefore, they embody the minimum point
. The resulting equations, therefore, in conjunction with the original constraint equations of the problem can independently be used to find the minimum solution
. The corresponding minimum objective function value
follow immediately, therefore.
2.2. Finding the Parameter λ
The optimal value of λ can also be obtained independently, from (2.2a) which is given by:
By obtaining the required derivatives of f and
, and evaluating them at
, we obtain m equations involving
only,
, which, therefore, can be solved independently for
.
2.3. The Solution Steps of the MLM
The solution process of the MLM is characterized by the following five (5) steps:
Step 1: Given problem (1.3), construct the Lagrangian function as in (2.1) and go to Step 2;
Step 2: Construct the first order necessary optimality condition as in (2.2a), and construct (2.4), then go to Step 3;
Step 3: Generate SCE as in (2.6), (2.9) or (2.10) and go to Step 4;
Determine whether or not the set of SCE together with the original problem constraints are sufficient to find
. If yes, then solve for
using any suitable numerical algorithm, otherwise, form alternative SCE using other forms of ratios or aggregates of equations such as 2.10 to solve for
; then go to Step 5.
Step 5: Substitute
in (2.2a) and solve for
, then stop.
We conclude the presentation of the MLM with the observation that: unlike the CLM, it does not require the first order necessary optimality condition given by (2.2b) in the solution of a problem (as given by (1.3)). Nevertheless, the solution
obtained would not violate (2.2b). This, in addition to the novel procedures for finding
and
are major simplifications in the process of solving a given convex quadratic optimization problem. The next section provides numerical results obtained by only hand computations involving small sized problems as a first stage evaluation of the MLM. In a subsequent work to be given in another paper, the new method would be evaluated on fairly large-scale problems and its relative performance against the CLM would be assessed to further demonstrate its capabilities as a viable method for solving equality constrained convex quadratic optimization problems.
3. Numerical Results
The new method (MLM), is applied to selected convex quadratic optimization problems to show that it produces the required solutions as the CLM. In this first stage of testing the method, we use small sized problems for which manual (or hand) computation is sufficient to find the solution. In the subsequent paper, where the method would be assessed against the CLM in terms of their relative performances in solving larger sized problems, the use of software would be necessary.
Seven problems, selected from literature for which solutions by the CLM have been obtained already, are solved with the MLM. The solution of each problem by the MLM is obtained for several cases of the SCE generated using (2.6) and (2.10) together with the original problem constraints, to demonstrate the variety of ways of solving the problems using the MLM.
3.1. Problem 1
This problem is extracted from Schittkowski [18] , given by:
(3.1a)
Subject to:
(3.1b)
(3.1c)
The solution by the CLM (see [18] ) is:
,
,
,
Solution by the MLM
Case 1: Solving by (2.6) and from (3.1a) and (3.1b), we obtain the SCE given by (3.1c) to (3.1f).
(3.1c)
(3.1d)
(3.1e)
(3.1f)
Notice that the equations given by (3.1d) and (3.1f) make (3.1e) redundant. However, substituting (3.1d) in (3.1c) gives
. By substituting
, (3.1d), and (3.1f) in the constraint Equations (3.1b) and (3.1c) and multiplying the resulting expression from (3.1b) by and adding that to the resulting equation from (3.1c) yields
and therefore
. Substituting
,
into (3.1b) we obtain
which gives the solution
, satisfying both constraint equations
and
. The objective function value is obtained as
. The minimum values of the multipliers
and
are obtained by solving the system of Equations (3.1g) given by:
(3.1g)
Substituting the values of
into (3.1g) produces the results
. Hence the minimum solution for Problem 1 using the MLM is
,
and
.
Case 2: Alternatively, we use aggregates of SCE formed as in (2.9) or (2.10) to generate equations and use in conjunction with the constraint equations to solve Problem 1, as follows:
Subtracting
from
, the result is
. Substituting
in
, the result is
. Substituting
in (3.1b), the resulting equation is
. Substituting
in
, the resulting equation is
. Solving
and
results in
.
Since the value of
satisfy both constraints (3.1b) and (3.1c), the objective function value is 0 as obtained previously. The values of
follow accordingly from (3.1g).
Case 3: The required number of SCE in this case are generated with respect to f and
using (2.6) leading to the following:
,
,
.
The resulting SCE generated above, together with the constraint Equations (3.1b) and (3.1c) fail to be of sufficient number to enable solving Problem 1, due to some redundant equations obtained. Therefore, we proceed to Case 4, where aggregates of the SCE are constructed to enable solving of Problem 1.
Case 4: The aggregates of SCE produced according to (2.10), yield the following equations:
The solution process of the MLM.
Substituting
in (3.1b) and multiplying the resulting equation by 2 and adding that to (3.1c) gives
. Substituting
in
results in
and so
. Substituting
in (3.1c) gives
. Substituting
in
gives
. Solving
and
simultaneously results in
. Therefore, we have
as was obtained in earlier cases. The rest of the results follow therefore.
3.2. Problem 2
This is an extract from [19] and given by:
(3.2a)
subject to
(3.2b)
(3.2c)
The solution obtained by CLM as in [19] , is:
and
Solution of Problem 2 by the MLM
Case 1: The nonredundant SCE produced according to (2.6) and using
yield:
,
,
Substituting
and
in (3.2b) results in
. From
and
we obtain
and
. The solution is therefore
,
,
. The solution
,
and
satisfy both constraints. The objective function value is
. We solve for
and
as follows:
This system of equations yields the solutions
,
.
Case 2: Aggregates of SCE this time is produced according to (2.10) involving the constraint
as follows:
Solving
simultaneously with (3.2b) and (3.2c) results in
,
,
which is the required solution. The corresponding values of the parameters
and
and the objective function follow accordingly. Alternatively, we can obtain the same solution by solving simultaneously
, (3.2b) and (3.2c).
Case 3: In this case, SCE are produced using (2.6), but this time using
, leading to the following equations:
,
,
Substituting
and
in (3.2c), the result is
. From
and
, we obtain
and
. Since the values do not satisfy both constraints, they are discarded. Alternatively, substituting
in both constraints and solving the resulting equations simultaneously results in
,
. Substituting these values in (3.2c) gives
.
The values
satisfy both constraint and yield corresponding objective function value calculated as
. This value is higher than the value
obtained earlier; and so, the current solution, though feasible, fails to be the minimum solution.
Case 4: In this case, aggregates of the SCE formed according to (2.10) and using
produce the following equations:
Solving (3.2b) and (3.2c) simultaneously with the first SCE results in the solution
,
,
, which satisfy both constraints. The corresponding objective function value is
. Alternatively, solving (3.2b), (3.2c) and the second equation simultaneously, gives
,
,
, which satisfy both constraints. The corresponding objective function value is
, which is marginally different from the value of 9 obtained previously.
The results obtained under this case reveal interesting outcomes that require further investigation to understand the phenomena responsible for this. We observe, nevertheless, that the solution (2.1154, 1.1538, 1.8077) compares well with (2, 1, 2) which is the required; the objective function value of 9.0381 also compares well with 9, the required minimum value.
3.3. Problem 3
The problem is taken from [20] and given by:
(3.3a)
subject to
(3.3b)
(3.3c)
The solution by the CLM is:
Solution of Problem 3 by the MLM
Case 1: SCE according to (2.6) are generated involving
as follows:
,
,
.
The equations
and
. Substituting
in (3.3b) and (3.3c) gives the equations
and
. Adding them gives
. Also, adding
to
and subtracting the result from (3.3c) yields
. Putting
in
gives
, and putting the results of
and
in
gives
. Finally, substituting
,
and
in (3.3b) or (3.3c) results in
. The values
,
,
,
satisfy both constraints (3.3b) and (3.3c), and the corresponding objective function value is
. The values of
,
are obtained from the equation
which is the system:
(3.3d)
Solving Equation (3.3d) produces the result
,
.
Case 2: The SCE, this time, are constructed according to (2.10) involving the constraint
, leading to the following:
(3.3e)
(3.3f)
(3.3g)
(3.3h)
Solving Equations (3.3b) (3.3c), (3.3e) and (3.3f) simultaneously, we have
,
,
,
, which satisfy the constraints (3.3b) and (3.3c). The corresponding objective function value is
. Comparing the objective function values of 27 (obtained in Case 1), we see that 27 is the least. Therefore, the minimum solution for Problem 3 is
,
,
,
,
,
,
.
Case 3: In this case, SCE are produced using (2.6) in conjunction with the constraint
as follows:
,
,
,
From
and
. Substituting
in (3.3b) and (3.3c), gives the equations
and
respectively. Adding
to
, results in
. Adding
to
results in
,
. Therefore,
, since
. Putting
,
,
in (3.3c) give
. The values
,
,
,
satisfy both constraints (3.3b) and (3.3c) with corresponding objective function value of 36. This value, however, is higher than 27 obtained under Cases 1 and 2; therefore, the solution while feasible, is not optimal.
3.4. Problem 4
This problem is taken from [18] and given by:
subject to:
(3.4a)
(3.4b)
(3.4c)
The solution of the problem by CLM is:
and
(Schittkowski, 2009).
Solution of Problem 4 by the MLM
Case 1: The SCE according to (2.6) and involving the constraint
are:
(3.4d)
, (3.4e)
Solving (3.4a), (3.4b), (3.4c), (3.4d) and (3.4e) simultaneously results in
, which satisfy all the three constraints with corresponding objective function value
. The first order necessary optimality conditions lead to the system:
(3.4f)
Solving the system of Equations (3.4f) results in
,
,
.
Case 2: The SCE this time are produced according to (2.10) as:
(3.4g)
(3.4h)
(3.4i)
(3.4j)
(3.4k)
Multiplying (3.4g) by 2 and subtracting the resulting equation from (3.4h) gives
. Subtracting (2.4h) from (3.4i) results in
. Solving simultaneously the equations
and
yields
. Substituting
in (3.4c) gives
. Substituting
in (3.4b) yields
. Substituting
in (3.4a) gives
. Therefore, the minimum solution is
with minimum objective function value of zero.
Case 3: The SCE in this case are produced according to (2.6), but this time involving the constraint
as follows:
, (3.4l)
, (3.4m)
.
Solving (3.4a), (3.4b), (3.4c), (3.4l), and (3.4m) simultaneously results in
which is the required minimum solution, with objective function value 0.
Case 4: The SCE are produced using (2.6) involving
this time around, as:
, (3.4n)
, (3.4o)
. (3.4p)
Solving constraint (3.4a) and Equation (3.4n) simultaneously results in
. Substituting
in constraint (3.4b) results in
. Substituting
in
and
in
gives
and
respectively. Therefore,
is the required minimum solution, with objective function value zero as previously. The parameter values obtained previously therefore follow accordingly.
Case 5: The SCE in this case are generated according to (2.10), involving the constraint
, as follows:
(3.4q)
(3.4r)
(3.4s)
(3.4t)
Solving the constraint Equation (3.4a) and Equation (3.4q) simultaneously yields
. Substituting
in (3.4c) results in
. Substituting (3.4t) gives
. Substituting
in (3.4t) results in
, which is what was obtained in the earlier cases of Problem 4. The objective function and parameter values, therefore, follow accordingly.
3.5. Problem 5
This is an extract from [20] . In this case we have a linear objective function being minimized subject to a nonlinear convex quadratic constraint equation.
subject to:
The solution obtained by the CLM is:
Solution of Problem 5 by the MLM
Case 1: The only SCE produced for Problem 5 using (2.6) and involving the only constraint
is:
.
Substituting
in
the result is
. Substituting
and
in
results in
and
respectively. The objective function values at (
,
) and (
,
) are respectively
and
. Hence, the minimum value is
and attained at (
,
). From the first order optimality condition, we have the system:
which produces the result
.
3.6. Problem 6
This problem is an extract from [21] . Like Problem 5, also involves minimizing a linear objective function, subject to a nonlinear convex quadratic constraint, given by:
subject to:
The solution (see [21] ) obtained by the CLM is:
Solution of Problem 6 by the MLM
Case 1: The only SCE obtained from (2.6) for this problem is:
Substituting
in
results in
or
. Substituting
and
, the objective function value is 2. Also, substituting
,
, the objective function value is −2. Therefore, minimum objective function value is −2 and attained at
,
. The first order optimality condition leads to the system:
Substituting
,
results in
.
3.7. Problem 7
In this problem, we consider minimizing a convex function, subject to a linear constraint. The problem is taken from [21] .
subject to:
Solution by the CLM is:
Solution of Problem 7 by the MLM
Case 1: The only SCE obtained from (2.6) is:
Substituting
in
, results in
. Substituting
in
, we obtain
. The minimum objective function value is therefore 25, attained at
. The first order optimality condition leads to the system:
which gives
.
4. Conclusions
This paper has reported a new way of solving an equality constrained convex quadratic optimization problems, based on what the authors have called Modified Lagrange Method (MLM), which, unlike the classical Lagrange Method (CLM), decomposes the task of solving for the unknown decision variable values of the problem and the values of the Lagrange multipliers by two independent procedures. The construction of what has been called Subsidiary Constraint Equations (SCE) by the authors, which result from a new procedure for independent solution for the decision variables is the novelty of the MLM. The arbitrarily large variety of ways of constructing the SCE is noted to be interesting and remains an area of further research.
The seven problems solved in this paper were intended to obtain preliminary results to demonstrate the ability of the MLM to find the required optimal solutions just as can be obtained from the CLM. In future works, the authors would demonstrate the superior performance of the MLM over the CLM in solving convex quadratic equality constrained problems and extend the approach to inequality constrained nonlinear convex quadratic problems as well as to convex problems not necessarily quadratic.