Abstract
We propose a new method to solve nonlinear semidefinite complementarity problem by combining a continuous method and a trust-region-type method. At every iteration, we need to calculate a second-order cone subproblem. We show the well-definedness of the method. The global convergent result is established.
1. Introduction
This paper deals with the semidefinite complementarity problem (SDCP) with respect to a mapping , denoted by SDCP(), to find an such that where is a set comprising those that are real symmetric. SDCP is the generalization of linear complementarity problems (LCPs) and semidefinite programs (SDPs) which has wide applications in engineering and economics [1]. The study of this problem can be dated back to the work of Shibata et al. [2]. Since then much attention has been attracted to SDCPs and various reformulations of SDCPs to minimization problem based on merit functions have been presented [3–6]. In general, there are two ways to derive the global convergence of an algorithm: trust-region methods and line search methods. The above methods proposed for solving SDCP() are all based on a line search strategy; methods based on trust-region technique are relatively fewer. Despite having been studied by many researchers [7, 8], trust-region methods are robust, can be applied to ill-conditioned problems, and have strong global convergence properties. Therefore, different from the above methods, we propose a new algorithm based on trust-region method to solve SDCPs.
A function is said to be monotone if for any . An SDCP() is called a monotone SDCP() if the involved function is a monotone function. The Frobenius norm of a matrix is defined by Let be a linear operator satisfying then, is said to be Fréchet differentiable at and is the Fréchet derivative of at . The function is said to be differentiable if it is differentiable at each and to be continuously differentiable if also is continuous at each . In this paper, we suppose that is a continuously differentiable monotone function.
Recently, there has been much interest in SDCP(). A few methods have been developed to solve this problem, such as interior point methods, merit function methods, and noninterior point continuation/smoothing methods [3–5].
Our new algorithm is based on the following smoothed Fischer-Burmeister function: where and is the identity matrix. This smoothing function was introduced by Kanzow [9] in the case of the NCP based on the Fischer-Burmeister function. Let From Lemma 1 of [3], we know that if , then where denotes the orthogonal projection of at , whereas by Lemma of [4], Thus, we can solve SDCP() by using the following approach: reformulate SDCP() as a system of nonsmooth equation and then approximate nonsmooth equations by parameterized smooth equations ; we solve the smooth equations at each iteration and make decrease gradually by reducing the smoothing parameter to zero. In practice, however, it is usually impossible to solve the equation exactly for .
In this paper, we present a continuous and approximate method to solve SDCP(). At each iteration, the method solves a quadratic semidefinite program, which can be converted to a linear semidefinite program with a second-order cone constraint. A subproblem of this kind can be solved quite efficiently by using some recent software for semidefinite and second-order cone programs. The method is shown to be globally convergent under certain assumption.
The rest of this paper is organized as follows. Section 2 gives the algorithm and discusses the well-posedness for the algorithm; Section 3 analyzes the global convergence for the new algorithm; Section 4 presents the numerical results for the new algorithm; Section 5 concludes this paper.
2. The Algorithm
In this section, we will propose a smoothing trust-region-type algorithm for solving SDCP() and prove that the proposed algorithm is well-defined.
Define
We begin with a formal statement of the algorithm which is in the spirit of [10–12].
Algorithm 1 (continuous trust-region-type method). (S0)(Initialization) Choose , , , , , , , , , and set . (S1)Find the solution of the subproblem
If and , then STOP. (S2)Compute the ratio
If , then the th iteration is called successful, and we set ; otherwise, the th iteration is called unsuccessful, and we set . (S3)If
then set
and choose such that
otherwise, let and . (S4)Update as follows. If , set . If , set . If , set . (S5)Set , and go to (S1).
To verify that Algorithm 1 is well-defined, we need the following properties of the smoothed Fischer-Burmeister function (5).
Lemma 2 (see [1]). Let and be defined by (5). Then(i)if , is continuously differentiable at any ;(ii)for any and , it follows that
Lemma 3. Let be a given iterate and let be the solution of the corresponding subproblem (10). Then
Proof. Since , the symmetric matrix is feasible for the subproblem (10). But is a solution of this subproblem, so we obtain This proves our statement.
The above lemma ensures that the denominator in the ratio is always nonnegative. Note that this implies that the sequence is monotonically nonincreasing. We next show that this denominator is equal to zero if and only if the termination criterion in step (S1) is satisfied. Hence, step (S2) is visited only if the denominator is positive, so that Algorithm 1 is well-defined.
Lemma 4. Let be a given iterate and the solution of the corresponding subproblem. Then if and only if .
Proof. First assume that . Then since the definition of implies . Conversely, let . Lemma 3 then implies and hence .
Next we have to justify our termination criterion in step (S1). To this end, we will show that this criterion is satisfied if and only if the current iterate is a stationary point of .
Before we arrive at this result, we first take a closer look at subproblem. Let be a given iterate and let be the unique solution of this subproblem. Since this subproblem is a convex program with a strictly feasible set, this problem is equivalent to its KKT conditions. In other words, is a solution of subproblem if and only if there exist Lagrange multipliers such that the following KKT conditions hold: Now, if is the unique solution of this subproblem, then the system yields However, these conditions are nothing, but the KKT conditions for the following problem are
Summarizing these observations, we obtain the following result.
Theorem 5. Let . If is the (unique) solution of the subproblem for some , then is a stationary point of the original problem. Conversely, if is a stationary point of the original problem, then is the unique solution of subproblem for every .
Proof. The statements follow immediately from the preceding arguments.
3. Convergence Analysis
Throughout this section, we assume that Algorithm 1 generates an infinite sequence . Our aim is to establish a global convergence result for Algorithm 1. More precisely, we will show any accumulation point of is a stationary point of the original problem.
Lemma 6. Let be a sequence generated by Algorithm 1, and let and be subsequences converging to 0 and some matrix , respectively, in such a way that . Then is a stationary point of the original problem.
Proof. First note that is symmetric positive semidefinite and hence feasible for original problem. Furthermore, since for all , the assumption implies . By continuity, we also have as , . This together with system (18) implies that
as , . Therefore, taking the limit on the subsequence in the KKT condition (18), we obtain
Hence we conclude that is a stationary point of the original problem.
Another main step toward our global convergence result is contained in the following technical lemma.
Lemma 7. Let be a sequence generated by Algorithm 1 and a subsequence converging to some matrix . If is not a stationary point, then one has .
Proof. Let . Then we have . We will show that . Assume the contrary. Then, if necessary, we may suppose without loss of generality that
The updating rule in step (S3) then implies that none of the iterations with sufficiently large is successful since otherwise we would have for all these . Hence, we have
and for all large enough. Since , this implies , too. Further, noticing that for all unsuccessful iterations, we also have
because of (23). We now want to show that
which would then lead to the desired contradiction to (24). To this end, we first note that
In fact, if on a subsequence, we would deduce from Lemma 6 that is a stationary point in contradiction to our assumption. Hence there is a constant such that
By Lemma 2, this implies
for all sufficiently large.
We further note that . Otherwise, it would follow from (25) that on a suitable subsequence. This, in turn, would imply that the optimal value of the subproblem tends to infinity. However, this cannot be true since the feasible matrix would give a smaller objective value. Hence we have .
Taking this into account and using and the fact that is continuously differentiable, we obtain through standard calculus arguments
Summarizing these observations, we get
as , . This contradiction to (24) completes the proof.
As a direct consequence of this lemma, we obtain the following result.
Lemma 8. Let be sequence generated by Algorithm 1. Then there are infinitely many successful iterations.
Proof. If not, there would exist an index with and for all . This implies due to the updating rule in (S3). However, since is not a stationary point and , we get a contradiction to Lemma 6.
We are now in the position to prove the main convergence result for Algorithm 1.
Theorem 9. Let be a sequence generated by Algorithm 1. Then, any accumulation point of this sequence is a stationary point of the original problem.
Proof. Let be an accumulation point and subsequence converging to . Since for all unsuccessful iterations and since there are infinitely many successful iterations by Lemma 7, we may assume without loss of generality that all iterations are successful.
Assume that is not a solution. Lemma 6 then implies
Hence there is a constant such that
Since each iteration is successful, we also have . Consequently, we obtain from Lemma 2
for all . Since is monotonically nonincreasing and bounded from below by, for example, , we have as . Therefore, we obtain from (34). By (33), this also implies . But then Lemma 6 shows that is a solution in contradiction to our assumption. This completes the proof.
In the following, we will give one stronger global convergence result. Define the index set
Lemma 10. If is a function, then the sequence generated by Algorithm 1 remains in the level set
Proof. Let be an arbitrary nonnegative integer, and let be the largest number in such that , as is as defined from (35). It is easy to deduce from step 3 of Algorithm 1 that
Set
As is an arbitrary integer and , it follows that .
Next, by induction, we will prove
In view of Lemma 2, we deduce that ,
If , then by (40) we have
This proves .
Suppose for some . Then, and hence . Set
It follows from step 3 of Algorithm 1 and Lemma 2 that
or
This implies that
where . Moreover, we have
From (45) and (46), we deduce that, for ,
Combining (47) with (40), we have
which implies . Hence (39) is proved and Lemma 10 is valid.
It follows from Fischer [13] that if is a function or, more generally, an -function, then the level set as defined in Lemma 10 is compact. Lemma 3 shows that the sequence is monotonically decreasing and converges.
Theorem 11. Assume that is a function. Let be a sequence generated by Algorithm 1. If there exists at least an accumulation point in the sequence , then the index set defined by (35) is infinite:
Proof. We first prove that set is infinite. By contradiction, assume that is finite. Let be the largest number in . Then for all , and . Denote
Then ,
From Theorem 9, it follows that there exists at least an accumulation point of such that
Next, assume that subsequence converges to . In view of (53), we have and hence there exists such that, for all with ,
This together with (51) and (52) shows that, for all with ,
that is,
which means
This contradicts (51). Hence the set is infinite.
Next, follows immediately from the updating rule of and the fact that the set is infinite. Moreover, by the proof of Lemma 10, we deduce
Because the set is infinite, it follows from Lemma 10 and (58) that
This completes the proof.
As a consequence of the above theorem, we get the following global convergence result.
Corollary 12. Assume that is a function. Let be a sequence generated by Algorithm 1. Then every accumulation point of the sequence is a solution of NCP().
4. Numerical Experiments
4.1. The Reformulation for Subproblem
To test the numerical performance of Algorithm 1, we implemented the method in MATLAB (Version 7.0) using the SDPT3-Solver (Version 3.0) for the corresponding subproblems. First, we will give the reformulation of the subproblem. In order to solve nonlinear semidefinite programs of the form (1) by Algorithm 1, we have to be able to deal with a subproblem given by For this purpose, we would like to use the SDPT3-Solver (version 3.0) from [14]. This software is designed to solve linear semidefinite programs with cone constraints of the form where , are symmetric matrices of dimension ; , are vectors in ; denotes the -dimension positive semidefinite cone defined by ; denotes the -dimensional second-order cone defined by ; and are vectors in ; are matrices with ; and are and matrices, respectively; and is the operator defined by for any symmetric matrix .
We now want to rewrite the problem (60) in the form of (61). To this end, we need to make some reformulations, which will be described step by step in the following.
First, we drop the constant from the objective function without affecting the problem. Next, we introduce the auxiliary variable and set . Because needs only to be symmetric and not to be positive semidefinite, we set and write the problem in terms of with . Then problem (60) is equivalent to By introducing the second-order cone constraint , the above problem can be further rewritten as
Unfortunately, the term is not linear as required in (61). So we replace by the new variable and add the constraint . But this constraint can be rewritten as the semidefinite constraint Introducing once again an auxiliary variable, problem (62) and hence the original subproblem (60) are equivalent to We write the equality constraint in the svec-notation. Then we get
We are now in a position to give the explicit correspondence between the parameters, variables, and input data in our last problem formulation (67) and those in the SDPT3 standard form. The problem parameters are given by The variables are given by The input datum in the objective function is given by with Finally, the matrices , , , and and the vector are given by This is the desired reformulation.
It may be worth mentioning that problem (60) can also be transformed as Since the constraint is equivalent to problem (73) can further be reformulated as a linear semidefinite program that involves a semidefinite cone constraint instead of a second-order cone constraint. However, such a semidefinite representation is much more expensive in terms of memory requirement. Therefore, we adopted the reformulation (73) in our numerical experiments.
4.2. Numerical Results
We present some numerical tests using Algorithm 1. All the codes are written in MATLAB 7.10. The tests are conducted on a DELL computer with Intel(R)Core(TM)i5-2400 processor (3.10 GHz) and 4.00 GB of memory on Windows 7.
Consider the following nonlinear semidefinite complementarity problem with . It is obvious that the solution set of is nonempty, since is its one solution. The parameters in the algorithm can be presented as follows: and can be randomly generated from and , respectively; and are randomly generated from and , respectively; and are randomly generated from and , respectively; is randomly chosen from ; is randomly generated from and , where with every entry being randomly generated from . The stopping criterion is set as and .
For the purpose of comparison, we also solve this problem by the following descent algorithm based on the method proposed in [15].
Algorithm 13 (decent direction method). (S0)(Initialization) Choose , , , , , , , and set . (S1)Find the solution of the subproblem If , then STOP. (S2)Compute such that (S3)Let . If then set and choose such that otherwise, let and . (S4)Set , and go to (S1).
The above descent algorithm uses different NCP function from the one used in [15]. The parameters in this algorithm are set as follows: can be randomly generated from ; ; the starting point and the stopping criteria are the same as Algorithm 1.
We now solve this problem 40 times by Algorithms 1 and 13, respectively, with the initial point being randomly generated as above. Table 1 lists the numerical results for the applications of Algorithms 1 and 13. The average number of iterations and the average computational time (CPU time) are reported in Table 1. The results generally show that our method is efficient in solving this problem.
5. Conclusion
In this paper, we propose a trust-region method to solve nonlinear semidefinite complementarity problem. The well-posedness of the new method is proved and the global convergence is also presented. The numerical comparisons with the descent algorithm show the efficiency of the proposed method. For further study, we will discuss the convergent rate of the algorithm.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work was supported by the Natural Scientific Foundation of China (no. 71201040 and no. 11201099) and by the Major Program of the National Natural Science Foundation of China (no. 71031003).