On how to solve large-scale log-determinant optimization problems Chengjing Wang January 25, 2014
Transcription
On how to solve large-scale log-determinant optimization problems Chengjing Wang January 25, 2014
On how to solve large-scale log-determinant optimization problems Chengjing Wang ∗ January 25, 2014 Abstract We propose a proximal augmented Lagrangian method and a hybrid method, i.e., employing the proximal augmented Lagrangian method to generate a good initial point and then employing the Newton-CG augmented Lagrangian method to get a highly accurate solution, to solve large-scale nonlinear semidefinite programming problems whose objective functions are a sum of a convex quadratic function and a log-determinant term. We demonstrate that the algorithms can supply a high quality solution efficiently even for some ill-conditioned problems. Keywords: Quadratic programming, Log-determinant optimization problem, Proximal augmented Lagrangian method, Augmented Lagrangian method, Newton-CG method 1 Introduction In this paper, by defining log 0 := −∞, we consider the following standard primal and dual nonlinear semidefinite programming problems whose objective functions are a sum of a convex quadratic function and a log-determinant term (QP-Logdet) : (P ) 1 min{ ⟨X, Q(X)⟩ + ⟨C, X⟩ − µ log det X : A(X) = b, X ≽ 0}, X 2 (D) 1 max {− ⟨X, Q(X)⟩ + bT y + µ log det Z + nµ(1 − log µ) : X,y,Z 2 −Q(X) + A∗ y + Z = C, Z ≽ 0}, ∗ School of Mathematics and Institute of Applied Mathematics, Southwest Jiaotong University, No.999, Xian Road, Chengdu 611756, China (ryan101879@gmail.com). The author’s research was supported by the National Natural Science Foundation of China under grant 11201382, the Youth Fund of Humanities and Social Sciences of the Ministry of Education under grant 12YJC910008, the project of the science and technology department of Sichuan province under grant 2012ZR0154, and the Fundamental Research Funds for the Central Universities under grants SWJTU12CX055 and SWJTU12ZT15. 1 where Q : S n → S n is a given self-adjoint positive semidefinite linear operator, C ∈ S n , b ∈ Rm , µ ≥ 0 is a given parameter, A : S n → Rm is a linear map and A∗ : Rm → S n is the adjoint of A. Note that the linear maps A and A∗ can be expressed, respectively, as [ A(X) = ⟨A1 , X⟩, . . . , ⟨Am , X⟩ ]T , ∗ A (y) = m ∑ yk Ak , (1) k=1 where Ak , k = 1, · · · , m are given matrices in S n . As for the explanation of all other main notations one may see Subsection 1.1. The perturbed QP-Logdet problem has the form {1 } (P ε ) min ⟨X, Q(X)⟩ + ⟨C, X⟩ − µ log det X : A(X) = b, X ≽ εI . X 2 (Dε ) 1 max {− ⟨X, Q(X)⟩ + bT y + ⟨Z, εI⟩ + µ log det(Q(X) + C − A∗ y − Z) X,y,Z 2 +nµ(1 − log µ) : −Q(X) + A∗ y + Z + µX −1 = C, Z ≽ 0}. The QP-Logdet problem (P ) is itself a classical convex model problem in optimization theory. It can be regarded as an extension of the qudratic semidefinite programming problem (QSDP) and the log-determinant (Logdet) problem, so it shares the structures of both problems, and it goes without saying that the QP-Logdet problem is considerable. For the QSDP, it is certainly a heart problem in nonlinear semidefinite programming problems, which has been considered by Toh [35], Toh, T¨ ut¨ unc¨ u and Todd [36, 37], Zhao [45], Jiang, Sun and Toh [14], etc.. For the Logdet problem, which has a very important application in covariance selection [5] and has been intensively studied over the past several years, including the work of Dahl, Vandenberghe and Roychowdhury [4], d’Aspremont, Banerjee and El Ghaoui [6], Li and Toh [15], Lu [16, 17], Lu and Zhang [18], Olsen, Oztoprak, Nocedal and Rennie [24], Scheinberg, Ma and Goldfarb [30], Scheinberg and Rish [31], Toh [34], Wang, Sun and Toh [40], Yang, Sun and Toh [41], Yang, Shen, Wonka, Lu and Ye [43], Yuan [44], etc.. As far as the QP-Logdet problem be concerned, it also arises in many practical applications such as robust simulation of global warming policies [13], speech recognition [39], and so on. Thus the algorithms developed to solve this kind of problems can potentially find wide applications. For the QP-Logdet problems of small and medium size, the interior-point method (IPM) with a direct solver is certainly an efficient and robust approach; however, for these large-scale problems, the IPM lacks the ability due to the need of computing, storing, and factorizing the Schur matrices that are typically dense. In view of this, we need to design new approaches. The proximal augmented Lagrangian (PAL) method proposed in [7] is a fast primaldual approach. The key idea of this approach is to apply the proximal technique to the augmented Lagrangian function of the primal problem at every iteration so that the 2 simplified inner problem has an analytical or at least a semi-analytical solution which can be solved very fast, and then to update the dual variables. The convergence of the PAL method has been shown in [7]. The biggest advantage of the PAL method is that there is no need to solve any linear system of equations to obtain step directions. It is a good approach for generating a good initial point. Furthermore, for some ill-conditioned inner problems whose Hessian matrices are of near low ranks, the proximal technique of the PAL method is actually better than the Newton-CG method since it can be regarded as a regularization treatment in some sense. Allowing for the advantages of the PAL method, we may apply it to the QP-Logdet problem, especially for generating an initial point. Although the PAL method is a good choice for the QP-Logdet problem, it is after all a gradient method, once the iteration point is near the optimal solution, we may consider an approach with locally faster convergence rate. The augmented Lagrangian (AL) method [29] is just in the position to play this role. It is a classical method that can be viewed as a special form of the proximal point algorithm (PPA). (As for the PPA, it was proposed by Martinet [19], and further studied by Rockafellar [28, 29].) Although the AL method for convex programs is a gradient ascent method applied to the corresponding dual problems [27], Sun, Sun and Zhang revealed that under the constraint nondegenerate conditions [1], it could be locally regarded as an approximate generalized Newton method applied to a semismooth equation. It is just this reason that inspired us to apply the AL method to solve the QP-Logdet problem. Great successes of the applications of the AL method to large-scale semidefinite programming problems can also be seen in [46, 40, 41] and the references therein. If applying the AL method to problem (P), we have to solve the inner problems which have no closed-form solutions. The well-tested quadratically convergent semismooth Newton method introduced by Qi and Sun [25] is an ideal approach to fulfill this task. As for the semismooth Newton direction, the preconditioned conjugate gradient (PCG) method is a good choice because the Newton system of equations is large and a direct solver is not proper. Mainly based on Rockafellar’s theoretical framework the global convergence and local convergence rate of the sequence generated by the Newton-CG AL (NAL) method can be established. The numerical results demonstrate that the proposed approach can be very efficient and robust to supply highly accurate solutions for large-scale problems not only for synthetic problems but also for real data problems. In this paper, we solve synthetic QP-Logdet problems with n up to 2, 000 and m up to 1, 186, 173 in about 2 hours. The remaining parts of this paper are organized as follows. In Section 2, we give a brief introduction on some preliminary contents. In Section 3, we describe the algorithms in details. In Section 4, we establish the convergence theory of the proposed approaches. In Section 5 we describe some numerical issues on the semismooth Newton-CG method. In Section 6, we present the numerical results. Finally, we give some concluding remarks. 3 1.1 Additional Notations In this paper, all vectors are assumed to be finite dimensional. The symbol Rn denotes the n-dimensional Euclidean space, and On denotes the set of n × n orthogonal matrices. The set of all m × n matrices with real entries is denoted by Rm×n , ⟨·, ·⟩ stands for the standard trace inner product in S n , ∥ · ∥ denotes the induced Frobenius norm, S+n n and S++ denote the sets of positive semidefinite matrices and positive definite matrices, respectively. The symbol ◦ denotes the Hadamard product, and ⊗ denotes the Kronecker product. The symbol FP ε := {X ∈ S n : A(X) = b, X ≽ εI} denotes the feasible set of the problem (P ε ), and ΠK (·) denotes the metric projector onto the closed convex set K. Let vec : Rm×n → Rmn be the vectorization operator on matrices defined by stacking the columns of a matrix to a long vector one by one. 2 Preliminaries In order to be able to present our ideas more clearly, we first introduce some concepts related to the AL method based on the classical papers by Rockafellar [28, 29]. 2.1 Maximal monotone operators Let H be a real Hilbert space with an inner product ⟨·, ·⟩. A multifunction T : H ⇒ H is said to be a monotone operator if ⟨z − z ′ , w − w′ ⟩ ≥ 0, whenever w ∈ T (z), w′ ∈ T (z ′ ). (2) It is said to be maximal monotone if, in addition, the graph G(T ) = {(z, w) ∈ H × H| w ∈ T (z)} is not properly contained in the graph of any other monotone operator T ′ : H ⇒ H. For example, if T is the subdifferential ∂f of a lower semicontinuous convex function f : H → (−∞, +∞], f ̸≡ +∞, then T is maximal monotone (see Minty [21] or Moreau [22]), and the relation 0 ∈ T (z) means that f (z) = min f . 2.2 Representations in terms of maximal monotone operators The following Karush-Kuhn-Tucker (KKT) conditions are necessary and sufficient for the optimality of (P ) and (D): A(X) − b = 0, Q(X) + C − A∗ y − Z = 0, XZ = µI, X ≽ 0, Z ≽ 0. 4 The following KKT conditions are also necessary and sufficient for the optimality of (P ε ) and (Dε ): A(X) − b = 0, Q(X) + C − A∗ y − Z − µX −1 = 0, ⟨X − εI, Z⟩ = 0, X − εI ≽ 0, Z ≽ 0. Throughout this paper, the following conditions for (P ε ) and (Dε ) are assumed to hold. Assumption 2.1. Problem (P ε ) satisfies the condition n ∃ X0 ∈ S++ such that A(X0 ) = b, X0 ≻ εI. Assumption 2.2. For problem (P ε ), there exists an α ∈ R such that the level set {x| f0 (X) ≤ α, A(X) = b, X ≽ εI} is nonempty and bounded, where f0 (x) stands for the objective function. Let l(X; y, Z) : S n × Rm × S n → R the extended form: L0 (X; y, Z) −∞ l(X; y, Z) = +∞ be the ordinary Lagrangian function for (P ε ) in if X ∈ FP ε and (y, Z) ∈ Rm × S+n , if X ∈ FP ε and (y, Z) ̸∈ Rm × S+n , if X ̸∈ FP ε , where L0 (X; y, Z) = 1 ⟨X, Q(X)⟩ + ⟨C, X⟩ − µ log det X − ⟨y, A(X) − b⟩ − ⟨Z, X − εI⟩. 2 The essential objective function in (P ε ) is given by { 1 ⟨X, Q(X)⟩ + ⟨C, X⟩ − µ log det X if X ∈ FP ε , 2 f (X) = sup l(X; y, Z) = +∞ otherwise, y∈Rm ,Z∈S n while the essential objective function in (Dε ) is given by { g0 (y, Z) if (y, Z) ∈ Rm × S+n , g(y, Z) = inf n l(X; y, Z) = X∈S −∞ if (y, Z) ̸∈ Rm × S+n , 5 As in [29], we can define the following operators Tl (X, y, Z) = {(V, u, W ) ∈ S n × Rm × S n | (V, −u, −W ) ∈ ∂l(X; y, Z)}, (X, y, Z) ∈ S n × Rm × S n , Tf (X) = {V ∈ S n | V ∈ ∂f (X)}, X ∈ S n , Tg (y, Z) = {(u, W ) ∈ Rm | (−u, −W ) ∈ ∂g(y, Z)}, (y, Z) ∈ Rm × S n . We can observe that l is a closed proper saddle function in the sense of [26, p.363], and the mapping Tl is a maximal monotone operator in S n ×Rm ×S n [26, Cor.37.5.2]. Meanwhile, f is a closed proper convex function, so Tf is a maximal monotone operator in S n [21, 22]. Similarly, g is a closed proper concave function, so Tg is a maximal monotone operator in Rm × S n . To discuss the rate of convergence, we introduce some related concepts and conclusions. Definition 2.1. (cf. [28]) For a maximal monotone operator T , we say that its inverse T −1 is Lipschitz continuous at the origin (with modulus a ≥ 0) if there is a unique solution z¯ to z = T −1 (0), and for some τ > 0 we have ∥z − z¯∥ ≤ a∥w∥, where z ∈ T −1 (w) and ∥w∥ ≤ τ. Assumption 2.3. Let X be the optimal solution to problem (P ε ) and (y, Z) is the corresponding Lagrangian multiplier. We say that the primal constraint nondegeneracy [1] holds at X to problem (P ε ) if ( ) ( ) ( m ) {0} A R n S + = , n lin(TS+ (X − εI)) I Sn where lin(TS+n (X − εI)) denotes the lineality space of TS+n (X − εI), and TS+n (X − εI) denotes the tangent cone of S+n at X − εI. Assumption 2.4. Let X be the optimal solution to problem (P ε ) and (y, Z) is the corresponding Lagrangian multiplier. We say that the strong second order sufficient condition [3] holds at X if ⟨H, ∇2XX L0 (X; y, Z)(H)⟩ + Υ(X−εI) (Z, H) > 0, ∀ H ∈ aff(C(X)) \ {0}, where the linear-quadratic function ΥB : S n × S n → R is defined by ΥB (Γ, A) := 2⟨Γ, AB + A⟩, (Γ, A) ∈ S n × S n , where B + is the Moore-Penrose pseudo-inverse of B. 6 Next we present the direct condition for the Lipschitz continuity of Tg−1 due to Rockafellar [29, Prop.2]. Proposition 2.1. If the strong second-order sufficient condition and the primal constraint nondegeneracy condition are satisfied, then Tl−1 is actually single-valued and continuously differentiable on a neighborhood of the origin, and so are Tf−1 and Tg−1 . Thus in particular, these mappings are all Lipschitz continuous at the origin. Remark 2.1. We must emphasize that the so called strong second-order condition in [29] actually includes the primal constraint nondegeneracy condition. 3 3.1 The algorithms The PAL method In this subsection, we describe the PAL method in details. The augmented Lagrangian function of (P) is Lσ (X; y) = 1 σ ⟨X, Q(X)⟩ + ⟨C, X⟩ − µ log det X − ⟨y, A(X) − b⟩ + ∥A(X) − b∥2 , X ≽ 0. 2 2 At the kth iteration, we solve the following subproblem } { 1 argmin Lσk (X; y k ) + ∥X − X k ∥2Tk | X ≽ 0 2 {1 } = argmin ⟨X, (Q + σk A∗ A + Tk )(X)⟩ − ⟨M k , X⟩ − µ log det X | X ≽ 0 , (3) 2 where M k = A∗ y k + σk A∗ b + Tk (X k ) − C. Pick Tk ≽ 0 such that Q + σk A∗ A + Tk = αk I, with αk = λmax (Q + σk A∗ A), then M k = −Q(X k ) − C + αk X k + A∗ (y k + σk (b − A(X k ))), and (3) simplifies to {1 } 1 k 2 argmin αk ∥X − M ∥ − µ log det X | X ≽ 0 . (4) 2 αk Based on Lemma 2.1 in [40], problem (4) has a closed-form solution X k+1 = ϕ+ γk ( 1 k k k T M ) := P k diag(ϕ+ γk (d ))(P ) , αk (5) Z k+1 = ϕ− γk ( 1 k k k T M ) := P k diag(ϕ− γk (d ))(P ) , αk (6) and further where α1k M k = P k Dk (P k )T is the eigenvalue decomposition of α1k M k with P k ∈ On and 1 k k − 1 k + Dk = diag(dk ), dk ∈ Rn , ϕ+ γk ( αk M ) and ϕγk ( αk M ) are matrix valued functions, ϕγk (d ) 7 √ k ϕ− γk (d ) ϕ+ γk (x) x2 +4γ k are vector valued functions, whose scalar counterparts are = 2 √ x2 −4γk and ϕ− , correspondingly, and γk = αµk . From (5) and (6), we can see γk (x) = 2 that X k+1 and Z k+1 are positive definite automatically. Once X k+1 is obtained, one may update the dual variable to get y k+1 . As for λmax (Q + σk A∗ A), we may apply the power method to compute it. Now we may summarize the PAL method as below and n Algorithm 1: The PAL method. Input X 0 ∈ S++ , y 0 ∈ Rm , σ0 = 10. Iterate the following steps: Step 1. Compute αk = λmax (Q + σk A∗ A). Step 2. Set Tk = αk I − Q − σk A∗ A, and M k = A∗ y k + σk A∗ b + Tk (X k ) − C. Compute X k+1 = ϕ+ γk ( 1 k 1 k M ) and Z k+1 = ϕ− M ). γk ( αk αk Step 3. Set y k+1 = y k − τ σk (A(X k+1 ) − b), where τ ∈ [1, 2) is a given parameter. Step 4. Compute RPk+1 = ∥b − A(X k+1 )∥ , max{1, ∥b∥} k+1 RD = ∥Q(X k+1 ) + C − A∗ y k+1 − Z k+1 ∥ , max{1, ∥C∥} k+1 If max{RPk+1 , RD } < Tol1, stop; else, choose σk+1 ; end. 3.2 The NAL method In this subsection, we apply the NAL method to the perturbed problem (P ε ) since this method can not guarantee the positive definiteness of the variable X automatically. Given a penalty parameter σ > 0, the augmented Lagrangian function for problem 8 (P ε ) is defined as Lσ (X; y, Z) = 1 σ ⟨X, Q(X)⟩ + ⟨C, X⟩ − µ log det X − ⟨y, A(X) − b⟩ + ∥A(X) − b∥2 + 2 2 1 1 ∥ΠS+n (Z − σ(X − εI))∥2 − ∥Z∥2 . 2σ 2σ At the kth iteration, we solve the following subproblem { } argmin Lσk (X; y k , Z k ) | X ≽ εI . (7) Since the subproblem (7) has no closed-form solution, it needs elaborating on. We can calculate the first-order derivative of Θk (X) := Lσk (X; y k , Z k ) with respect to X ∇Θk (X) = Q(X) + C − A∗ y k − µX −1 + σk A∗ (A(X) − b) − ΠS+n (Z k − σk (X − εI)). Furthermore, base on [26, Thm. 23.8] and [32, Lem. 2.1] we can calculate the Clarke generalized Jacobian ∂ 2 Θk (X) := ∂∇Θk (X) as below ∂ 2 Θk (X)[H] = (Q + σk A∗ A)[H] + µX −1 HX −1 + σk ∂ΠS+n (Z k − σk (X − εI)), ∀ H ∈ S n . Actually, from [20, Prop.1] every element in ∂ΠS+n (·) is positive semidefinite, so every element in ∂ 2 Θk (X) is positive definite. Therefore we can apply the semismooth Newton method developed in [25] with the line search technique which guarantees the global convergence. And we may obtain the Newton direction by solving the linear system of equations Vbσ0k [H] = −∇Θk (X k ), where Vbσ0k := Q + σk A∗ A + µ(X k )−1 ⊗ (X k )−1 + Vσ0k , here Vσ0k ∈ ∂ΠS+n (Z k − σk (X k − εI)) may be chosen in the same way as that in [46] Vσ0k [H] = σk P k (Ωk ◦ ((P k )T (H)P k ))(P k )T , where P k ∈ On and Z k − σk (X k − εI) = P k Λk (P k )T , where Λk is the diagonal matrix with diagonal entries consisting of the eigenvalues λk1 ≥ λk2 ≥ · · · ≥ λkn of Z k − σk (X k − εI); if defining three index sets αk := {i | λki > 0}, βk := {i | λki = 0}, and γk := {i | λki < 0}, [ then k Ω = Eγ¯k γ¯k νγ¯k γk 0 νγ¯Tk γk ] , νij := 9 λki , i ∈ γ¯k , j ∈ γk , λki − λkj here γ¯k = {1, · · · , n} \ γk , and Eγ¯k γ¯k ∈ S |¯γk | is the matrix of ones. Now we summarize the algorithm as below Algorithm 2: The NAL method. Input X 0 ∈ FP ε , y 0 ∈ Rm , Z 0 ∈ S+n , σ0 = 10. Iterate the following steps: Step 1. Find an approximate minimizer X k+1 ≈ arg minn Θk (X). X∈S++ (8) Step 2. Compute y k+1 = y k − σk (A(X k+1 ) − b), Z k+1 = ΠS+n (Z k − σk (X k+1 − εI)). Step 3. Compute RPk+1 = ∥b − A(X k+1 )∥ , max{1, ∥b∥} k+1 RD = ∥Q(X k+1 ) + C − A∗ y k+1 − Z k+1 − µ(X k+1 )−1 ∥ , max{1, ∥C∥} k+1 If max{RPk+1 , RD } < Tol2, stop; else, σk+1 = 2σk ; end. In Algorithm 2, the principal computing costs lie in Step 1, i.e., computing the approximate minimizer of the inner problem (8). So we state the algorithm about how to compute the approximate minimizer in details below. 10 Algorithm 3: The semismooth Newton-CG Method. Step 1. Given ζ ∈ (0, 12 ), τ1 , τ2 ∈ (0, 1), and δ ∈ (0, 1), choose X 0 (≽ εI). Step 2. For j = 0, 1, 2, . . . , Step 1.1. Apply the PCG method to find an approximate solution H j to (Vbσ0k + ϵj I)[H] = −∇Θk (X j ), (9) where ϵj := τ1 min{τ2 , ∥∇Θk (X j )∥}. Step 1.2. Set αj = δ mj , where mj is the first nonnegative integer m for which Θk (X j + δ m H j ) ≤ Θk (X j ) + ζδ m ⟨∇Θk (X j ), H j ⟩. Step 1.3. Set X j+1 = X j + αj H j . From Algorithm 3, we can see that the main computing costs lie in solving the linear operator equation (9). The linear operator corresponds to a fourth-order n dimemsional tensor or a n2 × n2 matrix. For large-scale problems, n is very large, so direct solvers are not suitable and PCG solvers are ideal approaches for (9). Among various PCG solvers, we adopt the symmetric QMR algorithm [9]. But we must note that comparing with the algorithm in [9], we change the matrix to the linear operator in (9), and change the vectors to the corresponding matrices. As these changes are only a trivial generalization of the symmetric QMR algorithm, we do not go to details. 4 Convergence analyses The convergence analyses of these two methods can be derived from Fazel, Pong, Sun and Tseng’s paper [7] and Rockafellar’s paper [28, 29] without many difficulties, respectively. For the sake of completeness, we also present these results below. 4.1 Convergence analysis for the PAL method Theorem 4.1. Assume that the solution set of (P) is nonempty and Assumption 2.1 holds. Assume that Q + σk A∗ A + √Tk is positive definite. Let {X k , y k , Z k } be generated from Algorithm 1, and if τ ∈ (0, 1+2 5 ), then the sequence {X k } converges to the optimal solution to (P) and {y k , Z k } converges to the optimal solution to the dual problem (D). Proof. Since Q + σk A∗ A + Tk = λmax (Q + σk A∗ A)I ≻ 0, based on Theorem B.1 in [7] and the KKT condition, the conclusion of the theorem is obvious. 11 4.2 Convergence analysis for the NAL method Since we can not solve the inner problems exactly, we will use the following stopping criteria considered by Rockafellar [28, 29] for terminating Algorithms 2 and 3: (A) Θk (X k+1 ) − inf Θk (X) ≤ ϵ2k /2σk , ϵk ≥ 0, ∞ ∑ ϵk < ∞; k=0 (B) Θk (X k+1 ) − inf Θk (X) ≤ δk2 /2σk ∥(y k+1 , Z k+1 ) − (y , Z )∥ , δk ≥ 0, k k 2 ∞ ∑ δk < ∞; k=0 (B ′ ) ∥∇Θk (X k+1 )∥ ≤ δk′ /σk ∥(y k+1 , Z k+1 ) − (y k , Z k )∥, 0 ≤ δk′ → 0. We directly obtain from [28, 29] the following convergence results. Theorem 4.2. Let Algorithm 2 be executed with stopping criterion (A). If Assumption 2.1 is satisfied, then the generated sequence {(y k , Z k )} is bounded and {(y k , Z k )} converges to (y, Z), where (y, Z) is the unique optimal solution to (Dε ), and {X k } is asymptotically minimizing for (P ε ) with max(Dε ) = inf(P ε ). The boundedness of {y k , Z k } under (A) is actually equivalent to the existence of an optimal solution to (Dε ). If {y k , Z k } is bounded and Assumption 2.2 is satisfied, then the sequence {X k } is also bounded, and the accumulation point of the sequence {X k } is the unique optimal solution to (P ε ). Theorem 4.3. Let Algorithm 2 be executed with stopping criteria (A) and (B). If Assumption 2.1 holds and Tg−1 is Lipschitz continuous at the origin with modulus ag , then {(y k , Z k )} converges to the unique optimal solution (y, Z) with max(Dε ) = inf(P ε ), and for all k sufficiently large, ∥(y k+1 , Z k+1 ) − (y, Z)∥ ≤ θk ∥(y k , Z k ) − (y, Z)∥, where 2 −1/2 θk = [ag (a2g + σk2 )−1/2 + δk ](1 − δk )−1 → θ∞ = ag (a2g + σ∞ ) < 1, σk → σ∞ , Moreover, the conclusions of Theorem 4.2 about {(y k , Z k )} are valid. If in addition to (B) and the condition on Tg−1 one has (B’) and the stronger condition that Tl−1 is Lipschitz continuous at the origin with modulus al (≥ ag ), then X k → X, where X is the unique optimal solution to (P ε ), and one has that for all k sufficiently large, ∥X k+1 − X∥ ≤ θk′ ∥(y k+1 , Z k+1 ) − (y k , Z k )∥, where ′ = al /σ∞ . θk′ = al (1 + δk′ )/σk → θ∞ 12 Remark 4.1. In Algorithm 2 we can also add the term 2σ1k ∥X −X k ∥2 to Θk (X). Actually, in our Matlab code, one can optionally add this term. This actually corresponds to the proximal method of multipliers considered in [29, Section 5]. Convergence analysis for this improvement can be conducted in a parallel way as for Algorithm 2. Note that in the stopping criteria (A) and (B), inf Θk (X) is an unknown value. Since b Θk (X) := Θk (X) + 2σ1k ∥X − X k ∥2 is a strongly convex function with modulus σ1k , then one has the estimation b k (X k+1 ) − inf Θ b k (X) ≤ 1 ∥∇Θ b k (y k+1 )∥2 , Θ 2σk thus criteria (A) and (B) can be practically modified as follows: b k (y k+1 )∥ ≤ ϵk , ϵk ≥ 0, ∥∇Θ b k (X k+1 )∥ ≤ δk ∥(y k+1 , Z k+1 ) − (y k , Z k )∥, δk ≥ 0, ∥∇Θ ∞ ∑ k=0 ∞ ∑ ϵk < ∞; δk < ∞. k=0 Remark 4.2. The condition that Tg−1 is Lipschitz continuous at the origin is not easy to be realized, however, if Assumptions 2.3 and 2.4 are satisfied, then Tl−1 , Tf−1 and Tg−1 are all single-valued and continuously differentiable on a neighborhood of the origin, and in particular they are all Lipschitz continuous at the origin. 5 Numerical issues in the associated semismooth NewtonCG method When applying Algorithm 3 to solve the inner problem (7), the most expensive step lies in solving the linear system of equations where (M + εI)vec(H) = vec(−∇Θk (X)), (10) M := Q + σAT A + µX −1 ⊗ X −1 + σP ⊗ P diag(vec(Ω))P ⊗ P, (11) Q and A denote the matrix representations of Q and A, to obtain the approximate Newton direction. In order to solve (10) as efficiently as possible, it is necessary to analyze the condition number of the coefficient matrix and design an efficient preconditioner. 5.1 Conditioning of the coefficient matrix Let S = X − εI. 13 For simplicity, we suppose that strict complementarity holds for Z, S, i.e., Z + S ≻ 0. From the fact that ZS = 0, we have [ Z ] Λ 0 PT, (12) Z − σS = P 0 −σΛS where ΛZ = diag(λZ ) ∈ Rr×r and ΛS = diag(λS ) ∈ R(n−r)×(n−r) are diagonal matrices of positive eigenvalues of Z and S, respectively. Define the index sets γ¯ := {1, . . . , r}, γ := {r + 1, . . . , n}. Let ] [ λZi Eγ¯γ¯ νγ¯γ , ν := Ω= , i ∈ γ¯ , j ∈ γ, (13) ij νγ¯Tγ 0 λZi + σλSj−r and c1 = min(λZ ) max(λZ ) , c = < σ. 2 min(λZ )/σ + max(λS ) max(λZ )/σ + min(λS ) Then we have Proposition 5.1. Suppose that the strict complementarity holds for Z and S, then we have the following bound on the condition number of M κ(M ) ≤ λmax (Q) + σλmax (AT A) + µ/λ2min (X) + σλmax (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T ) , λmin (Q) + σλmin (AT A) + µ/λ2max (X) + c1 λmin (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T ) (14) where Pe1 = Pγ¯ ⊗ Pγ¯ , Pe2 = Pγ ⊗ Pγ¯ , Pe3 = Pγ¯ ⊗ Pγ , Pγ¯ and Pγ are submatrices of P = [Pγ¯ Pγ ]. Proof. From (11), (12) and (13), we obtain M = Q + σAT A + µX −1 ⊗ X −1 + σ(Pe1 Pe1T + Pe2 D2 Pe2T + Pe3 D3 Pe3T ), where D2 = diag(vec(νγ¯γ )), and D3 = diag(vec(νγ¯Tγ )). Since c1 ≤ σνij ≤ c2 , i ∈ γ¯ , j ∈ γ, then we can derive c1 (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T ) ≼ σ(Pe1 Pe1T + Pe2 D2 Pe2T + Pe3 D3 Pe3T ) ≼ σ(Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T ). Furthermore, from [11, Thm 4.2.12], µ λ2max (X) I ≼ µX −1 ⊗ X −1 ≼ 14 µ λ2min (X) I. Hence from [10, Thm 4.3.7], µ + c1 λmin (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T )]I ≼ M ≼ 2 λmax (X) µ + σλmax (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T )]I. [λmax (Q) + σλmax (AT A) + 2 λmin (X) [λmin (Q) + σλmin (AT A) + So we can get the following bound on the condition number of M : κ(M ) ≤ λmax (Q) + σλmax (AT A) + µ/λ2min (X) + σλmax (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T ) . λmin (Q) + σλmin (AT A) + µ/λ2max (X) + c1 λmin (Pe1 Pe1T + Pe2 Pe2T + Pe3 Pe3T ) The upper bound in (14) suggests that: (i) with σ and 1/c1 increasing, the condition number κ(M ) may increase. (ii) With Q, AT A and P ⊗ P diag(vec(Ω))P ⊗ P being of (near) low ranks, µX −1 ⊗ X −1 can potentially improve the condition number, for the term µ/λ2max (X)I can “lift” the minimal eigenvalue of M ; in other words, if Q, AT A and P ⊗ P diag(vec(Ω))P ⊗ P are of (near) low ranks, then with µ decreasing, the condition number κ(M ) may probably increase. 5.2 A diagonal preconditioner To achieve faster convergence for the PCG method to solve (10), one may select a proper preconditioner. In our implementation, we devise an easy-to-compute diagonal preconditioner by using an idea first developed in [8]. The preconditioner has the following form MD := diag(Q) + σdiag(AT A) + µdiag(vec(Ψ)) + σdiag(d), where Ψ ∈ S n , d ∈ Rn , and 2 (Ψ)ij = (X −1 )ii (X −1 )jj + (X −1 )2ij , d(ij) = ((P ◦ P )Ω(P ◦ P )T )ij , 1 ≤ i, j ≤ n. The biggest advantage of the preconditioner MD is that only O(n3 ) flops are needed to compute it. 6 Numerical experiments In this section, we present some numerical results to demonstrate the performances of the approaches with both synthetic and real data. We implemented the approaches in Matlab R2013a. All runs were performed on a PC (Intel Core 2 Duo 2.60 GHz with 12 GB RAM). 15 We measure the infeasibilities and optimality for the primal and dual problems by RP , RD and RG which are described in Algorithms 1 and 2. In general cases, we use the PAL method to generate an initial point and then switch to the NAL method for accelerating the local convergence. We stop the PAL method when max{RP , RD } < Tol1, and stop the NAL method when max{RP , RD } < Tol2, where Tol1 and Tol2 are pre-specified accuracy tolerances with Tol1 = 5 × 10−3 and Tol2 = 10−6 as the default. In some hard cases, we use the PAL method only and terminate it with Tol1 = 10−6 . Furthermore, we also use the relative gap RG := |pobj − dobj| 1 + |pobj| + |dobj| to measure the quality of the solution, where pobj and dobj denote the primal and dual objective function values, respectively. For the PAL method, we cap the iteration number to be 1000; for the Newton-CG method, we set the maximal outer iteration number to be 100 and the maximal inner iteration number to be 15. We choose the initial iterate X0 = I, and σ0 = 10. If the NAL method is applied to the above problems, the inequality constraint X ≽ 0 is replaced by X ≽ εI with ε = 10−16 . 6.1 Synthetic experiments I In this subsection, we focus our numerical experiments on the following special problems } {1 ⟨X, H ◦ X⟩ + ⟨C, X⟩ − µ log det X | Xii = 1, i = 1, 2, · · · , n, X ≽ 0 (A1) min X 2 { } Xij = Xji = 0, ∀ (i, j) ∈ E 1 (A2) min ⟨X, H ◦ X⟩ + ⟨C, X⟩ − µ log det X X 2 Xii = 1, i = 1, 2, · · · , n, X ≽ 0 {1 } (A3) min ⟨X, H ◦ X⟩ + ⟨C, X⟩ − µ log det X | Tr(X) = 1, X ≽ 0 X 2 } { Xij = Xji = 0, ∀ (i, j) ∈ E 1 (A4) min ⟨X, H ◦ X⟩ + ⟨C, X⟩ − µ log det X . X 2 Tr(X) = 1, X ≽ 0 The matrices H and C are generated randomly, and E is a random subset of {(i, j) | 1 ≤ i < j, j = 2, . . . , n}. The performances of the algorithms are presented in the following tables. For each instance, we report the matrix dimension (n) and the number of the equality constraints 16 (m); the number of outer iterations (it) and the total number of inner iterations (itsub); the primal (pobj) and dual (dobj) objective values; the primal infeasibility (RP ), the dual infeasibility (RD ), the relative gap (RG ); the time (in the format of hours:minutes:seconds) taken. Firstly, we present the performances of the pure PAL method and the hybrid method (i.e., the PAL method and the NAL method) on the problems (A1) and (A2) with µ = 1 in Tables 1 and 2, respectively. But for the hybrid method, we only report the details of the Newton-CG method. From the two tables, we can see that although the PAL method can solve the problems rapidly, the hybrid method is still superior to it for all instances. Especially for instances with relatively few constraints, the hybrid method outperforms the PAL method obviously. Actually, the superiority of the hybrid method will be even obvious if we want to get a highly accurate solution. Table 1: Performances of the PAL method on problems (A1)-(A4) with µ = 1. Problem n|m it rand-A1-0500 rand-A1-1000 rand-A1-1500 rand-A1-2000 rand-A2-0500 rand-A2-1000 rand-A2-1500 rand-A2-2000 rand-A3-0500 rand-A3-1000 rand-A3-1500 rand-A3-2000 rand-A4-0500 rand-A4-1000 rand-A4-1500 rand-A4-2000 500 | 500; 1000 | 1000; 1500 | 1500; 2000 | 2000; 500 | 12322; 1000 | 25086; 1500 | 38389; 2000 | 51322; 500 | 1; 1000 | 1; 1500 | 1; 2000 | 1; 500 | 74080; 1000 | 296243; 1500 | 666966; 2000 | 1186173; 93| 88| 88| 95| 106| 102| 99| 95| 570| 1000| 1000| 1000| 1000| 1000| 1000| 1000| pobj 1.51611644 2.88889067 4.02954208 5.12194289 1.48549608 2.92216880 4.07660888 5.14445107 3.10830391 6.90874980 1.09705221 1.52001337 3.10830458 6.90848660 1.09609128 1.51822498 17 dobj 3 3 3 3 3 3 3 3 3 3 4 4 3 3 4 4 1.51612384 2.88891640 4.02959372 5.12194882 1.48549478 2.92216820 4.07660986 5.14444160 3.10830441 6.90875463 1.09708284 1.52027947 3.10830554 6.90875346 1.09707760 1.52026740 3 3 3 3 3 3 3 3 3 3 4 4 3 3 4 4 RP /RD /RG Time 4.2-7| 9.4-7| 2.4-6 6.8-7| 9.9-7| 4.5-6 8.9-7| 9.9-7| 6.4-6 9.2-7| 2.5-7| 5.8-7 5.1-7| 9.2-7| 4.4-7 2.7-7| 9.6-7| 1.0-7 6.6-8| 9.9-7| 1.2-7 3.3-7| 9.1-7| 9.2-7 9.8-7| 9.3-7| 7.9-8 4.8-6| 4.6-6| 3.5-7 2.0-4| 1.9-4| 1.4-5 1.3-3| 1.3-3| 8.8-5 1.6-4| 3.5-4| 1.5-7 2.8-4| 4.7-4| 1.9-5 6.6-3| 7.0-3| 4.5-4 1.0-2| 7.5-3| 6.7-4 18 1:20 3:41 8:13 20 1:47 4:28 8:53 1:08 10:36 31:24 1:16:31 3:29 17:55 46:26 1:41:03 Table 2: Performances of the hybrid method on problems (A1)(A4) with µ = 1. Problem n|m it|itsub rand-A1-0500 rand-A1-1000 rand-A1-1500 rand-A1-2000 rand-A2-0500 rand-A2-1000 rand-A2-1500 rand-A2-2000 rand-A3-0500 rand-A3-1000 rand-A3-1500 rand-A3-2000 rand-A4-0500 rand-A4-1000 rand-A4-1500 rand-A4-2000 500 | 500; 1000 | 1000; 1500 | 1500; 2000 | 2000; 500 | 12322; 1000 | 25086; 1500 | 38389; 2000 | 51322; 500 | 1; 1000 | 1; 1500 | 1; 2000 | 1; 500 | 74080; 1000 | 296243; 1500 | 666966; 2000 | 1186173; 5| 8 6| 9 6| 8 6| 7 10| 11 11| 12 10| 11 10| 11 13| 15 15| 17 20| 22 16| 19 32| 55 36| 63 52| 92 79| 140 pobj 1.51610927 2.88889217 4.02952847 5.12195286 1.48549372 2.92217094 4.07661066 5.14445263 3.10830441 6.90875466 1.09708296 1.52028038 3.10830709 6.90875652 1.09708312 1.52028052 dobj 3 3 3 3 3 3 3 3 3 3 4 4 3 3 4 4 1.51593972 2.88900007 4.02940576 5.12224022 1.48549527 2.92217815 4.07661215 5.14445533 3.10830442 6.90875467 1.09708296 1.52028038 3.10830709 6.90875652 1.09708312 1.52028052 3 3 3 3 3 3 3 3 3 3 4 4 3 3 4 4 RP /RD /RG Time 5.3-7| 1.7-7| 5.6-5 1.3-7| 3.9-8| 1.9-5 3.2-7| 8.2-8| 1.5-5 5.1-7| 1.4-7| 2.8-5 7.7-7| 6.4-7| 5.2-7 3.8-7| 2.0-7| 1.2-6 3.8-7| 2.3-7| 1.8-7 4.5-7| 2.4-7| 2.6-7 3.6-9| 5.8-7| 2.4-9 1.3-9| 4.1-7| 7.4-10 2.8-9| 9.0-7| 9.6-10 1.2-9| 7.9-7| 6.6-10 1.4-10| 7.6-7| 1.5-12 5.5-11| 7.4-7| 4.4-13 8.7-11| 6.9-7| 3.6-13 1.9-10| 9.1-7| 5.9-13 08 41 1:47 3:44 15 1:20 3:26 7:50 35 5:25 22:55 1:07:14 1:37 13:39 59:34 2:09:36 Secondly, we present the performances of the PAL method and the hybrid method on problem (A1) with various µ. From Tables 3 and 4, we can see that with µ decreasing, the computing time increases very mildly. In contrast, with µ decreasing, the computing time increases very obviously. The rationale is analyzed behind Proposition 5.1, that is, in this problem, Q, AT A and P ⊗ P diag(vec(Ω))P ⊗ P are of near low ranks and a small µ leads to the ill-conditioning of the inner problem, so the NAL method needs many PCG iterations to solve the linear system of equations, however, the PAL method directly gives a closed-form solution of a regularized problem. 18 Table 3: Performances of the PAL method on problem (A1) with various µ. Problem rand1000-1 rand1000-2 rand1000-3 rand1000-4 rand1000-5 rand1000-6 rand1000-7 rand1000-8 rand1000-9 rand1000-10 n|m µ it | | | | | | | | | | 1.0000; 0.7500; 0.5000; 0.2500; 0.1250; 0.0625; 0.0313; 0.0157; 0.0078; 0.0039; 88| 113| 112| 126| 171| 169| 179| 179| 180| 180| 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 pobj 2.88889067 2.84053903 2.77900646 2.69409906 2.63413549 2.59461152 2.56991656 2.55504884 2.54622819 2.54123460 RP /RD /RG dobj 3 3 3 3 3 3 3 3 3 3 2.88891640 2.84053619 2.77901061 2.69412261 2.63413210 2.59461261 2.56991919 2.55505345 2.54623730 2.54124523 3 3 3 3 3 3 3 3 3 3 6.8-7| 8.2-7| 8.8-7| 1.6-7| 8.6-7| 8.8-7| 7.7-7| 9.4-7| 7.7-7| 8.2-7| 9.9-7| 1.9-7| 5.5-7| 9.9-7| 2.2-7| 4.7-7| 4.9-7| 6.4-7| 7.2-7| 8.0-7| Time 4.5-6 5.0-7 7.5-7 4.4-6 6.4-7 2.1-7 5.1-7 9.0-7 1.8-6 2.1-6 1:20 1:42 1:41 1:57 2:45 2:41 2:44 2:47 2:45 2:45 Table 4: Performances of the hybrid method on problem (A1) with various µ. Problem rand1000-1 rand1000-2 rand1000-3 rand1000-4 rand1000-5 rand1000-6 rand1000-7 rand1000-8 rand1000-9 rand1000-10 n|m µ it|itsub | | | | | | | | | | 1.0000; 0.7500; 0.5000; 0.2500; 0.1250; 0.0625; 0.0313; 0.0157; 0.0078; 0.0039; 6| 9 6| 9 6| 11 12| 24 12| 24 11| 22 11| 22 11| 22 11| 26 11| 30 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 1000 pobj 2.88889217 2.84053768 2.77900549 2.69409749 2.63412896 2.59460453 2.56991042 2.55504168 2.54622239 2.54122850 19 RP /RD /RG dobj 3 3 3 3 3 3 3 3 3 3 2.88900007 2.84064543 2.77914546 2.69409705 2.63412821 2.59460367 2.56990961 2.55504091 2.54622168 2.54122780 3 3 3 3 3 3 3 3 3 3 1.3-7| 1.4-7| 1.9-7| 9.5-8| 1.6-7| 1.9-7| 1.8-7| 1.7-7| 1.6-7| 1.5-7| 3.9-8| 3.6-8| 5.6-8| 4.7-7| 8.3-7| 9.8-7| 9.3-7| 9.1-7| 8.5-7| 8.4-7| 1.9-5 1.9-5 2.5-5 8.2-8 1.4-7 1.7-7 1.6-7 1.5-7 1.4-7 1.4-7 Time 43 45 59 3:18 3:51 5:12 6:59 9:08 12:17 15:50 1000 800 800 time (seconds) time (seconds) 1000 600 400 200 0 0 10 600 400 200 −1 10 0 0 10 −2 µ 10 −1 10 −2 µ 10 Figure 1: Comparison of the performances of two methods on problem (A1) with various µ. 6.2 Synthetic experiments II In this subsection, we focus the numerical experiments on the following problem: {1 } (B1) min ∥H ◦ (X − G)∥2 − µ log det X | Xii = 1, i = 1, 2, · · · , n, X ≽ 0 X 2 } { Xij = Xji = 0, ∀ (i, j) ∈ E 1 ∥H ◦ (X − G)∥2 − µ log det X . (B2) min X 2 Xii = 1, i = 1, 2, · · · , n, X ≽ 0 The matrices H and G are generated randomly, where H is a weighted matrix whose entries are between 0 and 1, and many entries are nearly zero, and E is a random subset of {(i, j) | 1 ≤ i < j, j = 2, . . . , n}. The objective function of (B1) or (B2) is 1 1 ⟨X, H ◦ H ◦ X⟩ − ⟨H ◦ H ◦ G, X⟩ + ⟨H ◦ G, H ◦ G⟩ − µ log det X, 2 2 (15) which is a special case of (P ), so the algorithms we developed can be applied to solve (B1)(B2). We list the performances of the PAL method and the hybrid method on problems (B1) and (B2) in Tables 5 and 6 as below. We can see that for these two problems, the PAL method is much faster than the hybrid method. The reason lies in that the entries of the weighted matrix H in (15) are small, then the entries of H ◦ H are even smaller which leads to a more ill-conditioned Hessian of the inner problem and in this situation the NAL method has no superiority. 20 Table 5: Performances of the PAL method on problems (B1)-(B2) with µ = 1. Problem n|m it rand-B1-0500 rand-B1-1000 rand-B1-1500 rand-B1-2000 rand-B2-0500 rand-B2-1000 rand-B2-1500 rand-B2-2000 500 | 500; 1000 | 1000; 1500 | 1500; 2000 | 2000; 500 | 12322; 1000 | 25086; 1500 | 38389; 2000 | 51322; 147| 114| 106| 107| 175| 143| 135| 129| pobj 3.90611146 1.76566465 4.38503029 8.41256796 5.30756232 1.91004704 4.53956279 8.56585712 RP /RD /RG dobj 3 4 4 4 3 4 4 4 3.90613831 1.76570729 4.38510787 8.41269558 5.30756434 1.91005117 4.53957194 8.56586838 3 4 4 4 3 4 4 4 8.9-7| 7.5-7| 2.5-7| 1.6-7| 8.6-7| 4.7-7| 3.4-8| 2.3-8| 3.1-7| 9.7-7| 9.4-7| 9.6-7| 2.1-7| 9.8-7| 9.9-7| 9.6-7| 3.4-6 1.2-5 8.8-6 7.6-6 1.9-7 1.1-6 1.0-6 6.6-7 Time 29 1:49 4:47 9:47 35 2:21 6:05 11:53 Table 6: Performances of the hybrid method on problems (B1)(B2) with µ = 1. Problem n|m rand-B1-0500 rand-B1-1000 rand-B1-1500 rand-B1-2000 rand-B2-0500 rand-B2-1000 rand-B2-1500 rand-B2-2000 500 | 500; 1000 | 1000; 1500 | 1500; 2000 | 2000; 500 | 12322; 1000 | 25086; 1500 | 38389; 2000 | 51322; it 9| 7| 7| 7| 6| 7| 7| 7| 25 20 19 20 17 19 19 19 pobj 3.90611894 1.76566628 4.38503287 8.41257271 5.30756255 1.91004712 4.53956281 8.56585840 RP /RD /RG dobj 3 4 4 4 3 4 4 4 3.90609819 1.76566528 4.38502532 8.41255864 5.30757122 1.91004680 4.53956184 8.56585294 3 4 4 4 3 4 4 4 6.9-7| 2.7-7| 3.7-7| 4.8-7| 3.5-7| 1.4-7| 1.7-7| 2.7-7| 7.6-7| 5.2-7| 6.5-7| 7.8-7| 9.8-7| 2.8-7| 3.1-7| 4.4-7| 2.7-6 2.9-7 8.6-7 8.4-7 8.2-7 8.2-8 1.1-7 3.2-7 Time 1:24 7:22 22:01 58:27 1:01 6:51 22:13 51:15 Based on the synthetic experiments in the above two subsections, we conclude that if the Hessian of the inner problem is good-conditioned, we prefer to adopt the hybrid method, but if the Hessian is ill-conditioned, the PAL method only is a good choice. 6.3 Real data experiments In this section, we consider the following model problem which is a special case of problem (P ) with Q ≡ 0: { } (C) min ⟨C, X⟩ − log det X | Xij = Xji = 0, ∀ (i, j) ∈ E, X ≽ 0 , X 21 where the matrix C is real data coming from the two gene expression data sets, and E is a predetermined index set. The model (C) finds wide applications in covariance selection. One gene set is the Rosetta Inpharmatics Compendium of gene expression profiles described by Hughes et al. [12]. The data set contains 253 samples with n = 6136 variables. We aim to estimate the covariance matrix of a Gaussian graphic model whose conditional independence is unknown. Another gene set is the Iconix microarray data obtained from 255 drug-treated rat livers; see Natsoulis et al. [23] for details. For both data sets, although our method can handle problems with larger matrix dimensions, we test only on a subset of the data. We create 5 subsets by taking 500, 1000, 1500, 2000 and 2500 variables with the highest variances, respectively. And as the variances vary widely, we normalize the sample covariance matrices to have unit variances on the diagonal. As the model (C) is a special case of (P ) with Q ≡ 0, so the Hessian of the inner problem may probably be ill-conditioned and we only apply the PAL method to solve it. Furthermore, we also compare the performance of the PAL method with that of the PPA proposed in [40]. Table 7: Performances of the PAL method on problem (C). Problem Rosetta-0500 Rosetta-1000 Rosetta-1500 Rosetta-2000 Rosetta-2500 Iconix-0500 Iconix-1000 Iconix-1500 Iconix-2000 Iconix-2500 n|m it | | | | | | | | | | 13| 12| 12| 12| 12| 28| 25| 21| 21| 16| 500 1000 1500 2000 2500 500 1000 1500 2000 2500 999; 1999; 2999; 3999; 4999; 999; 1999; 2999; 3999; 4999; pobj 2.29472323 3.06492978 3.56384267 3.94407754 4.19969175 1.30457645 2.28035658 2.92514554 3.42933318 3.84427349 22 RP /RD /RG dobj 1 1 1 1 1 2 2 2 2 2 2.29473435 3.06495129 3.56386620 3.94410344 4.19971940 1.30457705 2.28035773 2.92514672 3.42933507 3.84427544 1 1 1 1 1 2 2 2 2 2 4.3-7| 4.7-7| 3.7-7| 3.3-7| 2.9-7| 3.8-7| 1.3-7| 1.6-7| 6.6-8| 7.5-8| 5.1-7| 8.7-7| 6.7-7| 5.9-7| 5.4-7| 7.9-7| 9.0-7| 8.2-7| 9.7-7| 9.1-7| 2.4-6 3.5-6 3.3-6 3.2-6 3.3-6 2.3-7 2.5-7 2.0-7 2.8-7 2.5-7 Time 03 16 46 1:49 3:27 08 28 58 2:14 3:05 Table 8: Comparison of the PAL method and the PPA on problem (C) with accuracy 10−6 . Problem Rosetta-0500 Rosetta-1000 Rosetta-1500 Rosetta-2000 Rosetta-2500 Iconix-0500 Iconix-1000 Iconix-1500 Iconix-2000 Iconix-2500 n|m 500 1000 1500 2000 2500 500 1000 1500 2000 2500 | | | | | | | | | | it PAL PPA 13 12 12 12 12 28 25 21 21 16 2 2 2 2 3 3 3 4 4 4 999 1999 2999 3999 4999 999 1999 2999 3999 4999 Time PAL PPA pobj PAL 2.29472323 3.06492978 3.56384267 3.94407754 4.19969175 1.30457645 2.28035658 2.92514554 3.42933318 3.84427349 PPA 1 1 1 1 1 2 2 2 2 2 2.29473468 3.06495844 3.56387453 3.94411131 4.19979828 1.30457676 2.28035671 2.92514506 3.42933324 3.84427356 1 1 1 1 1 2 2 2 2 2 03 16 46 1:49 3:27 08 28 58 2:14 3:05 08 43 2:01 4:51 13:36 14 1:20 5:59 13:27 25:45 Tables 7 and 8 not only show that the PAL method is very efficient to solve the real data problems, but also show that the PAL method is about 2 to 7 times faster than the PPA. Concluding remarks In this paper, we designed a PAL method and a hybrid method to solve log-determinant optimization problems. We established the rigorous convergence results based on the fundamental theoretical frameworks. Extensive numerical experiments conducted on both synthetic problems and real data problems demonstrated that our methods are very robust and efficient. Based on the results of the numerical experiments, we conclude that for general problems, we prefer to adopt the hybrid method; but for some ill-conditioned problems, the PAL method may be a better choice. Acknowledgements I sincerely appreciate the Institute for Mathematical Sciences, National University of Singapore for supporting me to visit the institute and attend the workshop “Optimization: Computation, Theory and Modeling” in 2012 so that I can have a good opportunity to have fruitful discussions with Professors Defeng Sun and Kim-Chuan Toh on the PAL method. I also appreciate Dr. Xinyuan Zhao in Beijing University of Technology for many discussions on this topic. 23 References [1] F. Alizadeh, J. P. A. Haeberly, and O. L. Overton, Complementarity and nondegeneracy in semidefinite programming, Math. Programming, 77 (1997), 111–128. [2] O. Banerjee, L. El Ghaoui, A. d’Aspremont, Model selection through sparse maximum likelihood estimation, J. Mach. Learn. Res., 9 (2008), 485–516. [3] J. F. Bonnans and A. Shapiro, Perturbation Analysis of Optimization Problems, Springer, New York, 2000. [4] J. Dahl, L. Vandenberghe, and V. Roychowdhury, Covariance selection for nonchordal graphs via chordal embedding, Optim. Methods Softw., 23 (2008), 501–520. [5] A. Dempster, Covariance selection, Biometrics, 28 (1972), 157–175. [6] A. d’Aspremont, O. Banerjee, and L. El Ghaoui, First-order methods for sparse covariance selection, SIAM J. Matrix Anal. Appl., 30 (2008), 56–66. [7] M. Fazel, T.-K. Pong, D. Sun, and P. Tseng, Hankel matrix rank minimization with applications to system identification and realization, SIAM J. Matrix Anal. Appl., 34 (2013), 946–977. [8] Y. Gao and D. Sun, Calibrating least squares semidefinite programming with equality and inequality constraints, SIAM J. Matrix Anal. Appl., 31 (2009), 1432–1457. [9] R. W. Freund and N. M. Nachtigal, A new Krylov subspace method for symmetric indefinite linear systems, ORNL/TM-12754, 1994. [10] R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, Cambridge, UK, 1985. [11] R. A. Horn and C. R. Johnson, Topics in Matrix Analysis, Cambridge University Press, Cambridge, UK, 1991. [12] T. R. Hughes, M. J. Marton, A. R. Jones, C. J. Roberts, R. Stoughton, C. D. Armour, H. A. Bennett, E. Coffey, H. Dai, Y. D. He, M. J. Kidd, A. M. King, M. R. Meyer, D. Slade, P. Y. Lum, S. B. Stepaniants, D. D. Shoemaker, D. Gachotte, K. Chakraburtty, J. Simon, M. Bard, and S. H. Friend, Functional discovery via a compendium of expression profiles, Cell, 102 (2000), 109–126. [13] Z. Hu, J. Cao, and L. J. Hong, Robust simulation of global warming policies using the DICE model, Manage. Sci., 58 (2012), 1–17. [14] K. F. Jiang, D. F. Sun, and K.-C. Toh, An inexact accelerated proximal gradient method for large scale linearly constrained convex SDP, SIAM J. Optim., 22 (2012), 1042–1064. 24 [15] L. Li, and K.-C. Toh,, An inexact interior point method for L1-regularized sparse covariance selection, Math. Program. Comput., 2 (2010), 291–315. [16] Z. Lu, Smooth optimization approach for sparse covariance selection, SIAM J. Optim., 19 (2009), 1807–1827. [17] Z. Lu, Adaptive first-order methods for general sparse inverse covariance selection, SIAM J. Matrix Anal. Appl., 31 (2010), 2000–2016. [18] Z. Lu and Y. Zhang, Penalty decomposition methods for L0-norm minimization, Proceedings of Neural Information Processing Systems (NIPS), 2011: 46–54. [19] B. Martinet, Regularisation d’in´equations variationelles par approximations successives, Rev. Fran¸caise d’Informat. Recherche Op´erationnelle, (1970), 154–159. [20] F. Meng, D. Sun, and G. Zhao, Semismoothness of solutions to generalized equations and the Moreau–Yosida regularization, Math. Programming, 104 (2005), 561–581. [21] G. J. Minty, On the monotonicity of the gradient of a convex function, Pacific J. Math., 14 (1964), 243–247. [22] J. J. Moreau, Proximit´e et dualit´e dans un espace Hilbertien, Bull. Soc. Math. France, 93 (1965), 273–299. [23] G. Natsoulis, C. I. Pearson, J. Gollub, B. P. Eynon, J. Ferng, R. Nair, R. Idury, M. D. Lee, M. R. Fielden, R. J. Brennan, A. H. Roter and K. Jarnagin, The liver pharmacological and xenobiotic gene response repertoire, Mol. Syst. Biol., 175 (2008), 1–12. [24] P. Olsen, F. Oztoprak, J. Nocedal, and S. Rennie, Newton-like methods for sparse inverse covariance estimation, available at http://www.optimizationonline.org/DB HTML/2012/06/3506.html. [25] H. Qi and D. Sun, A quadratically convergent Newton method for computing the nearest correlation matrix, SIAM J. Matrix Anal. Appl., 28 (2006), 360–385. [26] R. T. Rockafellar, Convex Analysis, Princeton University Press, Princeton, 1970. [27] R. T. Rockafellar, A dual approach to solving nonlinear programming problems by unconstrained optimization, Math. Programming 5 (1973), 354–373. [28] R. T. Rockafellar, Monotone operators and the proximal point algorithm, SIAM J. Control Optim., 14 (1976), 877–898. [29] R. T. Rockafellar, Augmented Lagrangains and applications of the proximal point algorithm in convex programming, Math. Oper. Res., 1 (1976), 97–116. 25 [30] K. Scheinberg, S. Ma, and D. Goldfarb, Sparse inverse covariance selection via alternating linearization methods, Twenty-Fourth Annual Conference on Neural Information Processing Systems (NIPS), 2010: 2101–2109. [31] K. Scheinberg and I. Rish, Learning sparse Gaussian Markov networks using a greedy coordinate ascent approach, in Machine Learning and Knowledge Discovery in Databases, Lecture Notes in Comput. Sci. 6323, J. L. Balcazar, F. Bonchi, A. Gionis, and M. Sebag, eds., Springer-Verlag, Berlin, 2010, 196–212. [32] D. Sun, The strong second order sufficient condition and constraint nondegeneracy in nonlinear semidefinite programming and their implications, Math. Oper. Res., 31 (2006), 761–776. [33] D. Sun, J. Sun, and L. Zhang, The rate of convergence of the augmented Lagrangian method for nonlinear semidefinite programming, Math. Programming, 114 (2008), 349–391. [34] K.-C. Toh, Primal-dual path-following algorithms for determinant maximization problems with linear matrix inequalities, Comput. Optim. Appl., 14 (1999), 309–330. [35] K.-C. Toh, An inexact primal-dual path following algorithm for convex quadratic SDP, Math. Programming, 112 (2008), 221–254. [36] R. H. T¨ ut¨ unc¨ u, K.-C. Toh, and M. J. Todd, Solving semidefinite-quadratic-linear programs using SDPT3, Math. Programming, 95 (2003), 189–217. [37] K.-C. Toh, R. H. T¨ ut¨ unc¨ u, and M. J. Todd, Inexact primal-dual path-following algorithms for a special class of convex quadratic SDP and related problems, Pac. J. Optim., 3 (2007), 135–164. [38] N.-K. Tsing, M. K. H. Fan, and E. I. Verriest, On analyticity of functions involving eigenvalues, Linear Algebra Appl. 207 (1994), 159–180. [39] B. Varadarajan, D. Povey, and S. M. Chu, Quick fmllr for speaker adaptation in speech recognition, Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on. [40] C. Wang, D. Sun, and K.-C. Toh, Solving log-determinant optimization problems by a Newton-CG primal proximal point algorithm, SIAM J. Optim., 20 (2010), 2994–3013. [41] J. Yang, D. Sun, and K.-C. Toh, A proximal point algorithm for log-determinant optimization with group Lasso regularization, SIAM J. Optim., 23 (2013), 857–893. [42] K. Yosida, Functional Analysis, Springer Verlag, Berlin, 1964. 26 [43] S. Yang, X. Shen, P. Wonka, Z. Lu, and J. Ye, Fused multiple graphical Lasso, available at http://people.math.sfu.ca/∼zhaosong/ResearchPapers/FMGL.pdf. [44] X. Yuan, Alternating direction methods for sparse covariance selection, J. Sci. Comput., 51 (2012), 261–273. [45] X.-Y. Zhao, A Semismooth Newton-CG Augmented Lagrangian Method for Large Scale Linear and Convex Quadratic SDPs, PhD thesis, National University of Singapore, 2009. [46] X.-Y. Zhao, D. Sun, and K.-C. Toh, A Newton-CG augmented Lagrangian method for semidefinite programming, SIAM J. Optim., 20 (2010), 1737–1765. 27