Key words

1 Introduction

In this paper we consider the following linear complementarity problem:

$$\begin{array}{rcl} s& =& Mx + q, \\ xs& =& 0, \\ x,s& \geq & 0, \end{array}$$
(1)

where M ∈ R n ×n is a P (κ) matrix and q, x, s are vectors of R n, and xs denotes the componentwise product (Hadamard product) of vectors x and s. Linear complementarity problems have many applications in mathematical programming and equilibrium problems. Indeed, it is known that by exploiting the first-order optimality conditions of the optimization problem, any differentiable convex quadratic program can be formulated into a monotone linear complementarity problem, i.e. P (0) LCP, and vice versa [16]. Variational inequality problems are widely used in the study of equilibrium in economics, transportation planning, and game theory, and have a close connection to the LCPs. The reader can refer to Sect. 5.9 in [5] for the basic theory, algorithms, and applications.

The primal-dual IPM for linear optimization (LO) problems was first introduced in [9, 12] and extended to various class of problems, e.g., [3, 14]. Kojima et al. [9] and Monteiro et al. [12] first proved the polynomial computational complexity of the algorithm for LO problem independently, and since then many other algorithms have been developed based on the primal-dual strategy. Kojima et al. [10] proved the existence of the central path for any P (κ) LCP, generalized the primal-dual interior-point algorithm in [9] to P (κ) LCP and proved the same complexity results. Miao [11] extended the Mizuno-Todd-Ye predictor-corrector method to P (κ) LCPs. His algorithm uses the l 2-neighborhood of the central path and has \(O\left(\left(1 + \kappa \right)\sqrt{n}L\right)\) iteration complexity. Ills and Nagy [8] give a version of the Mizuno-Todd-Ye predictor-corrector interior point algorithm for the P (κ) LCP and show that the complexity of the algorithm is \(O\left({\left(1 + \kappa \right)}^{\frac{3} {2} }\sqrt{n}L\right).\) They choose τ and τ neighborhood parameters in such a way that at each iteration a predictor step is followed by one corrector step. For larger value of κ the values of τ and τ decrease fast, therefore the constant in the complexity results is increasing.

Most of the polynomial-time interior point algorithms for LO are based on the use of the logarithmic barrier function [9, 15]. Peng et al. [14] introduced self-regular barrier functions for primal-dual interior-point methods (IPMs) for LO, semidefinite optimization (SDO), second order cone optimization (SOCO) and also extended to P (κ) LCPs. Recently in [2, 7] the authors proposed a new primal-dual IPM for LO based on a new class of kernel functions which are not logarithmic and not necessarily self-regular barrier functions.

In this paper we propose a new large-update primal-dual IPM which generalizes the results obtained in [7] to P (κ) LCPs. We use a new search direction based on kernel functions which are neither logarithmic nor self-regular barrier. The new analysis which is derived in this paper is different from the one used in early papers [8, 10, 11, 14]. Furthermore, our analysis provides a simpler way to analyze primal dual IPMs.

We use the following notational conventions. Throughout the paper, \(\left\Vert \cdot \right\Vert\) denotes the 2-norm of a vector. The nonnegative orthant and positive orthant are denoted as R  +  n and \({\mathbf{R}}_{++}^{n}\), respectively. If z ∈ R  +  n and \(f :{ \mathbf{R}}_{+} \rightarrow {\mathbf{R}}_{+}\), then \(f\left(z\right)\) denotes the vector in R  +  n whose ith component is \(f\left({z}_{i}\right)\), with 1 ≤ i ≤ n. Finally, for x ∈ R n, X = diag (x) is the diagonal matrix from vector x, and \(J = \left\{1,2,\ldots,n\right\}\) is the index set.

This paper is organized as follows. In Sect. 2 we recall basic concepts and the notion of the central path. In Sect. 3 we review known results relevant for the development of the analysis. Section 4 contains new results to compute the feasible step size and the study of the amount of decrease of the proximity function during an inner iteration. Section 5 combines the results from Sect. 3 and the derived results in Sect. 4 to show the bound for the total number of iterations of the algorithm. Finally, concluding remarks are given in Sect. 6.

2 Preliminaries

In this section we introduce the definition of P (κ) matrix and its proprieties [10].

Definition 1.

Let Y be an open convex subset of R n and κ ≥ 0. A matrix M ∈ R n ×n is called a P (κ)-matrix on Y if and only if

$$\left(1 + 4\kappa \right) \sum\limits_{i\in {J}_{+}(x)}{x}_{i}{\left(Mx\right)}_{i} + \sum\limits_{i\in {J}_{-}(x)}{x}_{i}{\left(Mx\right)}_{i} \geq 0,$$

for all x ∈ Y, where

$${J}_{+}(x) = \left\{i \in J : {x}_{i}{\left(Mx\right)}_{i} \geq 0\right\}\quad and \quad {J}_{-}(x) = \left\{i \in J : {x}_{i}{\left(Mx\right)}_{i} < 0\right\}.$$

Further, M is called a P -matrix if it is a P (κ)-matrix for some κ ≥ 0.

Note that the class of P -matrices is the union of all P (κ)-matrices for κ ≥ 0, and contains the class of positive semi-definite matrices, i.e. symmetric matrices M satisfying ∑ i ∈ J x i (Mx) i  ≥ 0 for all x ∈ R n, by choosing κ = 0. The class of P  matrices also contains matrices with all positive principal minors. In the following we recall some results which are essential in our analysis.

Proposition 1 (Lemma 4.1 in [10]). 

If M ∈ R n×n is a P matrix, then

$$M^{\prime} = \left(\begin{array}{cc} - M & I\\ S &X\\ \end{array} \right)$$

is a nonsingular matrix for any positive diagonal matrices X, SR n×n .

We use the following corollary of Proposition 1 to prove that the modified Newton system has a unique solution.

Corollary 1.

Let M ∈ R n×n be a P matrix and x, s \(\in {\mathbf{R}}_{++}^{n}\). Then for all aR n the system

$$\begin{array}{rcl} - M\bigtriangleup x + \bigtriangleup s& =& 0, \\ S\bigtriangleup x + X\bigtriangleup s& =& a, \\ \end{array}$$

has a unique solution \(\left(\bigtriangleup x,\bigtriangleup s\right)\) .

The basic idea of primal-dual interior-point methods is to replace the second equation in (1) by the nonlinear equation xs = μe, where e is the all-one vector, and μ > 0. Thus we have the following parameterized system:

$$\begin{array}{rcl} s& =& Mx + q, \\ xs& =& \mu e, \\ x& \geq & 0,\ \ s \geq 0,\end{array}$$
(2)

where μ > 0. We assume that there exists strictly positive x and s that satisfy (1).

Since M is a P (κ) matrix and (1) is strictly feasible, then the parameterized system (2) has a unique solution \(\left(x(\mu ),s(\mu )\right)\) for each μ > 0. \(\left(x(\mu ),s(\mu )\right)\) is called μ-center of (2), the set of μ-centers (μ > 0) defines a homotopy path, which is called the central path of (2). If μ → 0 the limit of the central path exists. This limit satisfies the complementarity condition and belongs to the solution set of (1) [10].

Let (x, s) be a strictly feasible point and μ > 0. We define the vector

$$v := \sqrt{\frac{xs} {\mu }}.$$
(3)

Note that the pair (x, s) coincides with the μ-center (x(μ), s(μ)) if and only if v = e. Let Ψ(v) be a smooth, strictly convex function defined for all v > 0, which is minimal at v = e, with Ψ(e) = 0. Following [1, 2, 4, 6, 14] we define search directions Δx, Δs by

$$\begin{array}{ll} -M \Delta x + \Delta s = 0, \\ s\Delta x + x\Delta s = -\mu {v} \nabla \Psi (v).\end{array}$$
(4)

Since M is a P  matrix, the system (4) uniquely defines (Δx, Δs) for any x > 0 and s > 0. Note that Δx = 0, Δs = 0, if and only if v = e, because the right-hand sides in (4) vanish if and only if ∇ Ψ(v) = 0, and this occurs if and only if v = e.

Let (x, s) be a strictly feasible point. We define the vector p by

$$p := \sqrt{\frac{x} {s}}.$$
(5)

Introducing the following notations

$$\bar{M} := PMP\mbox{ and }P := \mathrm{diag}\,(p),\,V := \mathrm{diag}\,(v)\mbox{ where }v = \sqrt{\frac{xs} {\mu }},$$
(6)

and

$${d}_{x} := \frac{v\Delta x} {x},\,\qquad {d}_{s} := \frac{v\Delta s} {s},$$
(7)

system (4) can be reformulated as

$$\begin{array}{rcl} -\bar{ M}{d}_{x} + {d}_{s}& =&0, \\ {d}_{x} + {d}_{s}& =& -\nabla \Psi (v).\end{array}$$
(8)

From the solution d x and d s , the vectors Δx and Δs can be computed from (7).

Note that the vectors d x and d s are not orthogonal. So our analysis in this paper will deviate significantly from the analysis used for LO in [7].

The algorithm considered in this paper is described in Fig. 24.1.

The inner while loop in the algorithm is called inner iteration and the outer while loop outer iteration. So each outer iteration consists of an update of the barrier parameter and a sequence of one or more inner iterations. We assume that (1) is strictly feasible, and the starting point \(\left({x}^{0},{s}^{0}\right)\) is strictly feasible for (1). Choose τ and \({v}^{0} = \sqrt{\frac{{x}^{0 } {s}^{0 } } {{\mu }^{0}}}\) initial strictly feasible point such that \(\Psi \left({v}^{0}\right) \leq \tau \) where τ is threshold value in Fig. 24.1. We then decrease μ to \(\mu := (1 - \theta )\mu \), for some θ ∈ (0, 1). In general this will increase the value of Ψ(v) above τ. To get this value smaller again, and coming closer to the current μ-center, we solve the scaled search directions from (8), and unscaled these directions by using (4). By choosing an appropriate step size α, we move along the search direction, and construct a new pair \(({x}_{+},{s}_{+})\) with

$$\begin{array}{rcl}{ x}_{+} = x + \alpha \bigtriangleup x\quad {s}_{+} = s + \alpha \bigtriangleup s.& &\end{array}$$
(9)

If necessary, we repeat the procedure until we find iterates such that Ψ(v) no longer exceed the threshold value τ, which means that the iterates are in a small enough neighborhood of (x(μ), s(μ)). Then μ is again reduced by the factor 1 − θ and we apply the same procedure targeting at the new μ-centers. This process is repeated until μ is small enough, i.e. until nμ ≤ ε. At this stage we have found an ε-solution of (1). Just as in the LO case, the parameters τ, θ, and the step size α should be chosen in such a way that the algorithm is ‘optimized’ in the sense that the number of iterations required by the algorithm is as small as possible. Obviously, the resulting iteration bound will depend on the kernel function underlying the algorithm, and our main task becomes to find a kernel function that minimizes the iteration bound.

Table 1 gives some examples of kernel functions that have been analyzed in [6] with the complexity results for the corresponding algorithms.

Table 1 Examples of kernel functions studied in early paper [6] with complexity results.

The aim of this paper is to investigate a new kernel function studied first in linear optimization case in [7], namely

$$\psi (t) = \frac{{t}^{2} - 1} {2} + \frac{6} {\pi }\tan \left(h(t)\right),\quad \mbox{ with}\quad h(t) = \frac{\pi \left(1 - t\right)} {4t + 2},$$
(10)

and to show that the interior-point methods for linear complementarity based on these function have favorable complexity results.

Note that the growth term of our kernel function is quadratic as all kernel functions in Table 1. However, this function (10) deviates from all other kernel functions [6] since its barrier term is trigonometric as \(\frac{6} {\pi }\tan \frac{\pi \left(1-t\right)} {4t+2}\). In order to study the new kernel function, several new arguments had to be developed for the analysis.

3 Properties of the New Proximity Function

This section is started by technical lemma, and then some properties of the new kernel function introduced in this paper are derived.

3.1 Some Technical Results

The first three derivatives of ψ are given by

$$\begin{array}{rcl} \psi ^{\prime}(t)& =& t + \frac{6h^{\prime}(t)} {\pi } \left(1 {+\tan }^{2}(h(t))\right),\,\end{array}$$
(11)
$$\begin{array}{rcl}{ \psi }^{{\prime}{\prime}}(t)& =& 1 + \frac{6} {\pi }\left(1 {+\tan }^{2}(h(t))\right)\left(h^{\prime\prime}(t) + 2h^{\prime}{(t)}^{2}\tan (h(t))\right).\end{array}$$
(12)
$$\begin{array}{rcl}{ \psi }^{{\prime}{\prime}{\prime}}(t)& =& \frac{6} {\pi }\left(1 {+\tan }^{2}(h(t))\right)k(t),\end{array}$$
(13)

with

$$\begin{array}{rcl} k(t) := 6h^{{\prime}{\prime}}(t)h^{\prime}(t)\tan (h(t)) + h^{{\prime}{\prime}{\prime}}(t) + 2h^{\prime}{(t)}^{3}\left({3\tan }^{2}(h(t)) + 1\right).& &\end{array}$$
(14)

The next lemma serves to prove that the new kernel function (10) is eligible.

Lemma 1 (Lemma 2 in [7]). 

Let ψ be as defined in (10) and t > 0. Then,

$$\begin{array}{rcl}\psi ^{{\prime}{\prime}}(t)& >& 1,\,\quad \\\end{array}$$
(15a)
$$\begin{array}{rcl} t\psi ^{{\prime}{\prime}}(t) + \psi ^{\prime}(t)& >& 0,\, \\ \end{array}$$
(15b)
$$\begin{array}{rcl} t\psi ^{{\prime}{\prime}}(t) - \psi ^{\prime}(t)& >& 0,\,\quad\\ \end{array}$$
(15c)
$$\begin{array}{rcl}\mbox{ and}\quad \ \psi ^{{\prime}{\prime}{\prime}}(t)& <& 0.\quad\\\end{array}$$
(15d)

It follows that \(\psi (1) = \psi ^{\prime}(1) = 0\) and ψ(t) ≥ 0, proving that ψ is defined by ψ(t).

$$\psi (t) ={ \int }_{1}^{t}{ \int }_{1}^{\xi }\psi ^{{\prime}{\prime}}(\zeta )\,\mathrm{d}\zeta \mathrm{d}\xi.$$
(16)

The second property (15b) in Lemma 1 is related to Definition 2.1.1 and Lemma 2.1.2 in [14]. This property is equivalent to convexity of the composed function z↦ψ(e z) and this holds if and only if \(\psi (\sqrt{{t}_{1 } {t}_{2}}) \leq \frac{1} {2}\left(\psi ({t}_{1}) + \psi ({t}_{2})\right)\) for any t 1, t 2 ≥ 0. Following [1], we therefore say that ψ is exponentially convex, or shortly, e-convex, whenever t > 0.

Lemma 2.

Let ψ be as defined in (10), one has

$$\psi (t) < \frac{1} {2}\psi ^{{\prime}{\prime}}(1){\left(t - 1\right)}^{2}, \mbox{ if} t >1.$$

Proof.

By Taylor’s theorem and \(\psi (1) = \psi ^{\prime}(1) = 0,\) we obtain

$$\psi (t) = \frac{1} {2}\psi ^{{\prime}{\prime}}(1){\left(t - 1\right)}^{2} + \frac{1} {6}\psi ^{{\prime}{\prime}{\prime}}(\xi ){\left(\xi - 1\right)}^{3},$$

where 1 < ξ < t if t > 1. Since ψ(ξ) < 0, the lemma follows.

Lemma 3.

Let ψ be as defined in (10), one has

$$t\psi ^{\prime}(t) \geq \psi (t), \mbox{ if}\ t \geq 1.$$

Proof.

Defining \(g(t) := t\psi ^{\prime}(t) - \psi (t)\) one has g(1) = 0 and g′(t) = tψ′′(t) ≥ 0. Hence g(t) ≥ 0 and the lemma follows.

Following [2], we now introduce a norm-based proximity measure δ(v), according to

$$\delta (v) := \frac{1} {2}\|\psi ^{\prime}(v)\| = \frac{1} {2}\sqrt{ \sum\limits_{i=1}^{n}\psi ^{\prime}{({v}_{i})}^{2}} = \frac{1} {2}\left\Vert {d}_{x} + {d}_{s}\right\Vert,$$
(17)

in terms of Ψ(v). Since Ψ(v) is strictly convex and attains its minimal value zero at v = e, we have

$$\Psi \left(v\right) = 0 \Leftrightarrow \delta \left(v\right) = 0 \Leftrightarrow v = e.$$

3.2 Relations Between Proximity Measure and Norm-Based Proximity Measure

For the analysis of the algorithm in Sect. 4 we need to establish relations between Ψ(v) and δ(v). A crucial observation is that the inverse function of ψ(t), for t ≥ 1, plays an important role in this relation.

The next theorem, which is one of main results in [2], gives a lower bound on δ(v) in term of Ψ(v). This is due to the fact that ψ(t) satisfies (15d). The theorem is a special case of Theorem 4.9 in [2], and is therefore stated without proof.

We denote by ρ : [0, ) → [1, ) and ρ : [0, ) → (0, 1] the inverse functions of ψ(t) for t ≥ 1, and \(-\frac{1} {2}\psi ^{\prime}(t)\) for t ≤ 1, respectively. In other words

$$s = \psi (t)\quad \Leftrightarrow \quad t = \rho (s),\,\quad t \geq 1,\,$$
(18)
$$s = -\frac{1} {2}\psi ^{\prime}(t)\quad \Leftrightarrow \quad t = \rho (s),\,\quad t \leq 1.$$
(19)

Theorem 1 (Theorem 4.9 in [2]). 

Let ϱ be as defined in (18). .One has

$$\delta (v) \geq \frac{1} {2}\psi ^{\prime}\left(\rho \,\left(\Psi (v)\right)\right).$$

Corollary 2.

Let ϱ be as defined in (18). Thus we have

$$\delta (v) \geq \frac{\Psi (v)} {2\rho \left(\Psi (v)\right)}.$$

Proof.

Using Theorem 1, i.e., \(\delta (v) \geq \frac{1}{2}\psi^\prime(\rho(\Psi(v)))\) , we obtain from Lemma 3

$$\delta (v) \geq \frac{\psi \left(\rho (\Psi (v))\right)} {2\rho \left(\Psi (v)\right)} = \frac{\Psi (v)} {2\rho \left(\Psi (v)\right)}.$$

This proves the corollary.

Theorem 2.

If Ψ(v) ≥ 1, then

$$\delta (v) \geq \frac{1} {6}{\Psi }^{\frac{1} {2} }.$$
(20)

Proof.

The inverse function of ψ(t) for t ∈ [1, ) is obtained by solving t from

$$\psi (t) = \frac{{t}^{2} - 1} {2} + \frac{6} {\pi }\tan \frac{\pi \left(1 - t\right)} {4t + 2} = s, t \geq 1.$$

We derive an upper bound for t, as this suffices for our goal. One has from (16) and ψ′′(t) ≥ 1,

$$s = \psi (t) ={ \int }_{1}^{t}{ \int }_{1}^{\xi }\psi ^{{\prime}{\prime}}(\zeta )\,\mathrm{d}\zeta \mathrm{d}\xi \geq {\int }_{1}^{t}{ \int }_{1}^{\xi }\mathrm{d}\zeta \mathrm{d}\xi = \frac{1} {2}{(t - 1)}^{2},$$

which implies

$$t = \rho \,(s) \leq 1 + \sqrt{2s}.$$
(21)

Assuming s ≥ 1, we get \(t = \rho \,(s) \leq \sqrt{s} + \sqrt{2s} \leq 3{s}^{\frac{1} {2} }.\) Omitting the argument v, and assuming Ψ(v) ≥ 1, we have \(\rho (\Psi (v)) \leq 3\Psi {(v)}^{\frac{1} {2} }.\) Now, using Corollary 2, we have

$$\delta (v) \geq \frac{\Psi (v)} {2\rho \left(\Psi (v)\right)} \geq \frac{1} {6}\Psi {(v)}^{\frac{1} {2} }.$$

This proves the lemma.

Note that if \(\Psi (v) \geq 1,\) substitution in (20) gives

$$\delta (v) \geq \frac{1} {6}.$$
(22)

3.3 Growth Behavior of the Barrier Function

Note that at the start of each outer iteration of the algorithm, just before the update of μ with the factor 1 − θ, we have Ψ(v) ≤ τ. Due to the update of μ the vector v is divided by the factor \( \sqrt{1-\theta}\), with 0 < θ < 1, which in general leads to an increase in the value of Ψ(v). Then, during the subsequent inner iterations, Ψ(v) decreases until it passes the threshold τ again. Hence, during the course of the algorithm the largest values of Ψ(v) occur just after the updates of μ. In this section we derive an estimate for the effect of a μ-update on the value of Ψ(v). We start with an important theorem which is valid for all kernel functions ψ(t) that are strictly convex (15a), and satisfies (15c).

Theorem 3 (Theorem 3.2 in [2]). 

Let ϱ : [0,∞) → [1,∞) be the inverse function of ψ on [0,∞). Then for any positive vector v and any β > 1 we have:

$$\Psi (\beta v) \leq n\psi \left(\beta \rho \left(\frac{\Psi (v)} {n} \right)\right).$$
(23)

Corollary 3.

Let 0 < θ < 1 and \({v}_{+} = \frac{v} {\sqrt{1-\theta }}.\) Then

$$\begin{array}{rcl} \Psi ({v}_{+}) \leq n\psi \left(\frac{\rho \left(\frac{\Psi (v)} {n} \right)} {\sqrt{1 - \theta }}\right).& &\end{array}$$
(24)

Proof.

Substitution of \(\beta = \frac{1} {\sqrt{1-\theta }}\) into (23), the corollary is proved.

Suppose that the barrier update parameter θ and threshold value τ are given. According to the algorithm, at the start of each outer iteration we have Ψ(v) ≤ τ. By Theorem 3, after each μ-update the growth of Ψ(v) is limited by (24). Therefore we define

$$\begin{array}{rcl} L = L(n,\theta,\tau ) := n\psi \left( \frac{\rho \left(\frac{\tau } {n}\right)} {\sqrt{1 - \theta }}\right).& &\end{array}$$
(25)

Obviously, L is an upper bound of Ψ(v  + ), the value of Ψ(v) after the μ-update.

4 Analysis of the Algorithm

In this section, we show how to compute a feasible step size α of a Newton step with the decrease of the barrier function. Since d x and d s are not orthogonal the analysis in this paper is different from that of LO case. After a damped step, with step size α, using (3) and (7) we have

$${x}_{+} = x + \alpha \Delta x = \frac{x} {v}\left(v + \alpha {d}_{x}\right),\quad {s}_{+} = s + \alpha \Delta s = \frac{s} {v}\left(v + \alpha {d}_{s}\right).$$

Thus we obtain

$${v}_{+}^{2} = \frac{{x}_{+}{s}_{+}} {\mu } = \left(v + \alpha {d}_{x}\right)\left(v + \alpha {d}_{s}\right).$$
(26)

In the sequel we use the following notation:

$$\begin{array}{rcl} u :{=\min }_{i\in J}{v}_{i},\,\quad \delta := \delta (v),\,\quad {\sigma }_{+} := \sum\limits_{{i\in {J}_{+}}{d}_{{x}_{i}}{d}_{{s}_{i}}},\,\quad {\sigma }_{-} := - \sum\limits_{i\in {J}_{-}}{d}_{{x}_{i}}{d}_{{s}_{i}}.& &\end{array}$$
(27)

Since M is a P (κ) matrix, we have

$$(1 + 4\kappa ) \sum\limits_{i\in {J}_{+}}\Delta {x}_{i}{(M\Delta x)}_{i} + \sum\limits_{i\in {J}_{-}}\Delta {x}_{i}{(M\Delta s)}_{i} \geq 0,$$

where \({J}_{+} = \left\{i \in J : \Delta {x}_{i}{(M\Delta x)}_{i} \geq 0\right\},\) \({J}_{-} = J - {J}_{+}\). Using the first equation in (4) we have for Δx ∈ R n, MΔx = Δs, and

$$(1 + 4\kappa ) \sum\limits_{i\in {J}_{+}}\Delta {x}_{i}\Delta {s}_{i} + \sum\limits_{i\in {J}_{-}}\Delta {x}_{i}\Delta {s}_{i} \geq 0.$$

From (7) it follows that \({d}_{x}{d}_{s} = \frac{{v}^{2}\Delta x\Delta s} {xs} = \frac{\Delta x\Delta s} {\mu }\) with μ > 0, and

$$\begin{array}{rcl} \left(1 + 4\kappa \right) \sum\limits_{i\in {J}_{+}}{d}_{{x}_{i}}{d}_{{s}_{i}} + \sum\limits_{i\in {J}_{-}}{d}_{{x}_{i}}{d}_{{s}_{i}} = \left(1 + 4\kappa \right){\sigma }_{+} - {\sigma }_{-}\geq 0.& &\end{array}$$
(28)

The next lemma gives an upper bound of σ +  and σ − 

Lemma 4.

One has

$${\sigma }_{+} \leq {\delta }^{2},\quad and \quad {\sigma }_{ -}\leq \left(1 + 4\kappa \right){\delta }^{2}.$$

Proof.

By definition of σ + , σ −  and δ, we have

$${\sigma }_{+} = \sum\limits_{i\in {J}_{+}}{d}_{{x}_{i}}{d}_{{s}_{i}} \leq \frac{1} {4} \sum\limits_{i\in {J}_{+}}{\left({d}_{{x}_{i}} + {d}_{{s}_{i}}\right)}^{2} \leq \frac{1} {4} \sum\limits_{i\in J}{\left({d}_{{x}_{i}} + {d}_{{s}_{i}}\right)}^{2} = \frac{1} {4}{\left\Vert {d}_{{x}_{i}} + {d}_{{s}_{i}}\right\Vert }^{2} = {\delta }^{2}.$$

Since M is a P (κ) matrix, using (28), we get

$$\left(1 + 4\kappa \right){\sigma }_{+} - {\sigma }_{-}\geq 0.$$

Thus

$${\sigma }_{-}\leq \left(1 + 4\kappa \right){\sigma }_{+} \leq \left(1 + 4\kappa \right){\delta }^{2}.$$

This proves the lemma.

The following lemma gives an upper bound for \(\left\Vert {d}_{x}\right\Vert\) and \(\left\Vert {d}_{s}\right\Vert\).

Lemma 5.

One has

$$ \sum\limits_{i=1}^{n}\left({d}_{{ x}_{i}}^{2} + {d}_{{ s}_{i}}^{2}\right) \leq 4\left(1 + 2\kappa \right){\delta }^{2},\quad \left\Vert {d}_{ x}\right\Vert \leq 2\sqrt{1 + 2\kappa }\ \delta,\quad and \quad \left\Vert {d}_{s}\right\Vert \leq 2\sqrt{1 + 2\kappa }\ \delta.$$

Proof.

From the definitions (17) and (27), we have

$$\delta = \frac{1} {2}\left\Vert {d}_{x} + {d}_{s}\right\Vert,\quad {and} \quad \sum\limits_{j\in J}{d}_{{x}_{i}}{d}_{{s}_{i}} = {\sigma }_{+} - {\sigma }_{-},$$

then

$$2\delta = \left\Vert {d}_{x} + {d}_{s}\right\Vert = \sqrt{ \sum\limits_{i=1}^{n}{\left({d}_{{x}_{i}} + {d}_{{s}_{i}}\right)}^{2}} = \sqrt{ \sum\limits_{i=1}^{n}\left({d}_{{x}_{i}}^{2} + {d}_{{s}_{i}}^{2}\right) + 2\left({\sigma }_{+} - {\sigma }_{-}\right)}.$$

Using (28), and Lemma 4, we get

$$2\delta \geq \sqrt{ \sum\limits_{i=1}^{n}\left({d}_{{x}_{i}}^{2} + {d}_{{s}_{i}}^{2}\right) + 2\left( \frac{1} {1 + 4\kappa }{\sigma }_{-}- {\sigma }_{-}\right)} = \sqrt{ \sum\limits_{i=1}^{n}\left({d}_{{x}_{i}}^{2} + {d}_{{s}_{i}}^{2}\right) - \frac{8\kappa } {1 + 4\kappa }{\sigma }_{-}}.$$

Then, we get

$$4{\delta }^{2} + \frac{8\kappa } {1 + 4\kappa }{\sigma }_{-}\geq \sum\limits_{i=1}^{n}\left({d}_{{ x}_{i}}^{2} + {d}_{{ s}_{i}}^{2}\right).$$

Using again Lemma 4, we have

$$4\left(1 + 2\kappa \right){\delta }^{2} \geq 4{\delta }^{2} + \frac{8\kappa } {1 + 4\kappa }{\sigma }_{-}\geq \sum\limits_{i=1}^{n}\left({d}_{{ x}_{i}}^{2} + {d}_{{ s}_{i}}^{2}\right).$$

Thus

$$\left\Vert {d}_{x}\right\Vert \leq \sqrt{ \sum\limits_{i=1}^{n}\left({d}_{{x}_{i}}^{2} + {d}_{{s}_{i}}^{2}\right)} \leq 2\sqrt{1 + 2\kappa }\ \delta.$$

Using the same argument, we can prove that

$$\left\Vert {d}_{s}\right\Vert \leq 2\sqrt{1 + 2\kappa }\ \delta.$$

Thus the lemma follows.

Our aim is to find an upper bound for

$$f(\alpha ) := \Psi \left({v}_{+}\right) - \Psi \left(v\right) := \Psi \left(\sqrt{\left(v + \alpha {d}_{x } \right) \left(v + \alpha {d}_{s } \right)}\right) - \Psi \left(v\right),$$

where Ψ : R n → R is given by

$$\begin{array}{rcl} \Psi (v) = \sum\limits_{i=1}^{n}\psi ({v}_{ i}).& &\end{array}$$
(29)

To do this, the next four technical lemmas are needed. It is clear that f(α) is not necessarily convex in α. To simplify the analysis we use a convex upper bound for f(α). Such a bound is obtained by using that ψ(t) satisfies the condition (15b). Hence, ψ(t) is e-convex. This implies

$$\Psi \left({v}_{+}\right) = \Psi \left(\sqrt{\left(v + \alpha {d}_{x } \right) \left(v + \alpha {d}_{s } \right)}\right) \leq \frac{1} {2}\left[\Psi \left(v + \alpha {d}_{x}\right) + \Psi \left(v + \alpha {d}_{s}\right)\right].$$

Thus we have f(α) ≤ f 1(α), where

$${f}_{1}(\alpha ) := \frac{1} {2}\left[\Psi \left(v + \alpha {d}_{x}\right) + \Psi \left(v + \alpha {d}_{s}\right)\right] - \Psi \left(v\right)$$

is a convex function of α, since Ψ(v) is convex. Obviously, \(f(0) = {f}_{1}(0) = 0.\) Taking the derivative of f 1(α) to α, we get

$${f}_{1}^{\prime}(\alpha ) = \frac{1} {2} \sum\limits_{i=1}^{n}\left(\psi ^{\prime}\left({v}_{ i} + \alpha {{d}_{x}}_{i}\right){{d}_{x}}_{i} + \psi ^{\prime}\left({v}_{i} + \alpha {{d}_{s}}_{i}\right){{d}_{s}}_{i}\right).$$

This gives, using last equation in (8) and (17),

$${f}_{1}^{\prime}(0) = \frac{1} {2}\nabla \Psi {(v)}^{T}\left({d}_{ x} + {d}_{s}\right) = -\frac{1} {2}\nabla \Psi {(v)}^{T}\nabla \Psi (v) = -2\delta {(v)}^{2}.$$
(30)

Differentiating once more, we obtain

$${f}_{1}^{{\prime}{\prime}}(\alpha ) = \frac{1} {2} \sum\limits_{i=1}^{n}\left(\psi ^{{\prime}{\prime}}\left({v}_{ i} + \alpha {{d}_{x}}_{i}\right){{d}_{x}}_{i}^{2} + \psi ^{{\prime}{\prime}}\left({v}_{ i} + \alpha {{d}_{s}}_{i}\right){{d}_{s}}_{i}^{2}\right).$$
(31)

From this stage on we can apply word-by-word the same arguments as in [6] to obtain the following results that are therefore stated without proof.

The following lemma gives an upper bound of f 1(α) in terms of δ and ψ′′t.

Lemma 6 (Lemma 4.3 in [6]). 

One has

$${f^{{\prime}{\prime}}}_{1}(\alpha ) \leq 2\left(1 + 2\kappa \right){\delta }^{2}\,\psi ^{{\prime}{\prime}}\left(v - 2\alpha \sqrt{1 + 2\kappa }\ \delta \right).$$

Putting

$${\delta }_{\kappa } := \sqrt{1 + 2\kappa }\ \delta,$$
(32)

we have

$$\begin{array}{rcl}{ f}_{1}^{{\prime}{\prime}}(\alpha ) \leq 2{\delta }_{\kappa }^{2}\,\psi ^{{\prime}{\prime}}\left(v - 2\alpha {\delta }_{ \kappa }\right),& &\end{array}$$
(33)

Since f 1(α) is convex, we will have f ′ 1(α) ≤ 0 for all α less than or equal to the value where f 1(α) is minimal, and vice versa. In this respect the next result is important.

Lemma 7 (Lemma 4.4 in [6]). 

One has f 1 ′(α) ≤ 0 if α satisfies the inequality

$$-\psi ^{\prime}\left(v - 2\alpha {\delta }_{\kappa }\right) + \psi ^{\prime}\left(v \right) \leq \frac{2{\delta }_{\kappa }} {\left(1 + 2\kappa \right)}.$$
(34)

The next lemma uses the inverse function ρ : [0, ) → (0, 1] of \(-\frac{1} {2}\psi ^{\prime}(t)\) for t ∈ (0, 1], as defined in (19).

Lemma 8 (Lemma 4.5 in [6]). 

The largest value of the step size α satisfying (33) is given by

$$\begin{array}{rcl} \bar{\alpha } := \frac{1} {2{\delta }_{\kappa }}\left[\rho \left(\delta \right) - \rho \left(\frac{1 + \sqrt{1 + 2\kappa }} {1 + 2\kappa } {\delta }_{\kappa }\right)\right].& &\end{array}$$
(35)

Moreover

$$\begin{array}{rcl} \bar{\alpha } \geq \frac{1} {\left(1 + 2\kappa \right)\psi ^{{\prime}{\prime}}\left(\rho \left(\frac{1+\sqrt{1+2\kappa }} {1+2\kappa } {\delta }_{\kappa }\right)\right)}.& &\end{array}$$
(36)

For future use we define

$$\widetilde{\alpha } := \frac{1} {\left(1 + 2\kappa \right)\psi ^{{\prime}{\prime}}\left(\rho \left(\frac{1+\sqrt{1+2\kappa }} {1+2\kappa } \ {\delta }_{\kappa }\right)\right)},$$
(37)

as the default step size. By Lemma 8 this step \(\widetilde{\alpha }\) satisfies (34). By (36) we have \(\bar{\alpha } \geq \tilde{ \alpha }\). We recall without proof the following lemma from [13].

Lemma 9 (Lemma 3.12 in [13]). 

Let h(t) be a twice differentiable convex function with h(0) = 0, h′(0) < 0 and let h(t) attain its (global) minimum at t > 0. If h′′(t) is increasing for t ∈ [0,t ], then

$$h(t) \leq \frac{th^{\prime}(0)} {2},\,\quad 0 \leq t \leq {t}^{{_\ast}}.$$

Lemma 10.

If the step size α satisfies (34), then

$$f(\alpha ) \leq -\alpha \,{\delta }^{2}.$$
(38)

Proof.

Let h(α) be defined by

$$h\left(\alpha \right) := -2\alpha {\delta }^{2} + \alpha {\delta }_{ \kappa }\psi ^{\prime}\left(v \right) -\frac{1} {2}\psi \left(v \right) + \frac{1} {2}\psi \left(v - 2\alpha {\delta }_{\kappa }\right).$$

Then

$$h(0) = {f}_{1}(0) = 0,\,\quad h^{\prime}(0) = {f^{\prime}}_{1}(0) = -2{\delta }^{2},\,\quad h^{{\prime}{\prime}}(\alpha ) = 2{\delta }_{ \kappa }^{2}\,\psi ^{{\prime}{\prime}}\left(v - 2\alpha {\delta }_{ \kappa }\right).$$

Due to Lemma 6, f′ 1(α) ≤ h′(α). As a consequence, f′ 1(α) ≤ h′(α) and f 1(α) ≤ h(α). Taking \(\alpha \leq \bar{\alpha }\), with \(\bar{\alpha }\) as defined in Lemma 8, we have

$$\begin{array}{rcl} h^{\prime}(\alpha )& =& -2{\delta }^{2} + 2{\delta }_{ \kappa }^{2}{ \int }_{0}^{\alpha }\,\psi ^{{\prime}{\prime}}\left(v - 2\xi {\delta }_{ \kappa }\right)\,\mathrm{d}\xi \\ & =& -2{\delta }^{2} - {\delta }_{ \kappa }\left(\psi ^{\prime}\left(v - 2\alpha {\delta }_{\kappa }\right) - \psi ^{\prime}\left(v \right)\right) \leq \end{array}$$
(0.)

Since h′′(α) is increasing in α, using Lemma 9, we may write

$${f}_{1}(\alpha ) \leq h(\alpha ) \leq \frac{1} {2}\alpha h^{\prime}(0) = -\alpha {\delta }^{2}.$$

Since f(α) ≤ f 1(α), the proof is complete.

Theorem 4.

Let ρ be defined in (19) and \(\widetilde{\alpha }\) in (37). Then

$$f(\widetilde{\alpha }) \leq - \frac{{\delta }^{2}} {\left(1 + 2\kappa \right)\psi ^{{\prime}{\prime}}\left(\rho \left(\frac{1+\sqrt{1+2\kappa }} {\sqrt{1+2\kappa }} \ \delta \right)\right)} \leq - \frac{{\delta }^{\frac{1} {2} }} {2593\left(1 + 2\kappa \right)},$$
(39)

Proof.

By combining (36) in Lemma 8 and results in Lemma 10, using also (32). Thus the first inequality in (39) follows.

To obtain the inverse function t = ρ(s) of \(-\frac{1} {2}\psi ^{\prime}(t)\) for t ∈ (0, 1], we need to solve t from the equation \(-\left(t + \frac{6h^{\prime}(t)} {\pi } \left(1 {+\tan }^{2}(h(t))\right)\right) = 2s.\) This implies

$$1 {+\tan }^{2}(h(t)) = \frac{-\pi } {6h^{\prime}(t)}\left(2s + t\right) = \frac{2\pi {\left(2t + 1\right)}^{2}} {18\pi } \left(2s + t\right) \leq 2s + 1\quad \mbox{ for}\quad t \leq 1.$$

Hence, putting \(t = \rho \left(\frac{1+\sqrt{1+2\kappa }} {\sqrt{1+2\kappa }} \ \delta \right),\) which is equivalent to \(\frac{2\left(1+\sqrt{1+2\kappa }\right)} {\sqrt{1+2\kappa }} \delta = -\psi ^{\prime}(t).\) Using that \(\frac{1+\sqrt{1+2\kappa }} {\sqrt{1+2\kappa }} \leq 2\) for all κ ≥ 0, we get

$$\tan (h(t)) \leq 2\sqrt{\delta }.$$
(40)

By (40), thus we have

$$\begin{array}{rcl} \widetilde{\alpha }& =& \frac{1} {\left(1 +2\kappa \right)\psi ^{{\prime}{\prime}}\left(\rho \left(\frac{1+\sqrt{1+2\kappa }} {\sqrt{1+2\kappa }} \ \delta \right)\right)} \\ & =& \frac{1} {\left(1 + 2\kappa \right)\left(1 +\frac{6} {\pi }\left(1 {+\tan }^{2}(h(t))\right)\left(h^{{\prime}{\prime}}(t) + 2h^{\prime}{(t)}^{2}\tan (h(t))\right)\right)} \\ & \geq & \frac{1} {\left(1 + 2\kappa \right)\left(1+ \frac{6} {\pi }\left(1 + 4\delta \right)\left(h^{{\prime}{\prime}}(t) + 4h^{\prime}{(t)}^{2}\sqrt{\delta }\right)\right)}.\end{array}$$

Since \(h^{{\prime}{\prime}}(t) = \frac{6\pi } {{\left(2t+1\right)}^{3}} \leq 6\pi \), and \(h^{\prime}{(t)}^{2} = \frac{9{\pi }^{2}} {4{\left(2t+1\right)}^{4}} \leq \frac{9{\pi }^{2}} {4}\) for all 0 ≤ t ≤ 1. Then we have

$$\begin{array}{rcl} \widetilde{\alpha }& \geq & \frac{1} {\left(1 + 2\kappa \right)\left(1 + \frac{6} {\pi }\left(1 + 4\delta \right)\left(6\pi + 9{\pi }^{2}\sqrt{\delta }\right)\right)} \\ & =& \frac{1} {\left(1 + 2\kappa \right)\left(1 + 18\left(1 + 4\delta \right)\left(2 + 3\pi \sqrt{\delta }\right)\right)}.\end{array}$$

Also using (22) (i.e., 6δ ≥ 1) we get,

$$\begin{array}{rcl} \widetilde{\alpha }& \geq & \frac{1} {\left(1 + 2\kappa \right)\left({\left(6\delta \right)}^{\frac{3} {2} } + 18\left(6\delta + 4\delta \right)\left(2\sqrt{6\delta } + 3\pi \sqrt{\delta }\right)\right)} \\ & =& \frac{1} {\left(1 + 2\kappa \right)\left(\left({6}^{\frac{3} {2} } + 180\left(2\sqrt{6} + 3\pi \right)\right){\delta }^{\frac{3} {2} }\right)} \geq \frac{1} {2593\left(1 + 2\kappa \right){\delta }^{\frac{3} {2} }}. \end{array}$$

Hence

$$f(\widetilde{\alpha }) \leq - \frac{{\delta }^{2}} {2593\left(1 + 2\kappa \right){\delta }^{\frac{3} {2} }} = - \frac{{\delta }^{\frac{1} {2} }} {2593\left(1 + 2\kappa \right)}.$$

Thus, the theorem follows.

Substitution in (20) gives

$$f(\tilde{\alpha }) \leq - \frac{{\delta }^{\frac{1} {2} }} {2593\left(1 + 2\kappa \right)} \leq - \frac{{\Psi }^{\frac{1} {4} }} {2593\sqrt{6}\left(1 + 2\kappa \right)} \leq - \frac{{\Psi }^{\frac{1} {4} }} {6532\left(1 + 2\kappa \right)}.$$

5 Iteration Complexity

In this section we derive the complexity bounds for large-update methods and small-update methods.

5.1 Upper Bound for the Total Number of Iterations

Let K denote the number of inner iterations. An upper bound for the total number of iterations is obtained by multiplying (the upper bound for) the number K by the number of barrier parameter updates, which is bounded above by \(\frac{1} {\theta }\log \frac{n} {\epsilon }\) (cf. [15] Lemma II.17, page 116).

Lemma 11 (Proposition 2.2 in [13]). 

Let t 0 ,t 1 ,…,t K be a sequence of positive numbers such that

$$\begin{array}{rcl}{ t}_{k+1} \leq {t}_{k} - \kappa {t}_{k}^{1-\gamma },\qquad k = 0,1,\ldots,K - 1,& & \\ \end{array}$$

where κ > 0 and 0 < γ ≤ 1. Then \(K \leq \left\lfloor \frac{{t}_{0}^{\gamma }} {\kappa \gamma } \right\rfloor \).

Lemma 12.

If K denotes the number of inner iterations, we have

$$K \leq \frac{26128\left(1 + 2\kappa \right)} {3} {\Psi }_{0}^{\frac{3} {4} } \leq 8710\left(1 + 2\kappa \right){\Psi }_{0}^{\frac{3} {4} }.$$

Proof.

The definition of K implies Ψ K − 1 > τ and Ψ K  ≤ τ and

$${\Psi }_{k+1} \leq {\Psi }_{k} - \kappa {\left({\Psi }_{k}\right)}^{1-\gamma },\,\qquad k = 0,1,\ldots,K - 1,$$

with \(\kappa = \frac{1} {6532\left(1+2\kappa \right)}\) and \(\gamma = \frac{3} {4}\). Application of Lemma 11, with t k  = Ψ k yields the desired inequality.

Using ψ0 ≤ L, where the number L is as given in (25), and Lemma 12 we obtain the following upper bound on the total number of iterations:

$$\frac{8710\left(1 + 2\kappa \right){L}^{\frac{3} {4} }} {\theta } \log \frac{n} {\epsilon }.$$
(41)

5.2 Large-Update

We just established that (41) is an upper bound for the total number of iterations, using

$$\psi (t) = \frac{{t}^{2} - 1} {2} + \frac{6} {\pi }\tan \frac{\pi \left(1 - t\right)} {4t + 2} \leq \frac{{t}^{2} - 1} {2}, \mbox{ for}\ t \geq 1,$$

and (21), by substitution in (25) we obtain

$$L \leq n\frac{{\left( \frac{\rho \left( \frac{\tau } {n}\right)} {\sqrt{1-\theta }}\right)}^{2} - 1} {2} \leq \frac{n} {2\left(1 - \theta \right)}\left(\theta + 2\sqrt{2\frac{\tau } {n}} + \frac{2\tau } {n} \right) = \frac{\left(\theta n + 2\sqrt{2\tau n} + 2\tau \right)} {2\left(1 - \theta \right)}.$$

Using (41), thus the total number of iterations is bounded above by

$$\frac{K} {\theta } \log \frac{n} {\epsilon } \leq \frac{8710\left(1 + 2\kappa \right)} {\theta \left(2{\left(1 - \theta \right)}^{\frac{3} {4} }\right)}{ \left(\theta n + 2\sqrt{2\tau n} + 2\tau \right)}^{\frac{3} {4} }\log \frac{n} {\epsilon }.$$

A large-update methods uses τ = O(n) and θ = Θ(1). The right-hand side expression is then \(O\left(\left(1 + 2\kappa \right){n}^{\frac{3} {4} }\log \frac{n} {\epsilon } \right),\) as easily may be verified.

5.3 Small-Update Methods

For small-update methods one has τ = O(1) and \(\theta = \Theta \left( \frac{1} {\sqrt{n}}\right)\). Using Lemma 2, with \(\psi ^{{\prime}{\prime}}(1) = \frac{2\pi +9} {9},\) we then obtain

$$L \leq \frac{n\left(2\pi + 9\right)} {18}{ \left( \frac{\rho \left(\frac{\tau } {n}\right)} {\sqrt{1 - \theta }} - 1\right)}^{2}.$$

Using (21), then

$$L \leq \frac{n\left(2\pi + 9\right)} {18}{ \left(\frac{1 + \sqrt{\frac{2\tau } {n}} } {\sqrt{1 - \theta }} - 1\right)}^{2}.$$

Using \(1 -\sqrt{1 - \theta } = \frac{\theta } {1+\sqrt{1-\theta }} \leq \theta,\) this leads to \(L \leq \frac{\left(2\pi +9\right)} {18\left(1-\theta \right)}{\left(\theta \sqrt{n} + \sqrt{2\tau }\right)}^{2}.\) We conclude that the total number of iterations is bounded above by

$$\begin{array}{rcl} \frac{K} {\theta } \log \frac{n} {\epsilon } \leq \frac{8710\left(1 + 2\kappa \right){\left(2\pi + 9\right)}^{\frac{3} {4} }} {\theta {\left(18\left(1 - \theta \right)\right)}^{\frac{3} {4} }}{ \left(\theta \sqrt{n} + \sqrt{2\tau }\right)}^{\frac{3} {2} }\log \frac{n} {\epsilon }.& & \\ \end{array}$$

Thus, the right-hand side expression is then \(O\left(\left(1 + 2\kappa \right)\sqrt{n}\log \frac{n} {\epsilon } \right).\)

6 Concluding Remarks

In this paper we extended the results obtained for kernel-function-based IPMs in [7] for LO to P (κ) linear complementarity problems. The observation that the vectors d x and d s are not in general orthogonal implies that the analysis in [7] does not hold. The analysis in this paper is new and different from the one using for LO. Several new tools and techniques are derived in this paper. The proposed function has a trigonometric barrier term but the function is not logarithmic and not self-regular. We proved that the iteration bound of a large-update interior-point method based on the kernel function considered in this paper is \(O\left(\left(1 + 2\kappa \right){n}^{\frac{3} {4} }\log \frac{n} {\epsilon } \right),\) which improves the classical iteration complexity with a factor \({n}^{\frac{1} {4} }\). For small-update methods we obtain the best known iteration bound, namely \(O\left(\left(1 + 2\kappa \right)\sqrt{n}\log \frac{n} {\epsilon } \right)\).

The resulting iteration bounds for P (κ) linear complementarity problems depend on the parameter κ. For κ = 0, the iteration bounds are the same as the bounds that were obtained in [7] for linear optimization.