Polynomial SOS

From The Right Wiki
Jump to navigationJump to search

In mathematics, a form (i.e. a homogeneous polynomial) h(x) of degree 2m in the real n-dimensional vector x is sum of squares of forms (SOS) if and only if there exist forms g1(x),,gk(x) of degree m such that h(x)=i=1kgi(x)2. Every form that is SOS is also a positive polynomial, and although the converse is not always true, Hilbert proved that for n = 2, 2m = 2, or n = 3 and 2m = 4 a form is SOS if and only if it is positive.[1] The same is also valid for the analog problem on positive symmetric forms.[2][3] Although not every form can be represented as SOS, explicit sufficient conditions for a form to be SOS have been found.[4][5] Moreover, every real nonnegative form can be approximated as closely as desired (in the l1-norm of its coefficient vector) by a sequence of forms {fϵ} that are SOS.[6]

Square matricial representation (SMR)

To establish whether a form h(x) is SOS amounts to solving a convex optimization problem. Indeed, any h(x) can be written as h(x)=x{m}(H+L(α))x{m} where x{m} is a vector containing a base for the forms of degree m in x (such as all monomials of degree m in x), the prime ′ denotes the transpose, H is any symmetric matrix satisfying h(x)=x{m}Hx{m} and L(α) is a linear parameterization of the linear space ={L=L:x{m}Lx{m}=0}. The dimension of the vector x{m} is given by σ(n,m)=(n+m1m), whereas the dimension of the vector α is given by ω(n,2m)=12σ(n,m)(1+σ(n,m))σ(n,2m). Then, h(x) is SOS if and only if there exists a vector α such that H+L(α)0, meaning that the matrix H+L(α) is positive-semidefinite. This is a linear matrix inequality (LMI) feasibility test, which is a convex optimization problem. The expression h(x)=x{m}(H+L(α))x{m} was introduced in [7] with the name square matricial representation (SMR) in order to establish whether a form is SOS via an LMI. This representation is also known as Gram matrix.[8]

Examples

  • Consider the form of degree 4 in two variables h(x)=x14x12x22+x24. We have m=2,x{m}=(x12x1x2x22),H+L(α)=(10α101+2α10α101). Since there exists α such that H+L(α)0, namely α=1, it follows that h(x) is SOS.
  • Consider the form of degree 4 in three variables h(x)=2x142.5x13x2+x12x2x32x1x33+5x24+x34. We have m=2,x{m}=(x12x1x2x1x3x22x2x3x32),H+L(α)=(21.250α1α2α31.252α10.5+α20α4α500.5+α22α3α4α51α10α450α6α2α4α502α60α3α51α601). Since H+L(α)0 for α=(1.18,0.43,0.73,1.13,0.37,0.57), it follows that h(x) is SOS.

Generalizations

Matrix SOS

A matrix form F(x) (i.e., a matrix whose entries are forms) of dimension r and degree 2m in the real n-dimensional vector x is SOS if and only if there exist matrix forms G1(x),,Gk(x) of degree m such that F(x)=i=1kGi(x)Gi(x).

Matrix SMR

To establish whether a matrix form F(x) is SOS amounts to solving a convex optimization problem. Indeed, similarly to the scalar case any F(x) can be written according to the SMR as F(x)=(x{m}Ir)(H+L(α))(x{m}Ir) where is the Kronecker product of matrices, H is any symmetric matrix satisfying F(x)=(x{m}Ir)H(x{m}Ir) and L(α) is a linear parameterization of the linear space ={L=L:(x{m}Ir)L(x{m}Ir)=0}. The dimension of the vector α is given by ω(n,2m,r)=12r(σ(n,m)(rσ(n,m)+1)(r+1)σ(n,2m)). Then, F(x) is SOS if and only if there exists a vector α such that the following LMI holds: H+L(α)0. The expression F(x)=(x{m}Ir)(H+L(α))(x{m}Ir) was introduced in [9] in order to establish whether a matrix form is SOS via an LMI.

Noncommutative polynomial SOS

Consider the free algebra RX⟩ generated by the n noncommuting letters X = (X1, ..., Xn) and equipped with the involution T, such that T fixes R and X1, ..., Xn and reverses words formed by X1, ..., Xn. By analogy with the commutative case, the noncommutative symmetric polynomials f are the noncommutative polynomials of the form f = fT. When any real matrix of any dimension r × r is evaluated at a symmetric noncommutative polynomial f results in a positive semi-definite matrix, f is said to be matrix-positive. A noncommutative polynomial is SOS if there exists noncommutative polynomials h1,,hk such that f(X)=i=1khi(X)Thi(X). Surprisingly, in the noncommutative scenario a noncommutative polynomial is SOS if and only if it is matrix-positive.[10] Moreover, there exist algorithms available to decompose matrix-positive polynomials in sum of squares of noncommutative polynomials.[11]

References

  1. Hilbert, David (September 1888). "Ueber die Darstellung definiter Formen als Summe von Formenquadraten". Mathematische Annalen. 32 (3): 342–350. doi:10.1007/bf01443605. S2CID 177804714.
  2. Choi, M. D.; Lam, T. Y. (1977). "An old question of Hilbert". Queen's Papers in Pure and Applied Mathematics. 46: 385–405.
  3. Goel, Charu; Kuhlmann, Salma; Reznick, Bruce (May 2016). "On the Choi–Lam analogue of Hilbert's 1888 theorem for symmetric forms". Linear Algebra and Its Applications. 496: 114–120. arXiv:1505.08145. doi:10.1016/j.laa.2016.01.024. S2CID 17579200.
  4. Lasserre, Jean B. (2007). "Sufficient conditions for a real polynomial to be a sum of squares". Archiv der Mathematik. 89 (5): 390–398. arXiv:math/0612358. CiteSeerX 10.1.1.240.4438. doi:10.1007/s00013-007-2251-y. S2CID 9319455.
  5. Powers, Victoria; Wörmann, Thorsten (1998). "An algorithm for sums of squares of real polynomials" (PDF). Journal of Pure and Applied Algebra. 127 (1): 99–104. doi:10.1016/S0022-4049(97)83827-3.
  6. Lasserre, Jean B. (2007). "A Sum of Squares Approximation of Nonnegative Polynomials". SIAM Review. 49 (4): 651–669. arXiv:math/0412398. Bibcode:2007SIAMR..49..651L. doi:10.1137/070693709.
  7. Chesi, G.; Tesi, A.; Vicino, A.; Genesio, R. (1999). "On convexification of some minimum distance problems". Proceedings of the 5th European Control Conference. Karlsruhe, Germany: IEEE. pp. 1446–1451.
  8. Choi, M.; Lam, T.; Reznick, B. (1995). "Sums of squares of real polynomials". Proceedings of Symposia in Pure Mathematics. pp. 103–125.
  9. Chesi, G.; Garulli, A.; Tesi, A.; Vicino, A. (2003). "Robust stability for polytopic systems via polynomially parameter-dependent Lyapunov functions". Proceedings of the 42nd IEEE Conference on Decision and Control. Maui, Hawaii: IEEE. pp. 4670–4675. doi:10.1109/CDC.2003.1272307.
  10. Helton, J. William (September 2002). ""Positive" Noncommutative Polynomials Are Sums of Squares". The Annals of Mathematics. 156 (2): 675–694. doi:10.2307/3597203. JSTOR 3597203.
  11. Burgdorf, Sabine; Cafuta, Kristijan; Klep, Igor; Povh, Janez (25 October 2012). "Algorithmic aspects of sums of Hermitian squares of noncommutative polynomials". Computational Optimization and Applications. 55 (1): 137–153. CiteSeerX 10.1.1.416.543. doi:10.1007/s10589-012-9513-8. S2CID 254416733.

See also