zoukankan      html  css  js  c++  java
  • Definition of matrix norms

    In my previous post, I introduced various definitions of matrix norms in (mathbb{R}^{n imes n}) based on the corresponding vector norms in (mathbb{R}^n). Meanwhile, the equivalence of different vector norms and their induced metrics and topologies in (mathbb{R}^n) is also inherited into (mathbb{R}^{n imes n}). In this article, we’ll show why the above defined matrix norms are valid.

    Generally, the definition of a matrix norm in (mathbb{R}^{n imes n}) should satisfy the following four conditions:

    1. Positive definiteness: for all (A in mathbb{R}^{n imes n}), ( orm{A} geq 0). ( orm{A} = 0) if and only if (A = 0).
    2. Absolute homogeneity: for all (alpha in mathbb{R}) and (A in mathbb{R}^{n imes n}), ( orm{alpha A} = abs{alpha} orm{A}).
    3. Triangle inequality: for all (A, B in mathbb{R}^{n imes n}), ( orm{A + B} leq orm{A} + orm{B}).
    4. Sub-multiplicity: for all (A, B in mathbb{R}^{n imes n}), ( orm{AB} leq orm{A} orm{B}).

    Therefore, we need to prove the following theorem in order to meet the above requirements.

    Theorem Let ( orm{cdot}) be a norm on (mathbb{R}^n). Then for all (A in mathbb{R}^{n imes n}), its matrix norm (zeta: mathbb{R}^{n imes n} ightarrow mathbb{R}) can be defined as
    [
    zeta(A) = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{A vect{x}}}{ orm{vect{x}}} = sup_{forall vect{x} in mathbb{R}^n, orm{vect{x}}=1} orm{A vect{x}}
    ]

    Proof a) Positive definiteness and absolute homogeneity directly inherit from vector norms.

    b) The triangle inequality can be proved as following.
    [
    egin{aligned}
    zeta(A + B) &= sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{(A + B) vect{x}}}{ orm{vect{x}}} = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x} + Bvect{x}}}{ orm{vect{x}}} \
    & leq sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x}} + orm{Bvect{x}}}{ orm{vect{x}}} leq sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x}}}{ orm{vect{x}}} + sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Bvect{x}}}{ orm{vect{x}}} \
    &= zeta(A) + zeta(B).
    end{aligned}
    ]

    c) For sub-multiplicity, we have
    [
    egin{aligned}
    zeta(AB) &= sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{ABvect{x}}}{ orm{vect{x}}} = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{ABvect{x}} orm{Bvect{x}}}{ orm{Bvect{x}} orm{vect{x}}} \
    &leq sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x}}}{ orm{vect{x}}} cdot sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Bvect{x}}}{ orm{vect{x}}} = orm{A} cdot orm{B}.
    end{aligned}
    ]
    d) Prove (zeta(A) = sup_{forall vect{x} in mathbb{R}^n, orm{vect{x}} = 1} orm{Avect{x}}).

    Note that (frac{1}{ orm{vect{x}}}) is a scalar value in (mathbb{R}), then with the proved absolute homogeneity, we have
    [
    zeta(A) = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x}}}{ orm{vect{x}}} = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} leftVert A cdot frac{vect{x}}{ orm{vect{x}}} ightVert.
    ]
    By letting (vect{x}' = frac{vect{x}}{ orm{vect{x}}}), we have this part proved.

    Summarizing a) to d), ( orm{cdot}) is literally a matrix norm induced from the corresponding vector norm.

    Next, we prove the validity of the detailed formulations of the matrix norms, i.e.

    1. 1-norm: ( orm{A}_1 = max_{1 leq j leq n} sum_{i=1}^n abs{a_{ij}}), which is the maximum column sum;
    2. 2-norm: ( orm{A}_2 = sqrt{ ho(A^T A)}), where ( ho) represents the spectral radius, i.e. the maximum eigenvalue of (A^TA);
    3. (infty)-norm: ( orm{A}_{infty} = max_{1 leq i leq n} sum_{j=1}^n abs{a_{ij}}), which is the maximum row sum.

    a) 1-norm: Because
    [
    egin{aligned}
    orm{Avect{x}}_1 &= sum_{i=1}^n leftvert sum_{j=1}^n a_{ij} x_j ightvert leq sum_{i=1}^n sum_{j=1}^n abs{a_{ij} x_j} = sum_{j=1}^n left( abs{x_j} sum_{i=1}^n abs{a_{ij}} ight) \
    &leq left( sum_{j=1}^n abs{x_j} ight) cdot max_{1 leq j leq n} left( sum_{i=1}^n abs{a_{ij}} ight),
    end{aligned}
    ]
    we have
    [
    orm{A}_1 leq sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x}}_1}{ orm{vect{x}}_1} leq frac{left( sum_{j=1}^n abs{x_j} ight) cdot max_{1 leq j leq n} left( sum_{i=1}^n abs{a_{ij}} ight)}{sum_{j=1}^n abs{x_j}} = max_{1 leq j leq n} left( sum_{i=1}^n abs{a_{ij}} ight).
    ]
    Then, we need to show that the maximum value on the right hand side is achievable.

    Assume that when (j = j_0), (sum_{i=1}^n abs{a_{ij}}) has the maximum value. If this value is zero, it means (A) is a zero matrix and the definition of matrix 1-norm is trivially true. If this value is not zero, by letting (vect{x} = (delta_{ij_0})_{i geq 1}^n) with (delta_{ij_0}) being the Kronecker delta, we have
    [
    frac{ orm{Avect{x}}_1}{ orm{vect{x}}_1} = frac{sum_{i=1}^n abs{a_{ij_0}}}{1} = max_{1 leq j leq n} left( sum_{i=1}^n abs{a_{ij}} ight).
    ]
    b) 2-norm: The proof for this part needs the intervention of inner product (langle cdot, cdot angle) of vectors in (mathbb{R}^n), from which the vector 2-norm can be induced. Then we have
    [
    orm{A}_2 = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} frac{ orm{Avect{x}}_2}{ orm{vect{x}}_2} = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} sqrt{frac{langle Avect{x}, Avect{x} angle}{langle vect{x}, vect{x} angle}} = sup_{forall vect{x} in mathbb{R}^n, vect{x} eq 0} sqrt{frac{langle A^*Avect{x}, vect{x} angle}{langle vect{x}, vect{x} angle}},
    ]
    where (A^*) is the adjoint operator, i.e. transpose of (A). Therefore, (A^*A) is a real valued symmetric matrix which has (n) real eigenvalues ({lambda_i}_{i=1}^n) with (0 leq lambda_1 leq cdots leq lambda_n) and (n) corresponding orthonormal eigenvectors ({vect{v}_i}_{i=1}^n) (N.B. There may be duplicates in the eigenvalues). For all (vect{x} in mathbb{R}^n), it can be expanded as (vect{x} = sum_{i=1}^n a_i vect{v}_i) and (A^*Avect{x} = sum_{i=1}^n a_i A^*A vect{v}_i = sum_{i=1}^n a_i lambda_i vect{v}_i). Then we have
    [
    egin{aligned}
    langle A^*Avect{x}, vect{x} angle &= leftlangle sum_{i=1}^n a_i lambda_i vect{v}_i, sum_{j=1}^n a_j vect{v}_j ight angle = sum_{i=1}^n sum_{j=1}^n lambda_i a_i^2 langle vect{v}_i, vect{v}_j angle \
    &= sum_{i=1}^n sum_{j=1}^n lambda_i a_i^2 delta_{ij} = sum_{i=1} lambda_i a_i^2.
    end{aligned}
    ]
    Meanwhile,
    [
    langle vect{x}, vect{x} angle = leftlangle sum_{i=1}^n a_i vect{v}_i, sum_{j=1}^n a_j vect{v}_j ight angle = sum_{i=1}^n sum_{j=1}^n a_i a_j langle vect{v}_i, vect{v}_j angle = sum_{i=1}^n a_i^2.
    ]
    Therefore,
    [
    frac{ orm{Avect{x}}_2}{ orm{vect{x}}_2} leq sqrt{frac{lambda_n sum_{i=1}^n a_i^2}{sum_{i=1}^n a_i^2}} = sqrt{lambda_n}.
    ]
    By letting (a_1 = a_2 = cdots = a_{n-1} = 0) and (a_n = 1), we have (frac{ orm{Avect{x}}_2}{ orm{vect{x}}_2} = sqrt{lambda_n}). Hence,
    [
    orm{A}_2 = sqrt{lambda_n} = sqrt{ ho(A^*A)}
    ]
    and the definition of matrix 2-norm is valid.

    c) (infty)-norm:
    [
    egin{aligned}
    orm{Avect{x}}_{infty} &= max_{1 leq i leq n} left( leftvert sum_{j=1}^n a_{ij} x_j ightvert ight) leq max_{1 leq i leq n} left( sum_{j=1}^n abs{a_{ij}} cdot abs{x_j} ight) \
    &= max_{1 leq i leq n} left( left( sum_{j=1}^n abs{a_{ij}} ight) cdot left( max_{1 leq j leq n} abs{x_j} ight) ight) = left( max_{1 leq i leq n} sum_{j=1}^n abs{a_{ij}} ight) cdot left( max_{1 leq j leq n} abs{x_j} ight) \
    orm{vect{x}}_{infty} &= max_{1 leq i leq n} abs{x_i}
    end{aligned}
    ]
    Therefore, (frac{ orm{Avect{x}}_{infty}}{ orm{vect{x}}_{infty}} leq max_{1 leq i leq n} sum_{j=1}^n abs{a_{ij}}). Then, we need to prove this maximum value is achievable.

    Assume when (i = i_0), (sum_{j=1}^n abs{a_{i_0 j}}) achieves the maximum. If this value is zero, (A) is a zero matrix and the definition of matrix (infty)-norm is trivially true. If this value is not zero, by letting (vect{x} = (sgn(a_{i_0 1}), cdots, sgn(a_{i_0 n}))^{ m T}), we have ( orm{vect{x}}_{infty} = 1) and ( orm{Avect{x}}_{infty} = sum_{j=1}^n abs{a_{i_0 j}} = max_{1 leq i leq n} sum_{j=1}^n abs{a_{ij}}). Hence, (frac{ orm{Avect{x}}_{infty}}{ orm{vect{x}}_{infty}} = max_{1 leq i leq n} sum_{j=1}^n abs{a_{ij}}) and the definition of (infty​)-norm is valid.

  • 相关阅读:
    MagicAjax.NET Framework
    实例说明sql优化的重要性
    求助:谁可以给我点share point 开发的资料
    整理了一些tsql技巧(收藏)
    熟悉/了解C#基础问题
    .NET在SQL Server中的图片存取技术(例子在管理里InOutImg)
    SQL Server 2000 函数使用CAST 和 CONVERT
    MagicAjax.NET Framework configuration
    找工作了,希望各位多多帮忙
    OSG节点访问和遍历
  • 原文地址:https://www.cnblogs.com/peabody/p/10257742.html
Copyright © 2011-2022 走看看