Full and reduced SVD of a 3x3 matrix.












0














I currently studying for an exam, and I'm currently working my way through some old exam problems and I'm currently at the following.



First, we have a matrix



$A=
begin{bmatrix}
2&0&0\2&1&0\0&-2&0
end{bmatrix}$




  • a. Determine the singular values of the matrix A.

  • b. Write down the reduced SVD-decomposition of A.

  • c. Determine the full SVD-decomposition of A.

  • d. Let $C=A^*A$ and $D=AA^*$. Determine whether these are positive semidefinite.

  • e. Are they positive definite?


My answers are,
For readers ease I write down the formulars.



$A=USigma V^T$, where $U$ and $V$ are orthogonal and $Sigma$ is a diagonal matrix.



$A^TA=VSigma^T Sigma V^T$



$AV=USigma$.



a. First i find the eigenvalues of $A^TA$, the root these values will be my singular values for A.



$det(A^TA-lambda I)=begin{vmatrix}
8-lambda&2&0\2&5-lambda&0\0&0&-lambda
end{vmatrix}=lambda(lambda-4)(lambda-9)$



Thus, $lambda_1=9, lambda_2=4$ and $lambda_3=0$.



So the singular values is 3,2 and 0. I ordered the singular values such that $Sigma$ has increasing values along the diagonal to simplify notation. $Sigma=begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



b. I now find $V=(e_1,e_2,e_3)$, where $e_j,j=1,2,3$ is the unit vector corresponding to the the $j'th$ eigenvalue. I compute these by applying the Gram-Schmidt procedure to the arbitrary list of eigenvectors for $A^TA$.



$V=begin{bmatrix}
frac{2}{sqrt5}&frac{1}{sqrt5}&0\
frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
0&0&1
end{bmatrix}$



Now I only need to find U, because $V=V^T$.
I find U by using the formular $AV=USigma$.



$AV=begin{bmatrix}
2&0&0\2&1&0\0&-2&0
end{bmatrix}begin{bmatrix}
frac{2}{sqrt5}&frac{1}{sqrt5}&0\
frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
0&0&1
end{bmatrix}=begin{bmatrix}
frac{4}{sqrt5}&frac{2}{sqrt5}&0\
sqrt5&0&0\
-frac{2}{sqrt5}&frac{4}{sqrt5}&0
end{bmatrix}$



$USigma=begin{bmatrix}
frac{4}{sqrt5}&frac{2}{sqrt5}&0\
sqrt5&0&0\
-frac{2}{sqrt5}&frac{4}{sqrt5}&0
end{bmatrix}
=
begin{bmatrix}
frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
frac{sqrt5}{3}&0&0\
-frac{2}{3sqrt5}&frac{2}{sqrt5}&0
end{bmatrix}
begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



Now I have all three components of the SVD-decomposition such that



$A=USigma V^T=begin{bmatrix}
frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
frac{sqrt5}{3}&0&0\
-frac{2}{3sqrt5}&frac{2}{sqrt5}&0
end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix}
frac{2}{sqrt5}&frac{1}{sqrt5}&0\
frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
0&0&1
end{bmatrix}$
.



I believe that this answers both b. and c. because this is the reduced SVD and it's regarding a square matrix, so it's already a full SVD?



d. and e. First I calculate the matrices and then find the determinants of the upper left principals of the matrix, if they are all non-negative numbers, they will be positive semidefinite, if the determinants are strictly positive, they will be positive definite.



Knowing that $A$ is a real matrix, thus the adjoint of it is just its transpose, such that,



$C=begin{bmatrix}
2&0&0\2&1&0\0&-2&0
end{bmatrix}begin{bmatrix}
2&2&0\0&1&-2\0&0&0
end{bmatrix}=begin{bmatrix}
4&4&0\4&5&-2\0&-2&4
end{bmatrix}$



The determinants of the upper left principals will be, $4,4$ and $0$, so $C$ is positive semidefinite.



$D=begin{bmatrix}
2&2&0\0&1&-2\0&0&0
end{bmatrix}begin{bmatrix}
2&0&0\2&1&0\0&-2&0
end{bmatrix}=begin{bmatrix}
8&2&0\2&5&0\0&0&0
end{bmatrix}$



Here the determinants of the upper left principals is $8, 36$ and $0$, so this matrix is also positive semidefinite.



This concludes the problem.



I'm not totally sure about my answers, so I hope that I get some tips, tricks and corrections.
I'm really iffy about the argument of that the reduced SVD is the same as the full SVD for a square matrix.



Best regards Jens










share|cite|improve this question





























    0














    I currently studying for an exam, and I'm currently working my way through some old exam problems and I'm currently at the following.



    First, we have a matrix



    $A=
    begin{bmatrix}
    2&0&0\2&1&0\0&-2&0
    end{bmatrix}$




    • a. Determine the singular values of the matrix A.

    • b. Write down the reduced SVD-decomposition of A.

    • c. Determine the full SVD-decomposition of A.

    • d. Let $C=A^*A$ and $D=AA^*$. Determine whether these are positive semidefinite.

    • e. Are they positive definite?


    My answers are,
    For readers ease I write down the formulars.



    $A=USigma V^T$, where $U$ and $V$ are orthogonal and $Sigma$ is a diagonal matrix.



    $A^TA=VSigma^T Sigma V^T$



    $AV=USigma$.



    a. First i find the eigenvalues of $A^TA$, the root these values will be my singular values for A.



    $det(A^TA-lambda I)=begin{vmatrix}
    8-lambda&2&0\2&5-lambda&0\0&0&-lambda
    end{vmatrix}=lambda(lambda-4)(lambda-9)$



    Thus, $lambda_1=9, lambda_2=4$ and $lambda_3=0$.



    So the singular values is 3,2 and 0. I ordered the singular values such that $Sigma$ has increasing values along the diagonal to simplify notation. $Sigma=begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



    b. I now find $V=(e_1,e_2,e_3)$, where $e_j,j=1,2,3$ is the unit vector corresponding to the the $j'th$ eigenvalue. I compute these by applying the Gram-Schmidt procedure to the arbitrary list of eigenvectors for $A^TA$.



    $V=begin{bmatrix}
    frac{2}{sqrt5}&frac{1}{sqrt5}&0\
    frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
    0&0&1
    end{bmatrix}$



    Now I only need to find U, because $V=V^T$.
    I find U by using the formular $AV=USigma$.



    $AV=begin{bmatrix}
    2&0&0\2&1&0\0&-2&0
    end{bmatrix}begin{bmatrix}
    frac{2}{sqrt5}&frac{1}{sqrt5}&0\
    frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
    0&0&1
    end{bmatrix}=begin{bmatrix}
    frac{4}{sqrt5}&frac{2}{sqrt5}&0\
    sqrt5&0&0\
    -frac{2}{sqrt5}&frac{4}{sqrt5}&0
    end{bmatrix}$



    $USigma=begin{bmatrix}
    frac{4}{sqrt5}&frac{2}{sqrt5}&0\
    sqrt5&0&0\
    -frac{2}{sqrt5}&frac{4}{sqrt5}&0
    end{bmatrix}
    =
    begin{bmatrix}
    frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
    frac{sqrt5}{3}&0&0\
    -frac{2}{3sqrt5}&frac{2}{sqrt5}&0
    end{bmatrix}
    begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



    Now I have all three components of the SVD-decomposition such that



    $A=USigma V^T=begin{bmatrix}
    frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
    frac{sqrt5}{3}&0&0\
    -frac{2}{3sqrt5}&frac{2}{sqrt5}&0
    end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix}
    frac{2}{sqrt5}&frac{1}{sqrt5}&0\
    frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
    0&0&1
    end{bmatrix}$
    .



    I believe that this answers both b. and c. because this is the reduced SVD and it's regarding a square matrix, so it's already a full SVD?



    d. and e. First I calculate the matrices and then find the determinants of the upper left principals of the matrix, if they are all non-negative numbers, they will be positive semidefinite, if the determinants are strictly positive, they will be positive definite.



    Knowing that $A$ is a real matrix, thus the adjoint of it is just its transpose, such that,



    $C=begin{bmatrix}
    2&0&0\2&1&0\0&-2&0
    end{bmatrix}begin{bmatrix}
    2&2&0\0&1&-2\0&0&0
    end{bmatrix}=begin{bmatrix}
    4&4&0\4&5&-2\0&-2&4
    end{bmatrix}$



    The determinants of the upper left principals will be, $4,4$ and $0$, so $C$ is positive semidefinite.



    $D=begin{bmatrix}
    2&2&0\0&1&-2\0&0&0
    end{bmatrix}begin{bmatrix}
    2&0&0\2&1&0\0&-2&0
    end{bmatrix}=begin{bmatrix}
    8&2&0\2&5&0\0&0&0
    end{bmatrix}$



    Here the determinants of the upper left principals is $8, 36$ and $0$, so this matrix is also positive semidefinite.



    This concludes the problem.



    I'm not totally sure about my answers, so I hope that I get some tips, tricks and corrections.
    I'm really iffy about the argument of that the reduced SVD is the same as the full SVD for a square matrix.



    Best regards Jens










    share|cite|improve this question



























      0












      0








      0







      I currently studying for an exam, and I'm currently working my way through some old exam problems and I'm currently at the following.



      First, we have a matrix



      $A=
      begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}$




      • a. Determine the singular values of the matrix A.

      • b. Write down the reduced SVD-decomposition of A.

      • c. Determine the full SVD-decomposition of A.

      • d. Let $C=A^*A$ and $D=AA^*$. Determine whether these are positive semidefinite.

      • e. Are they positive definite?


      My answers are,
      For readers ease I write down the formulars.



      $A=USigma V^T$, where $U$ and $V$ are orthogonal and $Sigma$ is a diagonal matrix.



      $A^TA=VSigma^T Sigma V^T$



      $AV=USigma$.



      a. First i find the eigenvalues of $A^TA$, the root these values will be my singular values for A.



      $det(A^TA-lambda I)=begin{vmatrix}
      8-lambda&2&0\2&5-lambda&0\0&0&-lambda
      end{vmatrix}=lambda(lambda-4)(lambda-9)$



      Thus, $lambda_1=9, lambda_2=4$ and $lambda_3=0$.



      So the singular values is 3,2 and 0. I ordered the singular values such that $Sigma$ has increasing values along the diagonal to simplify notation. $Sigma=begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



      b. I now find $V=(e_1,e_2,e_3)$, where $e_j,j=1,2,3$ is the unit vector corresponding to the the $j'th$ eigenvalue. I compute these by applying the Gram-Schmidt procedure to the arbitrary list of eigenvectors for $A^TA$.



      $V=begin{bmatrix}
      frac{2}{sqrt5}&frac{1}{sqrt5}&0\
      frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
      0&0&1
      end{bmatrix}$



      Now I only need to find U, because $V=V^T$.
      I find U by using the formular $AV=USigma$.



      $AV=begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}begin{bmatrix}
      frac{2}{sqrt5}&frac{1}{sqrt5}&0\
      frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
      0&0&1
      end{bmatrix}=begin{bmatrix}
      frac{4}{sqrt5}&frac{2}{sqrt5}&0\
      sqrt5&0&0\
      -frac{2}{sqrt5}&frac{4}{sqrt5}&0
      end{bmatrix}$



      $USigma=begin{bmatrix}
      frac{4}{sqrt5}&frac{2}{sqrt5}&0\
      sqrt5&0&0\
      -frac{2}{sqrt5}&frac{4}{sqrt5}&0
      end{bmatrix}
      =
      begin{bmatrix}
      frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
      frac{sqrt5}{3}&0&0\
      -frac{2}{3sqrt5}&frac{2}{sqrt5}&0
      end{bmatrix}
      begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



      Now I have all three components of the SVD-decomposition such that



      $A=USigma V^T=begin{bmatrix}
      frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
      frac{sqrt5}{3}&0&0\
      -frac{2}{3sqrt5}&frac{2}{sqrt5}&0
      end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix}
      frac{2}{sqrt5}&frac{1}{sqrt5}&0\
      frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
      0&0&1
      end{bmatrix}$
      .



      I believe that this answers both b. and c. because this is the reduced SVD and it's regarding a square matrix, so it's already a full SVD?



      d. and e. First I calculate the matrices and then find the determinants of the upper left principals of the matrix, if they are all non-negative numbers, they will be positive semidefinite, if the determinants are strictly positive, they will be positive definite.



      Knowing that $A$ is a real matrix, thus the adjoint of it is just its transpose, such that,



      $C=begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}begin{bmatrix}
      2&2&0\0&1&-2\0&0&0
      end{bmatrix}=begin{bmatrix}
      4&4&0\4&5&-2\0&-2&4
      end{bmatrix}$



      The determinants of the upper left principals will be, $4,4$ and $0$, so $C$ is positive semidefinite.



      $D=begin{bmatrix}
      2&2&0\0&1&-2\0&0&0
      end{bmatrix}begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}=begin{bmatrix}
      8&2&0\2&5&0\0&0&0
      end{bmatrix}$



      Here the determinants of the upper left principals is $8, 36$ and $0$, so this matrix is also positive semidefinite.



      This concludes the problem.



      I'm not totally sure about my answers, so I hope that I get some tips, tricks and corrections.
      I'm really iffy about the argument of that the reduced SVD is the same as the full SVD for a square matrix.



      Best regards Jens










      share|cite|improve this question















      I currently studying for an exam, and I'm currently working my way through some old exam problems and I'm currently at the following.



      First, we have a matrix



      $A=
      begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}$




      • a. Determine the singular values of the matrix A.

      • b. Write down the reduced SVD-decomposition of A.

      • c. Determine the full SVD-decomposition of A.

      • d. Let $C=A^*A$ and $D=AA^*$. Determine whether these are positive semidefinite.

      • e. Are they positive definite?


      My answers are,
      For readers ease I write down the formulars.



      $A=USigma V^T$, where $U$ and $V$ are orthogonal and $Sigma$ is a diagonal matrix.



      $A^TA=VSigma^T Sigma V^T$



      $AV=USigma$.



      a. First i find the eigenvalues of $A^TA$, the root these values will be my singular values for A.



      $det(A^TA-lambda I)=begin{vmatrix}
      8-lambda&2&0\2&5-lambda&0\0&0&-lambda
      end{vmatrix}=lambda(lambda-4)(lambda-9)$



      Thus, $lambda_1=9, lambda_2=4$ and $lambda_3=0$.



      So the singular values is 3,2 and 0. I ordered the singular values such that $Sigma$ has increasing values along the diagonal to simplify notation. $Sigma=begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



      b. I now find $V=(e_1,e_2,e_3)$, where $e_j,j=1,2,3$ is the unit vector corresponding to the the $j'th$ eigenvalue. I compute these by applying the Gram-Schmidt procedure to the arbitrary list of eigenvectors for $A^TA$.



      $V=begin{bmatrix}
      frac{2}{sqrt5}&frac{1}{sqrt5}&0\
      frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
      0&0&1
      end{bmatrix}$



      Now I only need to find U, because $V=V^T$.
      I find U by using the formular $AV=USigma$.



      $AV=begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}begin{bmatrix}
      frac{2}{sqrt5}&frac{1}{sqrt5}&0\
      frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
      0&0&1
      end{bmatrix}=begin{bmatrix}
      frac{4}{sqrt5}&frac{2}{sqrt5}&0\
      sqrt5&0&0\
      -frac{2}{sqrt5}&frac{4}{sqrt5}&0
      end{bmatrix}$



      $USigma=begin{bmatrix}
      frac{4}{sqrt5}&frac{2}{sqrt5}&0\
      sqrt5&0&0\
      -frac{2}{sqrt5}&frac{4}{sqrt5}&0
      end{bmatrix}
      =
      begin{bmatrix}
      frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
      frac{sqrt5}{3}&0&0\
      -frac{2}{3sqrt5}&frac{2}{sqrt5}&0
      end{bmatrix}
      begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}$



      Now I have all three components of the SVD-decomposition such that



      $A=USigma V^T=begin{bmatrix}
      frac{4}{3sqrt5}&frac{1}{sqrt5}&0\
      frac{sqrt5}{3}&0&0\
      -frac{2}{3sqrt5}&frac{2}{sqrt5}&0
      end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix}
      frac{2}{sqrt5}&frac{1}{sqrt5}&0\
      frac{1}{sqrt5}&-frac{2}{sqrt5}&0\
      0&0&1
      end{bmatrix}$
      .



      I believe that this answers both b. and c. because this is the reduced SVD and it's regarding a square matrix, so it's already a full SVD?



      d. and e. First I calculate the matrices and then find the determinants of the upper left principals of the matrix, if they are all non-negative numbers, they will be positive semidefinite, if the determinants are strictly positive, they will be positive definite.



      Knowing that $A$ is a real matrix, thus the adjoint of it is just its transpose, such that,



      $C=begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}begin{bmatrix}
      2&2&0\0&1&-2\0&0&0
      end{bmatrix}=begin{bmatrix}
      4&4&0\4&5&-2\0&-2&4
      end{bmatrix}$



      The determinants of the upper left principals will be, $4,4$ and $0$, so $C$ is positive semidefinite.



      $D=begin{bmatrix}
      2&2&0\0&1&-2\0&0&0
      end{bmatrix}begin{bmatrix}
      2&0&0\2&1&0\0&-2&0
      end{bmatrix}=begin{bmatrix}
      8&2&0\2&5&0\0&0&0
      end{bmatrix}$



      Here the determinants of the upper left principals is $8, 36$ and $0$, so this matrix is also positive semidefinite.



      This concludes the problem.



      I'm not totally sure about my answers, so I hope that I get some tips, tricks and corrections.
      I'm really iffy about the argument of that the reduced SVD is the same as the full SVD for a square matrix.



      Best regards Jens







      linear-algebra svd






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited yesterday

























      asked yesterday









      Jens Kramer

      557




      557






















          1 Answer
          1






          active

          oldest

          votes


















          0














          I think you have the right general approach to computing the SVD.
          A couple notes:




          • Once you find the eigenvectors of $A^*A$, you know that eigenvectors will be orthogonal by the spectral theorem if their eigenvalues are distinct. So you only need to normalize them rather than do the entire Gram-Schmidt.

          • if you have a zero singular value is that you cannot compute $sigma u = Av$ does not mean $u = Av / sigma$. In your example, this lead to your $U$ missing a column (and therefore not being unitary). You can fix this by replacing it with a unit vector orthogonal to the other two columns.


          The definition of reduced SVD varies some. If $A$ is $mtimes n$ with $(mgeq n)$, the reduced SVD generally means the terms are factors liked $(mtimes n), (ntimes n), (ntimes n)$. This is what you did, and so if $A$ is square, the reduced SVD would be the same as the regular SVD.



          On the other hand, reduced SVD could mean $Sigma$ is square and of size equal to the rank of $A$. In your case, this would mean dropping the last column of $U$, the last rows and columsn of $A$ and the last row of $V^*$.



          All of these decompositions are equivalent since you are just dropping entries which end up multiplied by zero, but you should check which one you're using in the class.



          For the positive definite questions, there are many equivalent definitions which might be faster to use. However, I'm not sure if you have seen them. Some useful ones are a hermitian matrix $M$ is positive definite if




          • all eigenvalues are positive

          • it can be factored $M = CC^T$ for some $C$ full rank.

          • for all $x neq 0$, $x^TMx > 0$






          share|cite|improve this answer























          • About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
            – Jens Kramer
            yesterday










          • In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
            – Jens Kramer
            yesterday












          • It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
            – tch
            yesterday













          Your Answer





          StackExchange.ifUsing("editor", function () {
          return StackExchange.using("mathjaxEditing", function () {
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          });
          });
          }, "mathjax-editing");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "69"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: true,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: 10,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          noCode: true, onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3060622%2ffull-and-reduced-svd-of-a-3x3-matrix%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          0














          I think you have the right general approach to computing the SVD.
          A couple notes:




          • Once you find the eigenvectors of $A^*A$, you know that eigenvectors will be orthogonal by the spectral theorem if their eigenvalues are distinct. So you only need to normalize them rather than do the entire Gram-Schmidt.

          • if you have a zero singular value is that you cannot compute $sigma u = Av$ does not mean $u = Av / sigma$. In your example, this lead to your $U$ missing a column (and therefore not being unitary). You can fix this by replacing it with a unit vector orthogonal to the other two columns.


          The definition of reduced SVD varies some. If $A$ is $mtimes n$ with $(mgeq n)$, the reduced SVD generally means the terms are factors liked $(mtimes n), (ntimes n), (ntimes n)$. This is what you did, and so if $A$ is square, the reduced SVD would be the same as the regular SVD.



          On the other hand, reduced SVD could mean $Sigma$ is square and of size equal to the rank of $A$. In your case, this would mean dropping the last column of $U$, the last rows and columsn of $A$ and the last row of $V^*$.



          All of these decompositions are equivalent since you are just dropping entries which end up multiplied by zero, but you should check which one you're using in the class.



          For the positive definite questions, there are many equivalent definitions which might be faster to use. However, I'm not sure if you have seen them. Some useful ones are a hermitian matrix $M$ is positive definite if




          • all eigenvalues are positive

          • it can be factored $M = CC^T$ for some $C$ full rank.

          • for all $x neq 0$, $x^TMx > 0$






          share|cite|improve this answer























          • About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
            – Jens Kramer
            yesterday










          • In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
            – Jens Kramer
            yesterday












          • It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
            – tch
            yesterday


















          0














          I think you have the right general approach to computing the SVD.
          A couple notes:




          • Once you find the eigenvectors of $A^*A$, you know that eigenvectors will be orthogonal by the spectral theorem if their eigenvalues are distinct. So you only need to normalize them rather than do the entire Gram-Schmidt.

          • if you have a zero singular value is that you cannot compute $sigma u = Av$ does not mean $u = Av / sigma$. In your example, this lead to your $U$ missing a column (and therefore not being unitary). You can fix this by replacing it with a unit vector orthogonal to the other two columns.


          The definition of reduced SVD varies some. If $A$ is $mtimes n$ with $(mgeq n)$, the reduced SVD generally means the terms are factors liked $(mtimes n), (ntimes n), (ntimes n)$. This is what you did, and so if $A$ is square, the reduced SVD would be the same as the regular SVD.



          On the other hand, reduced SVD could mean $Sigma$ is square and of size equal to the rank of $A$. In your case, this would mean dropping the last column of $U$, the last rows and columsn of $A$ and the last row of $V^*$.



          All of these decompositions are equivalent since you are just dropping entries which end up multiplied by zero, but you should check which one you're using in the class.



          For the positive definite questions, there are many equivalent definitions which might be faster to use. However, I'm not sure if you have seen them. Some useful ones are a hermitian matrix $M$ is positive definite if




          • all eigenvalues are positive

          • it can be factored $M = CC^T$ for some $C$ full rank.

          • for all $x neq 0$, $x^TMx > 0$






          share|cite|improve this answer























          • About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
            – Jens Kramer
            yesterday










          • In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
            – Jens Kramer
            yesterday












          • It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
            – tch
            yesterday
















          0












          0








          0






          I think you have the right general approach to computing the SVD.
          A couple notes:




          • Once you find the eigenvectors of $A^*A$, you know that eigenvectors will be orthogonal by the spectral theorem if their eigenvalues are distinct. So you only need to normalize them rather than do the entire Gram-Schmidt.

          • if you have a zero singular value is that you cannot compute $sigma u = Av$ does not mean $u = Av / sigma$. In your example, this lead to your $U$ missing a column (and therefore not being unitary). You can fix this by replacing it with a unit vector orthogonal to the other two columns.


          The definition of reduced SVD varies some. If $A$ is $mtimes n$ with $(mgeq n)$, the reduced SVD generally means the terms are factors liked $(mtimes n), (ntimes n), (ntimes n)$. This is what you did, and so if $A$ is square, the reduced SVD would be the same as the regular SVD.



          On the other hand, reduced SVD could mean $Sigma$ is square and of size equal to the rank of $A$. In your case, this would mean dropping the last column of $U$, the last rows and columsn of $A$ and the last row of $V^*$.



          All of these decompositions are equivalent since you are just dropping entries which end up multiplied by zero, but you should check which one you're using in the class.



          For the positive definite questions, there are many equivalent definitions which might be faster to use. However, I'm not sure if you have seen them. Some useful ones are a hermitian matrix $M$ is positive definite if




          • all eigenvalues are positive

          • it can be factored $M = CC^T$ for some $C$ full rank.

          • for all $x neq 0$, $x^TMx > 0$






          share|cite|improve this answer














          I think you have the right general approach to computing the SVD.
          A couple notes:




          • Once you find the eigenvectors of $A^*A$, you know that eigenvectors will be orthogonal by the spectral theorem if their eigenvalues are distinct. So you only need to normalize them rather than do the entire Gram-Schmidt.

          • if you have a zero singular value is that you cannot compute $sigma u = Av$ does not mean $u = Av / sigma$. In your example, this lead to your $U$ missing a column (and therefore not being unitary). You can fix this by replacing it with a unit vector orthogonal to the other two columns.


          The definition of reduced SVD varies some. If $A$ is $mtimes n$ with $(mgeq n)$, the reduced SVD generally means the terms are factors liked $(mtimes n), (ntimes n), (ntimes n)$. This is what you did, and so if $A$ is square, the reduced SVD would be the same as the regular SVD.



          On the other hand, reduced SVD could mean $Sigma$ is square and of size equal to the rank of $A$. In your case, this would mean dropping the last column of $U$, the last rows and columsn of $A$ and the last row of $V^*$.



          All of these decompositions are equivalent since you are just dropping entries which end up multiplied by zero, but you should check which one you're using in the class.



          For the positive definite questions, there are many equivalent definitions which might be faster to use. However, I'm not sure if you have seen them. Some useful ones are a hermitian matrix $M$ is positive definite if




          • all eigenvalues are positive

          • it can be factored $M = CC^T$ for some $C$ full rank.

          • for all $x neq 0$, $x^TMx > 0$







          share|cite|improve this answer














          share|cite|improve this answer



          share|cite|improve this answer








          edited yesterday

























          answered yesterday









          tch

          584210




          584210












          • About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
            – Jens Kramer
            yesterday










          • In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
            – Jens Kramer
            yesterday












          • It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
            – tch
            yesterday




















          • About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
            – Jens Kramer
            yesterday










          • In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
            – Jens Kramer
            yesterday












          • It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
            – tch
            yesterday


















          About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
          – Jens Kramer
          yesterday




          About the second note you provided, I calculated another orthogonal vector such that, $A=USigma V^T=begin{bmatrix} frac{4}{3sqrt5}&frac{1}{sqrt5}&-frac{2}{3}\ frac{sqrt5}{3}&0&frac{2}{3}\ -frac{2}{3sqrt5}&frac{2}{sqrt5}&frac{1}{3} end{bmatrix}begin{bmatrix}3&0&0\0&2&0\0&0&0end{bmatrix}begin{bmatrix} frac{2}{sqrt5}&frac{1}{sqrt5}&0\ frac{1}{sqrt5}&-frac{2}{sqrt5}&0\ 0&0&1 end{bmatrix}$ Does this provide the full SVD?
          – Jens Kramer
          yesterday












          In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
          – Jens Kramer
          yesterday






          In the case of reduced SVD, my book states that $Sigma$ would be a $rtimes r$ matrix, where $r=rank(A)$. So I guess that the reduced SVD, would be the case you stated where I remove some colums and rows from the matrices. I find it hard to understand the removal of the specific columns and rows.
          – Jens Kramer
          yesterday














          It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
          – tch
          yesterday






          It seems like the first and third column of the new $U$ are not orthogonal. As for removing the rows and columns, it might help to see that the svd can also be written $A = sum_{i=1}^{n} sigma_i u_iv_i^T$ so the entries of $U$ and $V$ corresponding to zero singular values do not contribute. You can see this by noting that $Sigma V^*$ scales the rows of $V^*$, and then that $U(Sigma V^*)$ is the sum of a bunch of rank 1 outer products.
          – tch
          yesterday




















          draft saved

          draft discarded




















































          Thanks for contributing an answer to Mathematics Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.





          Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


          Please pay close attention to the following guidance:


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3060622%2ffull-and-reduced-svd-of-a-3x3-matrix%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          1300-talet

          1300-talet

          Display a custom attribute below product name in the front-end Magento 1.9.3.8