A question about variance-covariance matrix.












0












$begingroup$



If $X$ is a vector of random variables such that no element of $X$ is a linear combination of the remaining elements[i.e. there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$],then $Var(X)$ is a positive-definite matrix.




This is a Theorem from George A.F. Seber's book Linear Regression Analysis.
The proof is as follow:



For any vector $c$,we have $;c'Var[X]c=Var[c'X]ge0;$,the equality hold iff $c'X$ is a constant.



My question is that the two statement 'no element of $X$ is a linear combination of the remaining elements' and 'there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$' in the theorem are equivalent?










share|cite|improve this question









$endgroup$












  • $begingroup$
    Yes, they are equivalent because no component in $X$ belongs to the span of the remaining others
    $endgroup$
    – Legend Killer
    Dec 28 '18 at 5:57
















0












$begingroup$



If $X$ is a vector of random variables such that no element of $X$ is a linear combination of the remaining elements[i.e. there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$],then $Var(X)$ is a positive-definite matrix.




This is a Theorem from George A.F. Seber's book Linear Regression Analysis.
The proof is as follow:



For any vector $c$,we have $;c'Var[X]c=Var[c'X]ge0;$,the equality hold iff $c'X$ is a constant.



My question is that the two statement 'no element of $X$ is a linear combination of the remaining elements' and 'there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$' in the theorem are equivalent?










share|cite|improve this question









$endgroup$












  • $begingroup$
    Yes, they are equivalent because no component in $X$ belongs to the span of the remaining others
    $endgroup$
    – Legend Killer
    Dec 28 '18 at 5:57














0












0








0





$begingroup$



If $X$ is a vector of random variables such that no element of $X$ is a linear combination of the remaining elements[i.e. there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$],then $Var(X)$ is a positive-definite matrix.




This is a Theorem from George A.F. Seber's book Linear Regression Analysis.
The proof is as follow:



For any vector $c$,we have $;c'Var[X]c=Var[c'X]ge0;$,the equality hold iff $c'X$ is a constant.



My question is that the two statement 'no element of $X$ is a linear combination of the remaining elements' and 'there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$' in the theorem are equivalent?










share|cite|improve this question









$endgroup$





If $X$ is a vector of random variables such that no element of $X$ is a linear combination of the remaining elements[i.e. there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$],then $Var(X)$ is a positive-definite matrix.




This is a Theorem from George A.F. Seber's book Linear Regression Analysis.
The proof is as follow:



For any vector $c$,we have $;c'Var[X]c=Var[c'X]ge0;$,the equality hold iff $c'X$ is a constant.



My question is that the two statement 'no element of $X$ is a linear combination of the remaining elements' and 'there do not exist $a(neq0)$ and b such that $a'X=b$ for all values of $X=x$' in the theorem are equivalent?







statistics random-variables covariance






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Dec 28 '18 at 5:39









Tao XTao X

865




865












  • $begingroup$
    Yes, they are equivalent because no component in $X$ belongs to the span of the remaining others
    $endgroup$
    – Legend Killer
    Dec 28 '18 at 5:57


















  • $begingroup$
    Yes, they are equivalent because no component in $X$ belongs to the span of the remaining others
    $endgroup$
    – Legend Killer
    Dec 28 '18 at 5:57
















$begingroup$
Yes, they are equivalent because no component in $X$ belongs to the span of the remaining others
$endgroup$
– Legend Killer
Dec 28 '18 at 5:57




$begingroup$
Yes, they are equivalent because no component in $X$ belongs to the span of the remaining others
$endgroup$
– Legend Killer
Dec 28 '18 at 5:57










1 Answer
1






active

oldest

votes


















1












$begingroup$

Actually not. The second statement says that no linear combination of the variables of $X$ is a constant variable (not zero, not any other constant). This is the same as saying that the variables in $X$ together with the r.v. which is $1$ with probability $1$ are linearly independent.



For instance, if $Y$ has, say, a normal standard distribution, then $Y$ and $Y+1$ are linearly independent ($Y$ is not a linear combination of $Y+1$—that is, a scalar multiple—nor is the opposite the case). But the set ${Y,Y+1,1}$ is in fact linearly dependent, since, for instance, $Y=1cdot (Y+1)+(-1)cdot 1$.



In any case, if $vec X=(Y,Y+1)$, then $Var(vec X)$ would be positive semi-definite but not definite. Indeed,
$$var(Y)=var(Y+1)=sigma^2,$$
and also
$$cov(Y,Y+1)=cov(Y,Y)=var(Y).$$



So
$$Var(vec X)=left(begin{matrix}sigma^2 &sigma^2 \ sigma^2 &sigma^2 \end{matrix}right).$$



After all, what the property means is that definiteness of the covariance matrix is equivalent to the probability density or mass of probability of $vec X$ not to be entirely contained in any line, plane or hyperplane (passing through the origin or not).






share|cite|improve this answer











$endgroup$













    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3054609%2fa-question-about-variance-covariance-matrix%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    1












    $begingroup$

    Actually not. The second statement says that no linear combination of the variables of $X$ is a constant variable (not zero, not any other constant). This is the same as saying that the variables in $X$ together with the r.v. which is $1$ with probability $1$ are linearly independent.



    For instance, if $Y$ has, say, a normal standard distribution, then $Y$ and $Y+1$ are linearly independent ($Y$ is not a linear combination of $Y+1$—that is, a scalar multiple—nor is the opposite the case). But the set ${Y,Y+1,1}$ is in fact linearly dependent, since, for instance, $Y=1cdot (Y+1)+(-1)cdot 1$.



    In any case, if $vec X=(Y,Y+1)$, then $Var(vec X)$ would be positive semi-definite but not definite. Indeed,
    $$var(Y)=var(Y+1)=sigma^2,$$
    and also
    $$cov(Y,Y+1)=cov(Y,Y)=var(Y).$$



    So
    $$Var(vec X)=left(begin{matrix}sigma^2 &sigma^2 \ sigma^2 &sigma^2 \end{matrix}right).$$



    After all, what the property means is that definiteness of the covariance matrix is equivalent to the probability density or mass of probability of $vec X$ not to be entirely contained in any line, plane or hyperplane (passing through the origin or not).






    share|cite|improve this answer











    $endgroup$


















      1












      $begingroup$

      Actually not. The second statement says that no linear combination of the variables of $X$ is a constant variable (not zero, not any other constant). This is the same as saying that the variables in $X$ together with the r.v. which is $1$ with probability $1$ are linearly independent.



      For instance, if $Y$ has, say, a normal standard distribution, then $Y$ and $Y+1$ are linearly independent ($Y$ is not a linear combination of $Y+1$—that is, a scalar multiple—nor is the opposite the case). But the set ${Y,Y+1,1}$ is in fact linearly dependent, since, for instance, $Y=1cdot (Y+1)+(-1)cdot 1$.



      In any case, if $vec X=(Y,Y+1)$, then $Var(vec X)$ would be positive semi-definite but not definite. Indeed,
      $$var(Y)=var(Y+1)=sigma^2,$$
      and also
      $$cov(Y,Y+1)=cov(Y,Y)=var(Y).$$



      So
      $$Var(vec X)=left(begin{matrix}sigma^2 &sigma^2 \ sigma^2 &sigma^2 \end{matrix}right).$$



      After all, what the property means is that definiteness of the covariance matrix is equivalent to the probability density or mass of probability of $vec X$ not to be entirely contained in any line, plane or hyperplane (passing through the origin or not).






      share|cite|improve this answer











      $endgroup$
















        1












        1








        1





        $begingroup$

        Actually not. The second statement says that no linear combination of the variables of $X$ is a constant variable (not zero, not any other constant). This is the same as saying that the variables in $X$ together with the r.v. which is $1$ with probability $1$ are linearly independent.



        For instance, if $Y$ has, say, a normal standard distribution, then $Y$ and $Y+1$ are linearly independent ($Y$ is not a linear combination of $Y+1$—that is, a scalar multiple—nor is the opposite the case). But the set ${Y,Y+1,1}$ is in fact linearly dependent, since, for instance, $Y=1cdot (Y+1)+(-1)cdot 1$.



        In any case, if $vec X=(Y,Y+1)$, then $Var(vec X)$ would be positive semi-definite but not definite. Indeed,
        $$var(Y)=var(Y+1)=sigma^2,$$
        and also
        $$cov(Y,Y+1)=cov(Y,Y)=var(Y).$$



        So
        $$Var(vec X)=left(begin{matrix}sigma^2 &sigma^2 \ sigma^2 &sigma^2 \end{matrix}right).$$



        After all, what the property means is that definiteness of the covariance matrix is equivalent to the probability density or mass of probability of $vec X$ not to be entirely contained in any line, plane or hyperplane (passing through the origin or not).






        share|cite|improve this answer











        $endgroup$



        Actually not. The second statement says that no linear combination of the variables of $X$ is a constant variable (not zero, not any other constant). This is the same as saying that the variables in $X$ together with the r.v. which is $1$ with probability $1$ are linearly independent.



        For instance, if $Y$ has, say, a normal standard distribution, then $Y$ and $Y+1$ are linearly independent ($Y$ is not a linear combination of $Y+1$—that is, a scalar multiple—nor is the opposite the case). But the set ${Y,Y+1,1}$ is in fact linearly dependent, since, for instance, $Y=1cdot (Y+1)+(-1)cdot 1$.



        In any case, if $vec X=(Y,Y+1)$, then $Var(vec X)$ would be positive semi-definite but not definite. Indeed,
        $$var(Y)=var(Y+1)=sigma^2,$$
        and also
        $$cov(Y,Y+1)=cov(Y,Y)=var(Y).$$



        So
        $$Var(vec X)=left(begin{matrix}sigma^2 &sigma^2 \ sigma^2 &sigma^2 \end{matrix}right).$$



        After all, what the property means is that definiteness of the covariance matrix is equivalent to the probability density or mass of probability of $vec X$ not to be entirely contained in any line, plane or hyperplane (passing through the origin or not).







        share|cite|improve this answer














        share|cite|improve this answer



        share|cite|improve this answer








        edited Jan 10 at 2:40

























        answered Dec 28 '18 at 5:58









        Alejandro Nasif SalumAlejandro Nasif Salum

        4,765118




        4,765118






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3054609%2fa-question-about-variance-covariance-matrix%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Probability when a professor distributes a quiz and homework assignment to a class of n students.

            Aardman Animations

            Are they similar matrix