Why is the maximum of i.i.d. Gaussians asymptotically $sqrt{2 log n}$?











up vote
3
down vote

favorite
1












Assuming that $xi$ is bounded (as a function of $x$?), the claim is that given the equation:



$$xi frac{sqrt{2pi}}{n} = frac{1}{x} e^{-frac{x^2}{2}} left( 1 + Oleft(frac{1}{x^2} right) right) $$



one can solve ("after some calculation") for $x$ to get:



$$x = sqrt{2 log n} - frac{log log n + log 4 pi}{2 sqrt{2 log n}} - frac{log xi}{sqrt{2 log n}} + Oleft( frac{1}{log n} right) ,. $$




Question: Would it be possible to get some hints about how to solve for $x$ in this situation?




There are several issues about this which I don't understand:




  • How is the assumption that $xi$ is bounded used or otherwise relevant?

  • Why is the "imprecise knowledge of $x$ transferred to $n$" when solving for $x$, and not "transferred" to some other variable? If we require an assumption on $xi$, then why isn't the "imprecise knowledge transferred" to $xi$? (Why don't we get an $O(f(xi))$ term for some $f$?

  • Do we have to use/calculate some inverse function to $g(x) := frac{1}{x} e^{-frac{x^2}{2}}$? This function isn't even defined, strictly speaking, at $x=0$, but perhaps it has a continuous extension over the entire real line? If it does have a continuous extension over $mathbb{R}$, is that continuous extension even an invertible function, such that talking about $g^{-1}(x)$ even makes sense?

  • If we do calculate such an inverse function, do we then basically proceed by applying that function to both sides of the equation and ignoring the $O(x^{-2})$ term, with the understanding that the "uncertainty" contained within it now needs to be transferred somewhere else? If so, this leads back to the above question about where the $O((log n)^{-1})$ term could come from.


Context: I don't think the context is actually relevant to solving this problem, but for the record this comes up on p. 374 of Cramer's 1946 Mathematical Methods of Statistics where an asymptotic form for the maximum of i.i.d. Gaussian random variables is sought.










share|cite|improve this question




























    up vote
    3
    down vote

    favorite
    1












    Assuming that $xi$ is bounded (as a function of $x$?), the claim is that given the equation:



    $$xi frac{sqrt{2pi}}{n} = frac{1}{x} e^{-frac{x^2}{2}} left( 1 + Oleft(frac{1}{x^2} right) right) $$



    one can solve ("after some calculation") for $x$ to get:



    $$x = sqrt{2 log n} - frac{log log n + log 4 pi}{2 sqrt{2 log n}} - frac{log xi}{sqrt{2 log n}} + Oleft( frac{1}{log n} right) ,. $$




    Question: Would it be possible to get some hints about how to solve for $x$ in this situation?




    There are several issues about this which I don't understand:




    • How is the assumption that $xi$ is bounded used or otherwise relevant?

    • Why is the "imprecise knowledge of $x$ transferred to $n$" when solving for $x$, and not "transferred" to some other variable? If we require an assumption on $xi$, then why isn't the "imprecise knowledge transferred" to $xi$? (Why don't we get an $O(f(xi))$ term for some $f$?

    • Do we have to use/calculate some inverse function to $g(x) := frac{1}{x} e^{-frac{x^2}{2}}$? This function isn't even defined, strictly speaking, at $x=0$, but perhaps it has a continuous extension over the entire real line? If it does have a continuous extension over $mathbb{R}$, is that continuous extension even an invertible function, such that talking about $g^{-1}(x)$ even makes sense?

    • If we do calculate such an inverse function, do we then basically proceed by applying that function to both sides of the equation and ignoring the $O(x^{-2})$ term, with the understanding that the "uncertainty" contained within it now needs to be transferred somewhere else? If so, this leads back to the above question about where the $O((log n)^{-1})$ term could come from.


    Context: I don't think the context is actually relevant to solving this problem, but for the record this comes up on p. 374 of Cramer's 1946 Mathematical Methods of Statistics where an asymptotic form for the maximum of i.i.d. Gaussian random variables is sought.










    share|cite|improve this question


























      up vote
      3
      down vote

      favorite
      1









      up vote
      3
      down vote

      favorite
      1






      1





      Assuming that $xi$ is bounded (as a function of $x$?), the claim is that given the equation:



      $$xi frac{sqrt{2pi}}{n} = frac{1}{x} e^{-frac{x^2}{2}} left( 1 + Oleft(frac{1}{x^2} right) right) $$



      one can solve ("after some calculation") for $x$ to get:



      $$x = sqrt{2 log n} - frac{log log n + log 4 pi}{2 sqrt{2 log n}} - frac{log xi}{sqrt{2 log n}} + Oleft( frac{1}{log n} right) ,. $$




      Question: Would it be possible to get some hints about how to solve for $x$ in this situation?




      There are several issues about this which I don't understand:




      • How is the assumption that $xi$ is bounded used or otherwise relevant?

      • Why is the "imprecise knowledge of $x$ transferred to $n$" when solving for $x$, and not "transferred" to some other variable? If we require an assumption on $xi$, then why isn't the "imprecise knowledge transferred" to $xi$? (Why don't we get an $O(f(xi))$ term for some $f$?

      • Do we have to use/calculate some inverse function to $g(x) := frac{1}{x} e^{-frac{x^2}{2}}$? This function isn't even defined, strictly speaking, at $x=0$, but perhaps it has a continuous extension over the entire real line? If it does have a continuous extension over $mathbb{R}$, is that continuous extension even an invertible function, such that talking about $g^{-1}(x)$ even makes sense?

      • If we do calculate such an inverse function, do we then basically proceed by applying that function to both sides of the equation and ignoring the $O(x^{-2})$ term, with the understanding that the "uncertainty" contained within it now needs to be transferred somewhere else? If so, this leads back to the above question about where the $O((log n)^{-1})$ term could come from.


      Context: I don't think the context is actually relevant to solving this problem, but for the record this comes up on p. 374 of Cramer's 1946 Mathematical Methods of Statistics where an asymptotic form for the maximum of i.i.d. Gaussian random variables is sought.










      share|cite|improve this question















      Assuming that $xi$ is bounded (as a function of $x$?), the claim is that given the equation:



      $$xi frac{sqrt{2pi}}{n} = frac{1}{x} e^{-frac{x^2}{2}} left( 1 + Oleft(frac{1}{x^2} right) right) $$



      one can solve ("after some calculation") for $x$ to get:



      $$x = sqrt{2 log n} - frac{log log n + log 4 pi}{2 sqrt{2 log n}} - frac{log xi}{sqrt{2 log n}} + Oleft( frac{1}{log n} right) ,. $$




      Question: Would it be possible to get some hints about how to solve for $x$ in this situation?




      There are several issues about this which I don't understand:




      • How is the assumption that $xi$ is bounded used or otherwise relevant?

      • Why is the "imprecise knowledge of $x$ transferred to $n$" when solving for $x$, and not "transferred" to some other variable? If we require an assumption on $xi$, then why isn't the "imprecise knowledge transferred" to $xi$? (Why don't we get an $O(f(xi))$ term for some $f$?

      • Do we have to use/calculate some inverse function to $g(x) := frac{1}{x} e^{-frac{x^2}{2}}$? This function isn't even defined, strictly speaking, at $x=0$, but perhaps it has a continuous extension over the entire real line? If it does have a continuous extension over $mathbb{R}$, is that continuous extension even an invertible function, such that talking about $g^{-1}(x)$ even makes sense?

      • If we do calculate such an inverse function, do we then basically proceed by applying that function to both sides of the equation and ignoring the $O(x^{-2})$ term, with the understanding that the "uncertainty" contained within it now needs to be transferred somewhere else? If so, this leads back to the above question about where the $O((log n)^{-1})$ term could come from.


      Context: I don't think the context is actually relevant to solving this problem, but for the record this comes up on p. 374 of Cramer's 1946 Mathematical Methods of Statistics where an asymptotic form for the maximum of i.i.d. Gaussian random variables is sought.







      probability statistics asymptotics problem-solving order-statistics






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Nov 16 at 21:54

























      asked Nov 13 at 0:10









      hasManyStupidQuestions

      335




      335



























          active

          oldest

          votes











          Your Answer





          StackExchange.ifUsing("editor", function () {
          return StackExchange.using("mathjaxEditing", function () {
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          });
          });
          }, "mathjax-editing");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "69"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          convertImagesToLinks: true,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: 10,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          noCode: true, onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














           

          draft saved


          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2996092%2fwhy-is-the-maximum-of-i-i-d-gaussians-asymptotically-sqrt2-log-n%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown






























          active

          oldest

          votes













          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes
















           

          draft saved


          draft discarded



















































           


          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2996092%2fwhy-is-the-maximum-of-i-i-d-gaussians-asymptotically-sqrt2-log-n%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Probability when a professor distributes a quiz and homework assignment to a class of n students.

          Aardman Animations

          Are they similar matrix