MLE of the unknown radius Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern)MLE of a function of a parameterMLE estimation of spatial effects radiusFinding the MLE of the Pareto distribution and distributionsA question about a Theorem about MLEMLE discrete uniform distributionDefinition of asymptotic varianceMLE for 2 parameter exponential distributionJoint distribution of least square estimates $(hatalpha,hatbeta)$ in a simple linear regression modelUnderstanding the Invariance Property Proof of MLEConsistency of MLE for independent but not identically distributed data

Why are vacuum tubes still used in amateur radios?

How to tell that you are a giant?

Why does the remaining Rebel fleet at the end of Rogue One seem dramatically larger than the one in A New Hope?

What initially awakened the Balrog?

Fundamental Solution of the Pell Equation

How to get all distinct words within a set of lines?

Why is Nikon 1.4g better when Nikon 1.8g is sharper?

What is the appropriate index architecture when forced to implement IsDeleted (soft deletes)?

Generate an RGB colour grid

How much damage would a cupful of neutron star matter do to the Earth?

Performance gap between vector<bool> and array

Chinese Seal on silk painting - what does it mean?

Can the Great Weapon Master feat's damage bonus and accuracy penalty apply to attacks from the Spiritual Weapon spell?

Do I really need to have a message in a novel to appeal to readers?

Effects on objects due to a brief relocation of massive amounts of mass

How does the secondary effect of the Heat Metal spell interact with a creature resistant/immune to fire damage?

A term for a woman complaining about things/begging in a cute/childish way

Maximum summed subsequences with non-adjacent items

How to compare two different files line by line in unix?

An adverb for when you're not exaggerating

How can I reduce the gap between left and right of cdot with a macro?

Is there any word for a place full of confusion?

Can anything be seen from the center of the Boötes void? How dark would it be?

Why do we bend a book to keep it straight?



MLE of the unknown radius



Announcing the arrival of Valued Associate #679: Cesar Manara
Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern)MLE of a function of a parameterMLE estimation of spatial effects radiusFinding the MLE of the Pareto distribution and distributionsA question about a Theorem about MLEMLE discrete uniform distributionDefinition of asymptotic varianceMLE for 2 parameter exponential distributionJoint distribution of least square estimates $(hatalpha,hatbeta)$ in a simple linear regression modelUnderstanding the Invariance Property Proof of MLEConsistency of MLE for independent but not identically distributed data



.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;








1












$begingroup$


Consider this question,




Suppose that $(X_1, Y_1),(X_2, Y_2), . . . ,(X_n, Y_n)$ are the
coordinates of $n$ points chosen independently and uniformly at random
within a circle with center $(0, 0)$ and unknown radius $r$. Obtain
the MLE $hatr_n$ of $r$.




My attempt:



I have thought about the question but I am not able to put it formally. This is my reasoning.
$X_i$ and $Y_i$ are both $uniformly$ $distributed$ independent random variables on a circle of radius $r$ and center at $(0,0)$. Since $(X_i, Y_i)$ are the coordinates of the $i^textth$ point, transforming these coordinates to polar coordinates we get for the $i^textth$ point $(theta_i,a_i)$ (say) where $theta_i$ follows $Uniform(0,2pi)$ and $a_i$ follows $Uniform(0,r)$ independently. Then $a_(n) = max(a_i)$ is the MLE of $r$.



Is this reasoning correct? How do I put it formally? And if not, how should this problem be solved?










share|cite|improve this question











$endgroup$


















    1












    $begingroup$


    Consider this question,




    Suppose that $(X_1, Y_1),(X_2, Y_2), . . . ,(X_n, Y_n)$ are the
    coordinates of $n$ points chosen independently and uniformly at random
    within a circle with center $(0, 0)$ and unknown radius $r$. Obtain
    the MLE $hatr_n$ of $r$.




    My attempt:



    I have thought about the question but I am not able to put it formally. This is my reasoning.
    $X_i$ and $Y_i$ are both $uniformly$ $distributed$ independent random variables on a circle of radius $r$ and center at $(0,0)$. Since $(X_i, Y_i)$ are the coordinates of the $i^textth$ point, transforming these coordinates to polar coordinates we get for the $i^textth$ point $(theta_i,a_i)$ (say) where $theta_i$ follows $Uniform(0,2pi)$ and $a_i$ follows $Uniform(0,r)$ independently. Then $a_(n) = max(a_i)$ is the MLE of $r$.



    Is this reasoning correct? How do I put it formally? And if not, how should this problem be solved?










    share|cite|improve this question











    $endgroup$














      1












      1








      1





      $begingroup$


      Consider this question,




      Suppose that $(X_1, Y_1),(X_2, Y_2), . . . ,(X_n, Y_n)$ are the
      coordinates of $n$ points chosen independently and uniformly at random
      within a circle with center $(0, 0)$ and unknown radius $r$. Obtain
      the MLE $hatr_n$ of $r$.




      My attempt:



      I have thought about the question but I am not able to put it formally. This is my reasoning.
      $X_i$ and $Y_i$ are both $uniformly$ $distributed$ independent random variables on a circle of radius $r$ and center at $(0,0)$. Since $(X_i, Y_i)$ are the coordinates of the $i^textth$ point, transforming these coordinates to polar coordinates we get for the $i^textth$ point $(theta_i,a_i)$ (say) where $theta_i$ follows $Uniform(0,2pi)$ and $a_i$ follows $Uniform(0,r)$ independently. Then $a_(n) = max(a_i)$ is the MLE of $r$.



      Is this reasoning correct? How do I put it formally? And if not, how should this problem be solved?










      share|cite|improve this question











      $endgroup$




      Consider this question,




      Suppose that $(X_1, Y_1),(X_2, Y_2), . . . ,(X_n, Y_n)$ are the
      coordinates of $n$ points chosen independently and uniformly at random
      within a circle with center $(0, 0)$ and unknown radius $r$. Obtain
      the MLE $hatr_n$ of $r$.




      My attempt:



      I have thought about the question but I am not able to put it formally. This is my reasoning.
      $X_i$ and $Y_i$ are both $uniformly$ $distributed$ independent random variables on a circle of radius $r$ and center at $(0,0)$. Since $(X_i, Y_i)$ are the coordinates of the $i^textth$ point, transforming these coordinates to polar coordinates we get for the $i^textth$ point $(theta_i,a_i)$ (say) where $theta_i$ follows $Uniform(0,2pi)$ and $a_i$ follows $Uniform(0,r)$ independently. Then $a_(n) = max(a_i)$ is the MLE of $r$.



      Is this reasoning correct? How do I put it formally? And if not, how should this problem be solved?







      self-study estimation maximum-likelihood inference






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited 2 hours ago







      Sanket Agrawal

















      asked 3 hours ago









      Sanket AgrawalSanket Agrawal

      596




      596




















          2 Answers
          2






          active

          oldest

          votes


















          1












          $begingroup$

          Your reasoning is not quite correct: if $X$ and $Y$ were both independent uniform on $(-r, r)$, then there is positive probability that, for example, the point $(X, Y)$ is close to the point $(r, r)$, which does not lie on the circle of radius $r$ centered at the origin.




          Edit.
          I mistakenly stopped reading after the second sentence of your attempt, where you claim that each $X_i$ and $Y_i$ are independently distributed on $(-r, r)$.
          While that claim is incorrect, the logic following that claim leads to the correct solution (and is frankly more elegant than my solution).
          You may think of my solution below as a formal way to approach this problem.




          What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
          You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.



          First, let's recall a general definition of what it means to be uniformly distributed on a subset of the plane:
          Let $A$ be a subset of the plane with a defined positive area (formally, a Borel subset of $mathbbR^2$ of positive Lebesgue measure, but don't worry about that if those words aren't familiar).
          Then we say that a random vector $(X, Y)$ is distributed uniformly on $A$ if the joint density of $(X, Y)$ is given by
          $$
          f_X, Y(x, y)
          = begincases
          frac1operatornamearea(A) & textif $(x, y) in A$ \
          0 & textotherwise
          endcases
          $$

          This should remind you of the definition of a random variable uniformly distributed on an interval in $mathbbR$:
          We say that a random variable $U$ is distributed uniformly on an interval $[a, b]$ if its density is
          $$
          f_U(u)
          = begincases
          frac1b - a & textif $a le u le b$ \
          0 & textotherwise
          endcases
          $$

          which can be rewritten as
          $$
          f_U(u)
          = begincases
          frac1operatornamelength([a, b]) & textif $u in [a, b]$ \
          0 & textotherwise
          endcases
          $$

          To go from this univariate case to the bivariate case, you replace intervals by subsets of the plane, and length by area (hopefully you can see that this can be done in any number of dimensions by replacing length or area by higher dimensional volume).



          Back to the question at hand.
          What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
          You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.
          Let's call this circle $C_r$; it has area $pi r^2$.
          By what we said above, if $(X, Y)$ is distributed uniformly on $C_r$, then the joint density of $(X, Y)$ is
          $$
          beginaligned
          f_X, Y(x, y mid r)
          &= begincases
          frac1operatornamearea(C_r) & textif $(x, y) in C_r$ \
          0 & textotherwise
          endcases \
          &= begincases
          frac1pi r^2 & textif $x^2 + y^2 le r^2$ \
          0 & textotherwise
          endcases \
          &= frac1pi r^2 mathbf1_C_r(x, y),
          endaligned
          $$

          where $mathbf1_A$ denotes the indicator function of the set $A$.



          If we now have a sample $(X_1, Y_1), ldots, (X_n, Y_n)$ of random vectors independently distributed uniformly on $C_r$ for some unknown $r > 0$, then we can estimate $r$ by maximum likelihood.
          By independence, the likelihood of the sample is
          $$
          beginaligned
          L(r)
          &= prod_i=1^n f_X_i, Y_i(X_i, Y_i mid r) \
          &= prod_i=1^n frac1pi r^2 mathbf1_C_r(X_i, Y_i) \
          &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
          &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
          &= begincases
          frac1pi^n r^2n & textif $X_i^2 + Y^2 le r^2$ for all $i$ \
          0 & textotherwise
          endcases \
          &= begincases
          frac1pi^n r^2n & textif $maxlimits_i (X_i^2 + Y^2) le r^2$ \
          0 & textotherwise
          endcases \
          &= frac1pi^n r^2n mathbf1_[0, r^2]left(maxlimits_i (X_i^2 + Y_i^2)right) \
          &= frac1pi^n r^2n mathbf1_[0, r]left(sqrtmaxlimits_i (X_i^2 + Y_i^2)right) \
          &= frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
          endaligned
          $$

          (By the way, this shows that $max_isqrtX_i^2 + Y_i^2$ is a sufficient statistic!)
          To summarize,
          $$
          L(r)
          = frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
          $$

          Now to find the maximum likelihood estimator $widehatr_n$ of $n$, we have to maximize $L(r)$ with respect to $r$.
          Notice that the indicator function in the final formula for $L(r)$ is either $0$ or $1$, and $1/(pi^n r^2 n) to infty$ as $r downarrow 0$.
          I'll leave it to you to use these clues to figure out the formula for the MLE; you should get a relatively simple (and, in my opinion, intuitive) expression.






          share|cite|improve this answer











          $endgroup$












          • $begingroup$
            I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
            $endgroup$
            – Sanket Agrawal
            2 hours ago










          • $begingroup$
            thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
            $endgroup$
            – Sanket Agrawal
            2 hours ago






          • 1




            $begingroup$
            @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
            $endgroup$
            – Artem Mavrin
            2 hours ago










          • $begingroup$
            okay @ArtemMavrin, thank you.
            $endgroup$
            – Sanket Agrawal
            2 hours ago



















          3












          $begingroup$

          This question is substantially simplified if we recognise that the points give information about the radius $r$ only through their distance from the centre (i.e., the zero point). Let $R_1,R_2,R_3,...$ be these distance values corresponding to each of the points and let $A(r) = pi cdot r^2$ be the area of a circle with radius $r$. Since each sample point is uniform in the sphere we must have:



          $$mathbbP(R_i leq t) = fracA(t)A(r) = fracpi cdot t^2pi cdot r^2 = Big( fractr Big)^2
          quad quad quad textfor all 0 leq t leq r.$$



          This gives the corresponding sample density $p_r(t) = 2t/r^2$ over the support $0 leq t leq r$. As you can see, this sampling density is not uniform over the support. This is actually unsurprising --- the probability density for the distance $R_i$ is proportional to the circumference of a circle with radius equal to that distance.




          The maximum likelihood estimator: The likelihood function for $n$ observed data points is:



          $$L_mathbfr(r) = prod_i=1^n p_r(r_i) propto frac1r^2 cdot mathbbI(r geq r_(n)).$$



          We can see that the likelihood function is strictly decreasing in $r$, so the maximum likelihood estimator (MLE) occurs at the boundary point $hatr_n = r_(n)$. That is, the MLE for the true radius is the maximum of the lengths from the zero point to the sample points. Since the true parameter $r$ is at least as large as the MLE, the MLE is biased, and will tend to underestimate the true radius. Specifically, we have:



          $$beginequation beginaligned
          mathbbE(hatR_n) = mathbbE(R_(n))
          &= int limits_0^r mathbbP(R_(n) > t) dt \[6pt]
          &= int limits_0^r (1 - mathbbP(R_(n) leq t)) dt \[6pt]
          &= int limits_0^r Big( 1 - fract^2nr^2n Big) dt \[6pt]
          &= Bigg[ t - frac12n+1 cdot fract^2n+1r^2n Bigg]_t=0^t=r \[6pt]
          &= r Big( 1 - frac12n+1 Big) \[6pt]
          &= frac2n2n+1 cdot r. \[6pt]
          endaligned endequation$$



          Thus, a bias-corrected scaled version of the MLE is:



          $$tilder_n = frac2n+12n cdot r_(n).$$



          Incidentally, this process can easily be extended to a hypersphere in higher dimensions. In each case the MLE is the maximum distance from the zero point. For a hypershere in $k$ dimensions, the bias-corrected scaled MLE is:



          $$tilder_n = frackn+1kn cdot r_(n).$$






          share|cite|improve this answer









          $endgroup$













            Your Answer








            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "65"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: false,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: null,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f403914%2fmle-of-the-unknown-radius%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            2 Answers
            2






            active

            oldest

            votes








            2 Answers
            2






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            1












            $begingroup$

            Your reasoning is not quite correct: if $X$ and $Y$ were both independent uniform on $(-r, r)$, then there is positive probability that, for example, the point $(X, Y)$ is close to the point $(r, r)$, which does not lie on the circle of radius $r$ centered at the origin.




            Edit.
            I mistakenly stopped reading after the second sentence of your attempt, where you claim that each $X_i$ and $Y_i$ are independently distributed on $(-r, r)$.
            While that claim is incorrect, the logic following that claim leads to the correct solution (and is frankly more elegant than my solution).
            You may think of my solution below as a formal way to approach this problem.




            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.



            First, let's recall a general definition of what it means to be uniformly distributed on a subset of the plane:
            Let $A$ be a subset of the plane with a defined positive area (formally, a Borel subset of $mathbbR^2$ of positive Lebesgue measure, but don't worry about that if those words aren't familiar).
            Then we say that a random vector $(X, Y)$ is distributed uniformly on $A$ if the joint density of $(X, Y)$ is given by
            $$
            f_X, Y(x, y)
            = begincases
            frac1operatornamearea(A) & textif $(x, y) in A$ \
            0 & textotherwise
            endcases
            $$

            This should remind you of the definition of a random variable uniformly distributed on an interval in $mathbbR$:
            We say that a random variable $U$ is distributed uniformly on an interval $[a, b]$ if its density is
            $$
            f_U(u)
            = begincases
            frac1b - a & textif $a le u le b$ \
            0 & textotherwise
            endcases
            $$

            which can be rewritten as
            $$
            f_U(u)
            = begincases
            frac1operatornamelength([a, b]) & textif $u in [a, b]$ \
            0 & textotherwise
            endcases
            $$

            To go from this univariate case to the bivariate case, you replace intervals by subsets of the plane, and length by area (hopefully you can see that this can be done in any number of dimensions by replacing length or area by higher dimensional volume).



            Back to the question at hand.
            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.
            Let's call this circle $C_r$; it has area $pi r^2$.
            By what we said above, if $(X, Y)$ is distributed uniformly on $C_r$, then the joint density of $(X, Y)$ is
            $$
            beginaligned
            f_X, Y(x, y mid r)
            &= begincases
            frac1operatornamearea(C_r) & textif $(x, y) in C_r$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi r^2 & textif $x^2 + y^2 le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi r^2 mathbf1_C_r(x, y),
            endaligned
            $$

            where $mathbf1_A$ denotes the indicator function of the set $A$.



            If we now have a sample $(X_1, Y_1), ldots, (X_n, Y_n)$ of random vectors independently distributed uniformly on $C_r$ for some unknown $r > 0$, then we can estimate $r$ by maximum likelihood.
            By independence, the likelihood of the sample is
            $$
            beginaligned
            L(r)
            &= prod_i=1^n f_X_i, Y_i(X_i, Y_i mid r) \
            &= prod_i=1^n frac1pi r^2 mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= begincases
            frac1pi^n r^2n & textif $X_i^2 + Y^2 le r^2$ for all $i$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi^n r^2n & textif $maxlimits_i (X_i^2 + Y^2) le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi^n r^2n mathbf1_[0, r^2]left(maxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(sqrtmaxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            endaligned
            $$

            (By the way, this shows that $max_isqrtX_i^2 + Y_i^2$ is a sufficient statistic!)
            To summarize,
            $$
            L(r)
            = frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            $$

            Now to find the maximum likelihood estimator $widehatr_n$ of $n$, we have to maximize $L(r)$ with respect to $r$.
            Notice that the indicator function in the final formula for $L(r)$ is either $0$ or $1$, and $1/(pi^n r^2 n) to infty$ as $r downarrow 0$.
            I'll leave it to you to use these clues to figure out the formula for the MLE; you should get a relatively simple (and, in my opinion, intuitive) expression.






            share|cite|improve this answer











            $endgroup$












            • $begingroup$
              I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
              $endgroup$
              – Sanket Agrawal
              2 hours ago










            • $begingroup$
              thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
              $endgroup$
              – Sanket Agrawal
              2 hours ago






            • 1




              $begingroup$
              @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
              $endgroup$
              – Artem Mavrin
              2 hours ago










            • $begingroup$
              okay @ArtemMavrin, thank you.
              $endgroup$
              – Sanket Agrawal
              2 hours ago
















            1












            $begingroup$

            Your reasoning is not quite correct: if $X$ and $Y$ were both independent uniform on $(-r, r)$, then there is positive probability that, for example, the point $(X, Y)$ is close to the point $(r, r)$, which does not lie on the circle of radius $r$ centered at the origin.




            Edit.
            I mistakenly stopped reading after the second sentence of your attempt, where you claim that each $X_i$ and $Y_i$ are independently distributed on $(-r, r)$.
            While that claim is incorrect, the logic following that claim leads to the correct solution (and is frankly more elegant than my solution).
            You may think of my solution below as a formal way to approach this problem.




            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.



            First, let's recall a general definition of what it means to be uniformly distributed on a subset of the plane:
            Let $A$ be a subset of the plane with a defined positive area (formally, a Borel subset of $mathbbR^2$ of positive Lebesgue measure, but don't worry about that if those words aren't familiar).
            Then we say that a random vector $(X, Y)$ is distributed uniformly on $A$ if the joint density of $(X, Y)$ is given by
            $$
            f_X, Y(x, y)
            = begincases
            frac1operatornamearea(A) & textif $(x, y) in A$ \
            0 & textotherwise
            endcases
            $$

            This should remind you of the definition of a random variable uniformly distributed on an interval in $mathbbR$:
            We say that a random variable $U$ is distributed uniformly on an interval $[a, b]$ if its density is
            $$
            f_U(u)
            = begincases
            frac1b - a & textif $a le u le b$ \
            0 & textotherwise
            endcases
            $$

            which can be rewritten as
            $$
            f_U(u)
            = begincases
            frac1operatornamelength([a, b]) & textif $u in [a, b]$ \
            0 & textotherwise
            endcases
            $$

            To go from this univariate case to the bivariate case, you replace intervals by subsets of the plane, and length by area (hopefully you can see that this can be done in any number of dimensions by replacing length or area by higher dimensional volume).



            Back to the question at hand.
            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.
            Let's call this circle $C_r$; it has area $pi r^2$.
            By what we said above, if $(X, Y)$ is distributed uniformly on $C_r$, then the joint density of $(X, Y)$ is
            $$
            beginaligned
            f_X, Y(x, y mid r)
            &= begincases
            frac1operatornamearea(C_r) & textif $(x, y) in C_r$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi r^2 & textif $x^2 + y^2 le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi r^2 mathbf1_C_r(x, y),
            endaligned
            $$

            where $mathbf1_A$ denotes the indicator function of the set $A$.



            If we now have a sample $(X_1, Y_1), ldots, (X_n, Y_n)$ of random vectors independently distributed uniformly on $C_r$ for some unknown $r > 0$, then we can estimate $r$ by maximum likelihood.
            By independence, the likelihood of the sample is
            $$
            beginaligned
            L(r)
            &= prod_i=1^n f_X_i, Y_i(X_i, Y_i mid r) \
            &= prod_i=1^n frac1pi r^2 mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= begincases
            frac1pi^n r^2n & textif $X_i^2 + Y^2 le r^2$ for all $i$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi^n r^2n & textif $maxlimits_i (X_i^2 + Y^2) le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi^n r^2n mathbf1_[0, r^2]left(maxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(sqrtmaxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            endaligned
            $$

            (By the way, this shows that $max_isqrtX_i^2 + Y_i^2$ is a sufficient statistic!)
            To summarize,
            $$
            L(r)
            = frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            $$

            Now to find the maximum likelihood estimator $widehatr_n$ of $n$, we have to maximize $L(r)$ with respect to $r$.
            Notice that the indicator function in the final formula for $L(r)$ is either $0$ or $1$, and $1/(pi^n r^2 n) to infty$ as $r downarrow 0$.
            I'll leave it to you to use these clues to figure out the formula for the MLE; you should get a relatively simple (and, in my opinion, intuitive) expression.






            share|cite|improve this answer











            $endgroup$












            • $begingroup$
              I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
              $endgroup$
              – Sanket Agrawal
              2 hours ago










            • $begingroup$
              thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
              $endgroup$
              – Sanket Agrawal
              2 hours ago






            • 1




              $begingroup$
              @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
              $endgroup$
              – Artem Mavrin
              2 hours ago










            • $begingroup$
              okay @ArtemMavrin, thank you.
              $endgroup$
              – Sanket Agrawal
              2 hours ago














            1












            1








            1





            $begingroup$

            Your reasoning is not quite correct: if $X$ and $Y$ were both independent uniform on $(-r, r)$, then there is positive probability that, for example, the point $(X, Y)$ is close to the point $(r, r)$, which does not lie on the circle of radius $r$ centered at the origin.




            Edit.
            I mistakenly stopped reading after the second sentence of your attempt, where you claim that each $X_i$ and $Y_i$ are independently distributed on $(-r, r)$.
            While that claim is incorrect, the logic following that claim leads to the correct solution (and is frankly more elegant than my solution).
            You may think of my solution below as a formal way to approach this problem.




            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.



            First, let's recall a general definition of what it means to be uniformly distributed on a subset of the plane:
            Let $A$ be a subset of the plane with a defined positive area (formally, a Borel subset of $mathbbR^2$ of positive Lebesgue measure, but don't worry about that if those words aren't familiar).
            Then we say that a random vector $(X, Y)$ is distributed uniformly on $A$ if the joint density of $(X, Y)$ is given by
            $$
            f_X, Y(x, y)
            = begincases
            frac1operatornamearea(A) & textif $(x, y) in A$ \
            0 & textotherwise
            endcases
            $$

            This should remind you of the definition of a random variable uniformly distributed on an interval in $mathbbR$:
            We say that a random variable $U$ is distributed uniformly on an interval $[a, b]$ if its density is
            $$
            f_U(u)
            = begincases
            frac1b - a & textif $a le u le b$ \
            0 & textotherwise
            endcases
            $$

            which can be rewritten as
            $$
            f_U(u)
            = begincases
            frac1operatornamelength([a, b]) & textif $u in [a, b]$ \
            0 & textotherwise
            endcases
            $$

            To go from this univariate case to the bivariate case, you replace intervals by subsets of the plane, and length by area (hopefully you can see that this can be done in any number of dimensions by replacing length or area by higher dimensional volume).



            Back to the question at hand.
            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.
            Let's call this circle $C_r$; it has area $pi r^2$.
            By what we said above, if $(X, Y)$ is distributed uniformly on $C_r$, then the joint density of $(X, Y)$ is
            $$
            beginaligned
            f_X, Y(x, y mid r)
            &= begincases
            frac1operatornamearea(C_r) & textif $(x, y) in C_r$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi r^2 & textif $x^2 + y^2 le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi r^2 mathbf1_C_r(x, y),
            endaligned
            $$

            where $mathbf1_A$ denotes the indicator function of the set $A$.



            If we now have a sample $(X_1, Y_1), ldots, (X_n, Y_n)$ of random vectors independently distributed uniformly on $C_r$ for some unknown $r > 0$, then we can estimate $r$ by maximum likelihood.
            By independence, the likelihood of the sample is
            $$
            beginaligned
            L(r)
            &= prod_i=1^n f_X_i, Y_i(X_i, Y_i mid r) \
            &= prod_i=1^n frac1pi r^2 mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= begincases
            frac1pi^n r^2n & textif $X_i^2 + Y^2 le r^2$ for all $i$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi^n r^2n & textif $maxlimits_i (X_i^2 + Y^2) le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi^n r^2n mathbf1_[0, r^2]left(maxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(sqrtmaxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            endaligned
            $$

            (By the way, this shows that $max_isqrtX_i^2 + Y_i^2$ is a sufficient statistic!)
            To summarize,
            $$
            L(r)
            = frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            $$

            Now to find the maximum likelihood estimator $widehatr_n$ of $n$, we have to maximize $L(r)$ with respect to $r$.
            Notice that the indicator function in the final formula for $L(r)$ is either $0$ or $1$, and $1/(pi^n r^2 n) to infty$ as $r downarrow 0$.
            I'll leave it to you to use these clues to figure out the formula for the MLE; you should get a relatively simple (and, in my opinion, intuitive) expression.






            share|cite|improve this answer











            $endgroup$



            Your reasoning is not quite correct: if $X$ and $Y$ were both independent uniform on $(-r, r)$, then there is positive probability that, for example, the point $(X, Y)$ is close to the point $(r, r)$, which does not lie on the circle of radius $r$ centered at the origin.




            Edit.
            I mistakenly stopped reading after the second sentence of your attempt, where you claim that each $X_i$ and $Y_i$ are independently distributed on $(-r, r)$.
            While that claim is incorrect, the logic following that claim leads to the correct solution (and is frankly more elegant than my solution).
            You may think of my solution below as a formal way to approach this problem.




            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.



            First, let's recall a general definition of what it means to be uniformly distributed on a subset of the plane:
            Let $A$ be a subset of the plane with a defined positive area (formally, a Borel subset of $mathbbR^2$ of positive Lebesgue measure, but don't worry about that if those words aren't familiar).
            Then we say that a random vector $(X, Y)$ is distributed uniformly on $A$ if the joint density of $(X, Y)$ is given by
            $$
            f_X, Y(x, y)
            = begincases
            frac1operatornamearea(A) & textif $(x, y) in A$ \
            0 & textotherwise
            endcases
            $$

            This should remind you of the definition of a random variable uniformly distributed on an interval in $mathbbR$:
            We say that a random variable $U$ is distributed uniformly on an interval $[a, b]$ if its density is
            $$
            f_U(u)
            = begincases
            frac1b - a & textif $a le u le b$ \
            0 & textotherwise
            endcases
            $$

            which can be rewritten as
            $$
            f_U(u)
            = begincases
            frac1operatornamelength([a, b]) & textif $u in [a, b]$ \
            0 & textotherwise
            endcases
            $$

            To go from this univariate case to the bivariate case, you replace intervals by subsets of the plane, and length by area (hopefully you can see that this can be done in any number of dimensions by replacing length or area by higher dimensional volume).



            Back to the question at hand.
            What you need to do instead is figure out the joint distribution of each pair $(X, Y)$.
            You are told that $(X, Y)$ is distributed uniformly on the circle of radius $r$ centered at $0$.
            Let's call this circle $C_r$; it has area $pi r^2$.
            By what we said above, if $(X, Y)$ is distributed uniformly on $C_r$, then the joint density of $(X, Y)$ is
            $$
            beginaligned
            f_X, Y(x, y mid r)
            &= begincases
            frac1operatornamearea(C_r) & textif $(x, y) in C_r$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi r^2 & textif $x^2 + y^2 le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi r^2 mathbf1_C_r(x, y),
            endaligned
            $$

            where $mathbf1_A$ denotes the indicator function of the set $A$.



            If we now have a sample $(X_1, Y_1), ldots, (X_n, Y_n)$ of random vectors independently distributed uniformly on $C_r$ for some unknown $r > 0$, then we can estimate $r$ by maximum likelihood.
            By independence, the likelihood of the sample is
            $$
            beginaligned
            L(r)
            &= prod_i=1^n f_X_i, Y_i(X_i, Y_i mid r) \
            &= prod_i=1^n frac1pi r^2 mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= frac1pi^n r^2 n prod_i=1^n mathbf1_C_r(X_i, Y_i) \
            &= begincases
            frac1pi^n r^2n & textif $X_i^2 + Y^2 le r^2$ for all $i$ \
            0 & textotherwise
            endcases \
            &= begincases
            frac1pi^n r^2n & textif $maxlimits_i (X_i^2 + Y^2) le r^2$ \
            0 & textotherwise
            endcases \
            &= frac1pi^n r^2n mathbf1_[0, r^2]left(maxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(sqrtmaxlimits_i (X_i^2 + Y_i^2)right) \
            &= frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            endaligned
            $$

            (By the way, this shows that $max_isqrtX_i^2 + Y_i^2$ is a sufficient statistic!)
            To summarize,
            $$
            L(r)
            = frac1pi^n r^2n mathbf1_[0, r]left(maxlimits_i sqrtX_i^2 + Y_i^2right)
            $$

            Now to find the maximum likelihood estimator $widehatr_n$ of $n$, we have to maximize $L(r)$ with respect to $r$.
            Notice that the indicator function in the final formula for $L(r)$ is either $0$ or $1$, and $1/(pi^n r^2 n) to infty$ as $r downarrow 0$.
            I'll leave it to you to use these clues to figure out the formula for the MLE; you should get a relatively simple (and, in my opinion, intuitive) expression.







            share|cite|improve this answer














            share|cite|improve this answer



            share|cite|improve this answer








            edited 2 hours ago

























            answered 3 hours ago









            Artem MavrinArtem Mavrin

            1,328714




            1,328714











            • $begingroup$
              I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
              $endgroup$
              – Sanket Agrawal
              2 hours ago










            • $begingroup$
              thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
              $endgroup$
              – Sanket Agrawal
              2 hours ago






            • 1




              $begingroup$
              @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
              $endgroup$
              – Artem Mavrin
              2 hours ago










            • $begingroup$
              okay @ArtemMavrin, thank you.
              $endgroup$
              – Sanket Agrawal
              2 hours ago

















            • $begingroup$
              I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
              $endgroup$
              – Sanket Agrawal
              2 hours ago










            • $begingroup$
              thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
              $endgroup$
              – Sanket Agrawal
              2 hours ago






            • 1




              $begingroup$
              @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
              $endgroup$
              – Artem Mavrin
              2 hours ago










            • $begingroup$
              okay @ArtemMavrin, thank you.
              $endgroup$
              – Sanket Agrawal
              2 hours ago
















            $begingroup$
            I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
            $endgroup$
            – Sanket Agrawal
            2 hours ago




            $begingroup$
            I am so sorry, I actually meant that $(X_i,Y_i)$ are uniformly chosen from the given circle of radius $r$ and not from the square. It must have slipped while I was typing. I have fixed that.
            $endgroup$
            – Sanket Agrawal
            2 hours ago












            $begingroup$
            thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
            $endgroup$
            – Sanket Agrawal
            2 hours ago




            $begingroup$
            thanks for this answer. However you also say that my logic is also correct. How can I build up on that?
            $endgroup$
            – Sanket Agrawal
            2 hours ago




            1




            1




            $begingroup$
            @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
            $endgroup$
            – Artem Mavrin
            2 hours ago




            $begingroup$
            @SanketAgrawal I see, then your solution looks good to me. The only thing potentially missing is explaining why the polar coordinate variables $(theta_i, a_i)$ are independent and uniformly distributed, which can be done by considering the transformation $(x, y) mapsto (theta, a)$
            $endgroup$
            – Artem Mavrin
            2 hours ago












            $begingroup$
            okay @ArtemMavrin, thank you.
            $endgroup$
            – Sanket Agrawal
            2 hours ago





            $begingroup$
            okay @ArtemMavrin, thank you.
            $endgroup$
            – Sanket Agrawal
            2 hours ago














            3












            $begingroup$

            This question is substantially simplified if we recognise that the points give information about the radius $r$ only through their distance from the centre (i.e., the zero point). Let $R_1,R_2,R_3,...$ be these distance values corresponding to each of the points and let $A(r) = pi cdot r^2$ be the area of a circle with radius $r$. Since each sample point is uniform in the sphere we must have:



            $$mathbbP(R_i leq t) = fracA(t)A(r) = fracpi cdot t^2pi cdot r^2 = Big( fractr Big)^2
            quad quad quad textfor all 0 leq t leq r.$$



            This gives the corresponding sample density $p_r(t) = 2t/r^2$ over the support $0 leq t leq r$. As you can see, this sampling density is not uniform over the support. This is actually unsurprising --- the probability density for the distance $R_i$ is proportional to the circumference of a circle with radius equal to that distance.




            The maximum likelihood estimator: The likelihood function for $n$ observed data points is:



            $$L_mathbfr(r) = prod_i=1^n p_r(r_i) propto frac1r^2 cdot mathbbI(r geq r_(n)).$$



            We can see that the likelihood function is strictly decreasing in $r$, so the maximum likelihood estimator (MLE) occurs at the boundary point $hatr_n = r_(n)$. That is, the MLE for the true radius is the maximum of the lengths from the zero point to the sample points. Since the true parameter $r$ is at least as large as the MLE, the MLE is biased, and will tend to underestimate the true radius. Specifically, we have:



            $$beginequation beginaligned
            mathbbE(hatR_n) = mathbbE(R_(n))
            &= int limits_0^r mathbbP(R_(n) > t) dt \[6pt]
            &= int limits_0^r (1 - mathbbP(R_(n) leq t)) dt \[6pt]
            &= int limits_0^r Big( 1 - fract^2nr^2n Big) dt \[6pt]
            &= Bigg[ t - frac12n+1 cdot fract^2n+1r^2n Bigg]_t=0^t=r \[6pt]
            &= r Big( 1 - frac12n+1 Big) \[6pt]
            &= frac2n2n+1 cdot r. \[6pt]
            endaligned endequation$$



            Thus, a bias-corrected scaled version of the MLE is:



            $$tilder_n = frac2n+12n cdot r_(n).$$



            Incidentally, this process can easily be extended to a hypersphere in higher dimensions. In each case the MLE is the maximum distance from the zero point. For a hypershere in $k$ dimensions, the bias-corrected scaled MLE is:



            $$tilder_n = frackn+1kn cdot r_(n).$$






            share|cite|improve this answer









            $endgroup$

















              3












              $begingroup$

              This question is substantially simplified if we recognise that the points give information about the radius $r$ only through their distance from the centre (i.e., the zero point). Let $R_1,R_2,R_3,...$ be these distance values corresponding to each of the points and let $A(r) = pi cdot r^2$ be the area of a circle with radius $r$. Since each sample point is uniform in the sphere we must have:



              $$mathbbP(R_i leq t) = fracA(t)A(r) = fracpi cdot t^2pi cdot r^2 = Big( fractr Big)^2
              quad quad quad textfor all 0 leq t leq r.$$



              This gives the corresponding sample density $p_r(t) = 2t/r^2$ over the support $0 leq t leq r$. As you can see, this sampling density is not uniform over the support. This is actually unsurprising --- the probability density for the distance $R_i$ is proportional to the circumference of a circle with radius equal to that distance.




              The maximum likelihood estimator: The likelihood function for $n$ observed data points is:



              $$L_mathbfr(r) = prod_i=1^n p_r(r_i) propto frac1r^2 cdot mathbbI(r geq r_(n)).$$



              We can see that the likelihood function is strictly decreasing in $r$, so the maximum likelihood estimator (MLE) occurs at the boundary point $hatr_n = r_(n)$. That is, the MLE for the true radius is the maximum of the lengths from the zero point to the sample points. Since the true parameter $r$ is at least as large as the MLE, the MLE is biased, and will tend to underestimate the true radius. Specifically, we have:



              $$beginequation beginaligned
              mathbbE(hatR_n) = mathbbE(R_(n))
              &= int limits_0^r mathbbP(R_(n) > t) dt \[6pt]
              &= int limits_0^r (1 - mathbbP(R_(n) leq t)) dt \[6pt]
              &= int limits_0^r Big( 1 - fract^2nr^2n Big) dt \[6pt]
              &= Bigg[ t - frac12n+1 cdot fract^2n+1r^2n Bigg]_t=0^t=r \[6pt]
              &= r Big( 1 - frac12n+1 Big) \[6pt]
              &= frac2n2n+1 cdot r. \[6pt]
              endaligned endequation$$



              Thus, a bias-corrected scaled version of the MLE is:



              $$tilder_n = frac2n+12n cdot r_(n).$$



              Incidentally, this process can easily be extended to a hypersphere in higher dimensions. In each case the MLE is the maximum distance from the zero point. For a hypershere in $k$ dimensions, the bias-corrected scaled MLE is:



              $$tilder_n = frackn+1kn cdot r_(n).$$






              share|cite|improve this answer









              $endgroup$















                3












                3








                3





                $begingroup$

                This question is substantially simplified if we recognise that the points give information about the radius $r$ only through their distance from the centre (i.e., the zero point). Let $R_1,R_2,R_3,...$ be these distance values corresponding to each of the points and let $A(r) = pi cdot r^2$ be the area of a circle with radius $r$. Since each sample point is uniform in the sphere we must have:



                $$mathbbP(R_i leq t) = fracA(t)A(r) = fracpi cdot t^2pi cdot r^2 = Big( fractr Big)^2
                quad quad quad textfor all 0 leq t leq r.$$



                This gives the corresponding sample density $p_r(t) = 2t/r^2$ over the support $0 leq t leq r$. As you can see, this sampling density is not uniform over the support. This is actually unsurprising --- the probability density for the distance $R_i$ is proportional to the circumference of a circle with radius equal to that distance.




                The maximum likelihood estimator: The likelihood function for $n$ observed data points is:



                $$L_mathbfr(r) = prod_i=1^n p_r(r_i) propto frac1r^2 cdot mathbbI(r geq r_(n)).$$



                We can see that the likelihood function is strictly decreasing in $r$, so the maximum likelihood estimator (MLE) occurs at the boundary point $hatr_n = r_(n)$. That is, the MLE for the true radius is the maximum of the lengths from the zero point to the sample points. Since the true parameter $r$ is at least as large as the MLE, the MLE is biased, and will tend to underestimate the true radius. Specifically, we have:



                $$beginequation beginaligned
                mathbbE(hatR_n) = mathbbE(R_(n))
                &= int limits_0^r mathbbP(R_(n) > t) dt \[6pt]
                &= int limits_0^r (1 - mathbbP(R_(n) leq t)) dt \[6pt]
                &= int limits_0^r Big( 1 - fract^2nr^2n Big) dt \[6pt]
                &= Bigg[ t - frac12n+1 cdot fract^2n+1r^2n Bigg]_t=0^t=r \[6pt]
                &= r Big( 1 - frac12n+1 Big) \[6pt]
                &= frac2n2n+1 cdot r. \[6pt]
                endaligned endequation$$



                Thus, a bias-corrected scaled version of the MLE is:



                $$tilder_n = frac2n+12n cdot r_(n).$$



                Incidentally, this process can easily be extended to a hypersphere in higher dimensions. In each case the MLE is the maximum distance from the zero point. For a hypershere in $k$ dimensions, the bias-corrected scaled MLE is:



                $$tilder_n = frackn+1kn cdot r_(n).$$






                share|cite|improve this answer









                $endgroup$



                This question is substantially simplified if we recognise that the points give information about the radius $r$ only through their distance from the centre (i.e., the zero point). Let $R_1,R_2,R_3,...$ be these distance values corresponding to each of the points and let $A(r) = pi cdot r^2$ be the area of a circle with radius $r$. Since each sample point is uniform in the sphere we must have:



                $$mathbbP(R_i leq t) = fracA(t)A(r) = fracpi cdot t^2pi cdot r^2 = Big( fractr Big)^2
                quad quad quad textfor all 0 leq t leq r.$$



                This gives the corresponding sample density $p_r(t) = 2t/r^2$ over the support $0 leq t leq r$. As you can see, this sampling density is not uniform over the support. This is actually unsurprising --- the probability density for the distance $R_i$ is proportional to the circumference of a circle with radius equal to that distance.




                The maximum likelihood estimator: The likelihood function for $n$ observed data points is:



                $$L_mathbfr(r) = prod_i=1^n p_r(r_i) propto frac1r^2 cdot mathbbI(r geq r_(n)).$$



                We can see that the likelihood function is strictly decreasing in $r$, so the maximum likelihood estimator (MLE) occurs at the boundary point $hatr_n = r_(n)$. That is, the MLE for the true radius is the maximum of the lengths from the zero point to the sample points. Since the true parameter $r$ is at least as large as the MLE, the MLE is biased, and will tend to underestimate the true radius. Specifically, we have:



                $$beginequation beginaligned
                mathbbE(hatR_n) = mathbbE(R_(n))
                &= int limits_0^r mathbbP(R_(n) > t) dt \[6pt]
                &= int limits_0^r (1 - mathbbP(R_(n) leq t)) dt \[6pt]
                &= int limits_0^r Big( 1 - fract^2nr^2n Big) dt \[6pt]
                &= Bigg[ t - frac12n+1 cdot fract^2n+1r^2n Bigg]_t=0^t=r \[6pt]
                &= r Big( 1 - frac12n+1 Big) \[6pt]
                &= frac2n2n+1 cdot r. \[6pt]
                endaligned endequation$$



                Thus, a bias-corrected scaled version of the MLE is:



                $$tilder_n = frac2n+12n cdot r_(n).$$



                Incidentally, this process can easily be extended to a hypersphere in higher dimensions. In each case the MLE is the maximum distance from the zero point. For a hypershere in $k$ dimensions, the bias-corrected scaled MLE is:



                $$tilder_n = frackn+1kn cdot r_(n).$$







                share|cite|improve this answer












                share|cite|improve this answer



                share|cite|improve this answer










                answered 41 mins ago









                BenBen

                28.8k233129




                28.8k233129



























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Cross Validated!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f403914%2fmle-of-the-unknown-radius%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Isurus Índice Especies | Notas | Véxase tamén | Menú de navegación"A compendium of fossil marine animal genera (Chondrichthyes entry)"o orixinal"A review of the Tertiary fossil Cetacea (Mammalia) localities in wales port taf Museum Victoria"o orixinalThe Vertebrate Fauna of the Selma Formation of Alabama. Part VII. Part VIII. The Mosasaurs The Fishes50419737IDsh85068767Isurus2548834613242066569678159923NHMSYS00210535017845105743

                    Wolfenstein 3D Contents Availability Essential improvements Game data Video settings Input settings Audio settings Network VR support Issues fixed Other information System requirements NotesReferences    3D Realms Wolfenstein 3D pageGOG.com Community DiscussionsGOG.com Support PageSteam Community DiscussionsWolfenstein WikiOfficial websiteAmazon.comBethesda.netGamersGateGOG.comGreen Man GamingHumble StoreSteamweb browser versionWolfenstein 3D: Super UpgradesherehereUltraWolfhereWolfMenuECWolf Wolf4SDL WolfGL WinWolf3d NewWolf BetterWolf Sprite Fix and Rotation Project    Wolfenstein 3D VRSplitWolfWolfenstein 3D VRWolfenstein 3D VRWolfenstein 3D VR4DOS command shellFreeDOS's MORE.COMMacBin themthis shim fileWine regeditRELEASE: QUAKE II + III, WOLFENSTEIN 3D, RETURN TO CASTLE WOLFENSTEIN - GOG.com NewsMac Family - Wolfenstein Wiki - WikiaNerdly Pleasures: How many FPS? - DOS Games and Framerates

                    Король Коль Исторические данные | Стихотворение | Примечания | Навигацияверсии1 правкаверсии1 правкаA New interpretation of the 'Artognou' stone, TintagelTintagel IslandАрхивировано