The maximal variance.
by 3ch03s, Aug 11, 2015, 9:23 PM
In this year, I found something interesting. this post in AOPS
What about a random variable with support in [0,1]?
well, we can think, there is so many distributions, some well known (like the beta family, bernoulli trials, and a large etc), some very strange (like the rm distribution in my first entry blog). all with their mean and variance, could we think a distribution given a arbitrary mean and variance? the answer is no. in this entry, we find a several interesting things for the mean and variance of any distribution in [0,1].
First, ¿Why not all the pair mean-variance are possible in this kind of distributions?.
The answer is pretty simple. for X random variable distributed in [0,1] we can check that all the moments exists, and his mgf also exists always, in particular for first and second moment, who determines the mean and variance.
as for
we have
for p<q naturals, is easy to show that
for this kind of distributions, specially setting p=1 and q=2:
. We know that the variance is defined by:
, so how to use the preceding fact?. just note this simply operation:
![$E[X^2] \leq E[X]$](//latex.artofproblemsolving.com/4/5/8/4587a6a6c050de907a69a15fdce7b439802e32e5.png)
![$E[X^2]-E[X]^2 \leq E[X]-E[X]^2$](//latex.artofproblemsolving.com/b/3/c/b3cbff3b414b6646f9fd2fa31a5d4c0172068005.png)
then![$V[X] \leq E[X](1-E[X])$](//latex.artofproblemsolving.com/b/4/f/b4f680b685f9d0da73fb85b1f2e483bd5fb9a196.png)
This simple fact shows for a given mean
in [0,1] and a distribution
in [0,1] the variance is bounded by the term
(in fact is a parabola) and most important: as
(a well-known inequality) we conclude that the variance can't be greater than 1/4. this is his maximal variance, so thinking a distribution with an arbitrary pair of mean-variance is not posible unless this pair lies on the "factible set" given by:
![$\Phi=\{(\mu,\sigma^2) \in [0,1]^2 \sigma^2 \leq \mu(1-\mu)\}$](//latex.artofproblemsolving.com/4/d/7/4d79617ba1ed29fdc38f41b46b59668ee28f17b2.png)
more interestingly, if we think a pair mean-standard.deviation, the factible set is a semicircle of radius 1/2 and center in x=1/2.
The main questions are: ¿what's happening about the distribution for a given pair in the factible set? can the proximity of the pair respect the boundary of this set tell us how is the distribution behavior?, what distribution reaches is maximal variance? is unique?, what about the moments? how does affect for a more general distributions with bounded support? what's happening if the variance is very near to the maximal line?.
We know that the distribution who reaches is maximal variance is the bernoulli distribution, this distribution is often regarded a "discrete" distribution, but by the pointwiew of measuring theory, can be also regarded a concentrated distribution at the points 0 and 1 in some proportion given by the mean
in [0,1], in fact the relation are
for the point 1 and
for the point 0 (that is, a simple trial giving 1 with prob
). the uniqueness of the "maximal variance" distribution can be proved using chebyshev inequalities. setting
for fixed h<1/2 we have using some algebra that
so putting the function
and the chebyshev inequalities we have:
![$P[|x-1/2|\leq h]\leq \frac{E[X(1-X)]}{\frac 14-h^2}$](//latex.artofproblemsolving.com/f/7/b/f7ba465021694ec40dd099669b8567edc4c967be.png)
But![$E[X(1-X)]=\mu(1-\mu)-\sigma^2$](//latex.artofproblemsolving.com/0/0/d/00d0421e113a26ce295d167dada570bfd7572d7a.png)
![$P[|x-1/2|\geq h]\leq \frac{\mu(1-\mu)-\sigma^2}{\frac 14-h^2}$](//latex.artofproblemsolving.com/7/6/7/767e1c54b404188e68685cc3e3dbc9f4d21dfd0a.png)
the preceding inequality tell us for a bernoulli distribution, for all h<1/2 the probability is
(in fact is an equivalency), so, if there is another distribution different of bernoulli there exists "h" such that
, if also, this distribution reaches his maximal variance so we have for the same h:
which is a contradiction.
Again this inequality tell us something more about how the distribution is because the probability of X taking values around 1/2, or even, every set A (medible) in ]0,1[ it's bounded by a quantity who depends of the diference of the maximal variance and the variance and the maximal distance respect of x=1/2, then, if the variances comes closer to his maximal, the probability begins to be more smaller, and evidently, more concentrated in the extremes, so, the more the variance approaches is maximum, more concentrated in 0 and 1 is the distribution.
Then, converges into a bernoulli.
It will continue.....
What about a random variable with support in [0,1]?
well, we can think, there is so many distributions, some well known (like the beta family, bernoulli trials, and a large etc), some very strange (like the rm distribution in my first entry blog). all with their mean and variance, could we think a distribution given a arbitrary mean and variance? the answer is no. in this entry, we find a several interesting things for the mean and variance of any distribution in [0,1].
First, ¿Why not all the pair mean-variance are possible in this kind of distributions?.
The answer is pretty simple. for X random variable distributed in [0,1] we can check that all the moments exists, and his mgf also exists always, in particular for first and second moment, who determines the mean and variance.
as for


![$E[X^q] \leq E[X^p]$](http://latex.artofproblemsolving.com/f/8/7/f875066350012cf9444b978bb867d0629ce71fdb.png)
![$E[X^2] \leq E[X]$](http://latex.artofproblemsolving.com/4/5/8/4587a6a6c050de907a69a15fdce7b439802e32e5.png)
![$V[X]=E[X^2]-E[X]^2$](http://latex.artofproblemsolving.com/f/0/1/f0143b84ba14579b6c041366777737c496e8a6b7.png)
![$E[X^2] \leq E[X]$](http://latex.artofproblemsolving.com/4/5/8/4587a6a6c050de907a69a15fdce7b439802e32e5.png)
![$E[X^2]-E[X]^2 \leq E[X]-E[X]^2$](http://latex.artofproblemsolving.com/b/3/c/b3cbff3b414b6646f9fd2fa31a5d4c0172068005.png)
then
![$V[X] \leq E[X](1-E[X])$](http://latex.artofproblemsolving.com/b/4/f/b4f680b685f9d0da73fb85b1f2e483bd5fb9a196.png)
This simple fact shows for a given mean




![$\Phi=\{(\mu,\sigma^2) \in [0,1]^2 \sigma^2 \leq \mu(1-\mu)\}$](http://latex.artofproblemsolving.com/4/d/7/4d79617ba1ed29fdc38f41b46b59668ee28f17b2.png)
more interestingly, if we think a pair mean-standard.deviation, the factible set is a semicircle of radius 1/2 and center in x=1/2.
The main questions are: ¿what's happening about the distribution for a given pair in the factible set? can the proximity of the pair respect the boundary of this set tell us how is the distribution behavior?, what distribution reaches is maximal variance? is unique?, what about the moments? how does affect for a more general distributions with bounded support? what's happening if the variance is very near to the maximal line?.
We know that the distribution who reaches is maximal variance is the bernoulli distribution, this distribution is often regarded a "discrete" distribution, but by the pointwiew of measuring theory, can be also regarded a concentrated distribution at the points 0 and 1 in some proportion given by the mean







![$P[|x-1/2|\leq h]\leq \frac{E[X(1-X)]}{\frac 14-h^2}$](http://latex.artofproblemsolving.com/f/7/b/f7ba465021694ec40dd099669b8567edc4c967be.png)
But
![$E[X(1-X)]=\mu(1-\mu)-\sigma^2$](http://latex.artofproblemsolving.com/0/0/d/00d0421e113a26ce295d167dada570bfd7572d7a.png)
![$P[|x-1/2|\geq h]\leq \frac{\mu(1-\mu)-\sigma^2}{\frac 14-h^2}$](http://latex.artofproblemsolving.com/7/6/7/767e1c54b404188e68685cc3e3dbc9f4d21dfd0a.png)
the preceding inequality tell us for a bernoulli distribution, for all h<1/2 the probability is
![$P[|x-1/2|\geq h]=0$](http://latex.artofproblemsolving.com/b/4/b/b4b0f55950f12b8be7bba821d4142bfae69a6d0b.png)
![$P[|x-1/2|\geq h]>0$](http://latex.artofproblemsolving.com/c/1/2/c123cb7e1a551affb9368aa56077915bb604cf3a.png)
![$P[|x-1/2|\geq h]=0$](http://latex.artofproblemsolving.com/b/4/b/b4b0f55950f12b8be7bba821d4142bfae69a6d0b.png)
Again this inequality tell us something more about how the distribution is because the probability of X taking values around 1/2, or even, every set A (medible) in ]0,1[ it's bounded by a quantity who depends of the diference of the maximal variance and the variance and the maximal distance respect of x=1/2, then, if the variances comes closer to his maximal, the probability begins to be more smaller, and evidently, more concentrated in the extremes, so, the more the variance approaches is maximum, more concentrated in 0 and 1 is the distribution.
Then, converges into a bernoulli.
It will continue.....