A Weird but Interesting Distribution pt II

by 3ch03s, Mar 27, 2016, 12:47 AM

In the first part I define an interesting distribution. Let $X_n,n\in \mathbb{N}$ a bernoulli iid sequence with parameter $p \in ]0,1[$, and let
$R=\sum_{n \in \mathbb{N}} \frac{X_n}{2^n}$

We called it $X\sim rm(p)$.
We also showed some interesting properties of his probability distribution (cumulated) function, and we give some clues about the behavior of the distribution when p=1/2.

I claim when p=1/2, $R$ has the uniform distribution (which is in fact, the lebesgue measure in [0,1]). Also we give this property of cdf:

$F(r)=(1-p)F(2r)+pF(2r-1)$

Then we can compute any value of this distribution using dyadic aproximations of any number $r$ in [0,1]. Is pretty clear the sequence of aproximations has the form $x_n=\sum_{i=1}^n a_i/2^i$. this implies that for any number of the form $r=k/2^n$, with $k<2^n$ natural, F(r) has a "closed" form, because we are dealing with a finite number of bernoulli variables in that case.

we know every number $x$ in [0,1] has a unique (in the sense of unique sequence of 0's and 1's without getting finite 0's) binary representation, with this, we can compute $F(r)$ if $r=k/2^n$, which using the recurssion formula leads to:

$F(k/2^n)=\sum_{\sum x_i/2^i \leq k/2^n} p^{\sum x_i}(1-p)^{n-\sum x_k}$

Another aproach to the formula is thinking in the secuence of 0 and 1 for the finite dyadic representation.
So for every aproximation $r_n$ we can compute using this recurssion:

$F(x_n)=F(x_{n-1})+a_i(p^{\sum x_i})(1-p)^{n-\sum x_i}$

¿What happens when $p=1/2$?

the formula for dyadic numbers just reduces to:

$F(k/2^n)=\sum_{\sum x_i/2^i \leq k/2^n} 1/2^n=k/2^n$, hence, any dyadic number $r$ has the propierty $F(r)=r$, then, the recursion for any dyadic approximation gives $F(x_n)=x_n$, So in the limit, by density arguments that i will justificate soon, we have $F(x)=x$ for any $x  \in [0,1]$, so, for p=1/2, the distribution is in fact, the uniform distribution and hence has a pdf.

But... what about the image of X? are [0,1]?... in fact, because the representation, the uniform distribution does not go trough the whole interval, just, a select set of numbers in [0,1] with the property: $x \in [0,1], x=\sum x_n/2^n|\frac 1n \sum x_i \to 1/2$.

So we define:

$A_p=\{ x \in {0,1}, x=\sum x_n/2_n, x_n \in \{0,1\}, \textrm{such that } \frac 1n\sum x_n \to p\}$

and hence $[0,1]=\bigcup_{p \in [0,1]} A_p$.

if $X \sim rm(p)$ then the image of X is $A_p$.
its clear that $A_p$ and $A_q$ are disjoint if $p \neq q$. We will use this property using the fact if p=1/2, $X \sim  U[0,1]$, and hence the probability measure is in fact, the Lebesgue measure in [0,1], so, exists a pdf by radon-nykodym theorem. Let $\lambda$ the Lebesgue measure in [0,1], write:
$[0,1]=A_{1/2} \cup \bigcup_{p \neq 1/2} A_p$
Clearly $A_{1/2}$ and the rest of the union are disjoint.
so:
$\lambda([0,1])=1=\lambda (A_{1/2})+\lambda (\bigcup_{p \neq 1/2} A_p)$

The maximal variance.

by 3ch03s, Aug 11, 2015, 9:23 PM

In this year, I found something interesting. this post in AOPS

What about a random variable with support in [0,1]?

well, we can think, there is so many distributions, some well known (like the beta family, bernoulli trials, and a large etc), some very strange (like the rm distribution in my first entry blog). all with their mean and variance, could we think a distribution given a arbitrary mean and variance? the answer is no. in this entry, we find a several interesting things for the mean and variance of any distribution in [0,1].

First, ¿Why not all the pair mean-variance are possible in this kind of distributions?.

The answer is pretty simple. for X random variable distributed in [0,1] we can check that all the moments exists, and his mgf also exists always, in particular for first and second moment, who determines the mean and variance.

as for $0 \leq x \leq 1$ we have $x^q \leq x^p$ for p<q naturals, is easy to show that $E[X^q] \leq E[X^p]$ for this kind of distributions, specially setting p=1 and q=2: $E[X^2] \leq E[X]$. We know that the variance is defined by: $V[X]=E[X^2]-E[X]^2$, so how to use the preceding fact?. just note this simply operation:
$E[X^2] \leq E[X]$
$E[X^2]-E[X]^2 \leq E[X]-E[X]^2$
then $V[X] \leq E[X](1-E[X])$

This simple fact shows for a given mean $\mu$ in [0,1] and a distribution $X$ in [0,1] the variance is bounded by the term $\mu(1-\mu)$ (in fact is a parabola) and most important: as $\mu(1-\mu) \leq 1/4$ (a well-known inequality) we conclude that the variance can't be greater than 1/4. this is his maximal variance, so thinking a distribution with an arbitrary pair of mean-variance is not posible unless this pair lies on the "factible set" given by:
$\Phi=\{(\mu,\sigma^2) \in [0,1]^2 \sigma^2 \leq \mu(1-\mu)\}$
more interestingly, if we think a pair mean-standard.deviation, the factible set is a semicircle of radius 1/2 and center in x=1/2.

The main questions are: ¿what's happening about the distribution for a given pair in the factible set? can the proximity of the pair respect the boundary of this set tell us how is the distribution behavior?, what distribution reaches is maximal variance? is unique?, what about the moments? how does affect for a more general distributions with bounded support? what's happening if the variance is very near to the maximal line?.

We know that the distribution who reaches is maximal variance is the bernoulli distribution, this distribution is often regarded a "discrete" distribution, but by the pointwiew of measuring theory, can be also regarded a concentrated distribution at the points 0 and 1 in some proportion given by the mean $\mu$ in [0,1], in fact the relation are $\mu$ for the point 1 and $1-\mu$ for the point 0 (that is, a simple trial giving 1 with prob $\mu$). the uniqueness of the "maximal variance" distribution can be proved using chebyshev inequalities. setting $|x-1/2|\leq h$ for fixed h<1/2 we have using some algebra that $x(1-x)\geq \frac 14-h^2$ so putting the function $g(x)=x(1-x)$ and the chebyshev inequalities we have:
$P[|x-1/2|\leq h]\leq \frac{E[X(1-X)]}{\frac 14-h^2}$
But $E[X(1-X)]=\mu(1-\mu)-\sigma^2$
$P[|x-1/2|\geq h]\leq \frac{\mu(1-\mu)-\sigma^2}{\frac 14-h^2}$
the preceding inequality tell us for a bernoulli distribution, for all h<1/2 the probability is $P[|x-1/2|\geq h]=0$ (in fact is an equivalency), so, if there is another distribution different of bernoulli there exists "h" such that $P[|x-1/2|\geq h]>0$, if also, this distribution reaches his maximal variance so we have for the same h: $P[|x-1/2|\geq h]=0$ which is a contradiction.
Again this inequality tell us something more about how the distribution is because the probability of X taking values around 1/2, or even, every set A (medible) in ]0,1[ it's bounded by a quantity who depends of the diference of the maximal variance and the variance and the maximal distance respect of x=1/2, then, if the variances comes closer to his maximal, the probability begins to be more smaller, and evidently, more concentrated in the extremes, so, the more the variance approaches is maximum, more concentrated in 0 and 1 is the distribution.
Then, converges into a bernoulli.

It will continue.....
Shouts
Submit
  • hi
    also dead blog

    by aidan0626, Oct 10, 2022, 9:03 PM

  • hai.....

    by fluff_E, Dec 17, 2021, 5:13 PM

  • omg so pro

    by mathboy282, Sep 16, 2021, 4:52 AM

  • how is he so good?

    by centslordm, Aug 8, 2021, 3:45 PM

  • hello comrade

    by tauros, Sep 3, 2020, 2:55 PM

  • HOLA AMIGOS!

    by dolphindesigner, Mar 12, 2020, 12:41 AM

  • puedes hablar espanol conmigo?

    by Eugenis, Aug 12, 2015, 3:04 AM

  • tengo shouts? yaaaa.....

    by 3ch03s, Aug 11, 2015, 7:38 PM

  • hola. Mi nombre es claudio, soy chileno (y la ctm) Ingeniero civil Matematico con menci??n en estadistica aceptado en el doctorado de Estadpistica de la puc. Mucho Floyd Buki TB-303 y Lacie Baskerville y disco house.

    by 3ch03s, Nov 18, 2014, 4:29 AM

9 shouts
Tags
About Owner
  • Posts: 1323
  • Joined: May 23, 2014
Blog Stats
  • Blog created: Jun 9, 2014
  • Total entries: 3
  • Total visits: 1247
  • Total comments: 0
Search Blog
a