Angenommen, eine Zufallsvariable hat eine untere und eine obere Schranke [0,1]. Wie berechnet man die Varianz einer solchen Variablen?
22
Angenommen, eine Zufallsvariable hat eine untere und eine obere Schranke [0,1]. Wie berechnet man die Varianz einer solchen Variablen?
Antworten:
Sie können Popovicius Ungleichung wie folgt beweisen. Verwenden die Notationm=infX und M=supX . Definiere eine Funktion g durch
g(t)=E[(X−t)2].
Berechnen der Ableitungg′ und Lösen von
g′(t)=−2E[X]+2t=0,
wir, dassg sein Minimum beit=E[X] (beachte, dassg′′>0 ).
Betrachten Sie nun den Wert der Funktiong am Sonderpunkt t=M+m2 . Es muss der Fall seindass
Var[X]=g(E[X])≤g(M+m2).
Aber
g(M+m2)=E[(X−M+m2)2]=14E[((X−m)+(X−M))2].
DaX−m≥0 undX−M≤0 , haben wir
((X−m)+(X−M))2≤((X−m)−(X−M))2=(M−m)2,
was bedeutet, dass
14E[((X−m)+(X−M))2]≤14E[((X−m)−(X−M))2]=(M−m)24.
Daher haben wir Popovicius Ungleichung V a r [ X ] ≤ ( M - m ) 2 bewiesen
Var[X]≤(M−m)24.
quelle
LetF be a distribution on [0,1] . We will show that if the variance of F is maximal, then F can have no support in the interior, from which it follows that F is Bernoulli and the rest is trivial.
As a matter of notation, letμk=∫10xkdF(x) be the k th raw moment of F (and, as usual, we write μ=μ1 and σ2=μ2−μ2 for the variance).
We knowF does not have all its support at one point (the variance is minimal in that case). Among other things, this implies μ lies strictly between 0 and 1 . In order to argue by contradiction, suppose there is some measurable subset I in the interior (0,1) for which F(I)>0 . Without any loss of generality we may assume (by changing X to 1−X if need be) that F(J=I∩(0,μ])>0 : in other words, J is obtained by cutting off any part of I above the mean and J has positive probability.
Let us alterF to F′ by taking all the probability out of J and placing it at 0 . In so doing, μk changes to
As a matter of notation, let us write[g(x)]=∫Jg(x)dF(x) for such integrals, whence
Calculate
The second term on the right,(μ[x]−[x]2) , is non-negative because μ≥x everywhere on J . The first term on the right can be rewritten
The first term on the right is strictly positive because (a)μ>0 and (b) [1]=F(J)<1 because we assumed F is not concentrated at a point. The second term is non-negative because it can be rewritten as [(μ−x)(x)] and this integrand is nonnegative from the assumptions μ≥x on J and 0≤x≤1 . It follows that σ′2−σ2>0 .
We have just shown that under our assumptions, changingF to F′ strictly increases its variance. The only way this cannot happen, then, is when all the probability of F′ is concentrated at the endpoints 0 and 1 , with (say) values 1−p and p , respectively. Its variance is easily calculated to equal p(1−p) which is maximal when p=1/2 and equals 1/4 there.
Now whenF is a distribution on [a,b] , we recenter and rescale it to a distribution on [0,1] . The recentering does not change the variance whereas the rescaling divides it by (b−a)2 . Thus an F with maximal variance on [a,b] corresponds to the distribution with maximal variance on [0,1] : it therefore is a Bernoulli(1/2) distribution rescaled and translated to [a,b] having variance (b−a)2/4 , QED.
quelle
If the random variable is restricted to[a,b] and we know the mean μ=E[X] , the variance is bounded by (b−μ)(μ−a) .
Let us first consider the casea=0,b=1 . Note that for all x∈[0,1] , x2≤x , wherefore also E[X2]≤E[X] . Using this result,
To generalize to intervals[a,b] with b>a , consider Y restricted to [a,b] . Define X=Y−ab−a , which is restricted in [0,1] . Equivalently, Y=(b−a)X+a , and thus
quelle
At @user603's request....
A useful upper bound on the varianceσ2 of a random variable that takes on values in [a,b] with probability 1 is σ2≤(b−a)24 . A proof for the
special case a=0,b=1 (which is what the OP asked about) can be found
here on math.SE, and
it is easily adapted to
the more general case. As noted in my comment above and also in the answer referenced
herein, a discrete random variable that takes on values a and b with equal
probability 12 has variance (b−a)24 and thus no tighter
general bound can be found.
Another point to keep in mind is that a bounded random variable has finite variance, whereas for an unbounded random variable, the variance might not be finite, and in some cases might not even be definable. For example, the mean cannot be defined for Cauchy random variables, and so one cannot define the variance (as the expectation of the squared deviation from the mean).
quelle
are you sure that this is true in general - for continuous as well as discrete distributions? Can you provide a link to the other pages? For a general distibution on[a,b] it is trivial to show that
On the other hand one can find it with the factor1/4 under the name Popoviciu's_inequality on wikipedia.
This article looks better than the wikipedia article ...
For a uniform distribution it holds that
quelle