Calculus Applied to Probability and Statistics
by
Stefan Waner and Steven R. Costenoble

This Section: 3. Mean, Median, Variance and Standard Deviation

2. Probability Density Functions: Uniform, Exponential, Normal, and Beta Section 3 Exercises 4. You're the Expert Creating a Family Trust Calculus and Probability Main Page "Real World" Page

3. Mean, Median, Variance and Standard Deviation

Mean

In the last section we saw that if saving and loan institutions are continuously failing at a rate of $5%$ per year, then the associated probability density function is

with domain $[0,  +∞).$ An interesting and important question to ask is: What is the average length of time such an institution will last before failing? To answer this question, we use the following.

Mean or Expected Value

If $X$ is a continuous random variable with probability density function $f$ defined on an interval with (possibly infinite) endpoints a and b, then the mean or expected value of $X$ is

    $E(X) = ∫$ab $x f(x) dx.$

$E(X)$ is also called the average value of $X.$ It is what we expect to get if we take the average of many values of $X$ obtained in experiments.


Example 1


Example 2 Failing S&Ls


Question

Why is $E(X)$ given by that integral formula?

Answer

Suppose for simplicity that the domain of $f$ is a finite interval $[a,  b].$ Break up the interval into n subintervals $[x^{k-1},  x^{k}],$ each of length $Δx,$ as we did for Riemann sums. Now, the probability of seeing a value of $X$ in $[x^{k-1},  x^k]$ is approximately $f(x^k) Δx$ (the approximate area under the graph of $f$ over $[x^{k-1},  x^k]$). Think of this as the fraction of times we expect to see values of $X$ in this range. These values, all close to $x^k,$ then contribute approximately $x^k f(x^k) Δx$ to the average, if we average together many observations of $X.$ Adding together all of these contributions, we get

Now these approximations get better as $n→∞,$ and we notice that the sum above is a Riemann sum converging to

which is the formula we have been using.

Question

What are the expected values of the standard distributions we discussed in the previous section?

Answer

Let's compute them one by one.

Mean of a Uniform Distribution

If $X$ is uniformly distributed on $[a,  b],$ then

    $E(X) = (a + b)/2.$

This is not surprising, if you think about it for a minute. We'll leave the actual computation as one of the exercises.

Mean of an Exponential Distribution

If $X$ has the exponential distribution function $f(x) = ae^{-ax},$ then

    $E(X) = 1/a.$

We saw how to compute this in Example 2.

Mean of an Normal Distribution

If $X$ is normally distributed with parameters $µ$ and $σ,$ then

    $E(X) = µ.$

This is why we called $µ$ the mean, but we ought to do the calculation. Click on the footnote marker to to see it.

Mean of a Beta Distribution

If $X$ has the beta distribution function $f(x) = (β+1)(β+2)x^β(1-x),$ then

    $E(X) = (β + 1)/(β + 3)$

Again, we shall leave this as an exercise.


Example 3 Downsizing in the Utilities Industry


There is a generalization of the mean that we shall use below. If $X$ is a random variable on the interval $(a,  b)$ and probability density function $f,$ and if $g$ is any function defined on that interval, then we can define the expected value of $g$ to be

Thus, in particular, the mean is just the expected value of the function $g(x) = x.$ We can interpret this as the average we expect if we compute $g(X)$ for many experimental values of $X.$

Variance and Standard Deviation

Statisticians use the variance and standard deviation of a continuous random variable $X$ as a way of measuring its dispersion, or the degree to which is it "scattered." The definitions are as follows.

Variance and Standard Deviation

Let $X$ be a continuous random variable with density function $f$ defined on the interval $(a,  b),$ and let $µ = E(X)$ be the mean of $X.$ Then the variance of $X$ is given by

    $Var(X) = E((X-µ)^2) =$ ∫ab$(x-µ)^2 f(x) dx.$

The standard deviation of $X$ is the square root of the variance,

    $σ(X) = (Var(X))^{0.5}$

Notes

(1) In order to calculate the variance and standard deviation, we need first to calculate the mean.

(2) $Var(X)$ is the expected value of the function $(x-µ)^2,$ which measures the square of the distance of $X$ from its mean. It is for this reason that $Var(X)$ is sometimes called the mean square deviation, and $σ(X)$ is called the root mean square deviation. $Var(X)$ will be larger if $X$ tends to wander far away from its mean, and smaller if the values of $X$ tend to cluster near its mean.

(3) The reason we take the square root in the definition of $σ(X)$ is that $Var(X)$ is the expected value of the square of the deviation from the mean, and thus is measured in square units. Its square root $σ(X)$ therefore gives us a measure in ordinary units.

Question

What are the variances and standard deviations of the standard distributions we discussed in the previous section?

Answer

Let's compute them one by one. We'll leave the actual computations (or special cases) for the exercises.

Variance and Standard Deviation of a Uniform Distribution

If $X$ is uniformly distributed on $[a, b],$ then

    $Var(X) = (b-a)^2/12$

and

    $σ(X) =(b-a) /12^{0.5}$
Variance and Standard Deviation of an Exponential Distribution

If $X$ has the exponential distribution function $f(x) = ae^{-ax},$ then

    $Var(X) = 1/a^2$

and

    $σ(X) = 1/a.$
Variance and Standard Deviation of a Normal Distribution

If $X$ is normally distributed with parameters $µ$ and $σ,$ then

    $Var(X) = σ^2$

and

    $σ(X) = σ.$

(This is what you might have expected!)

Variance and Standard Deviation of a Beta Distribution

If $X$ has the beta distribution function $f(x) = (β+1)(β+2)x^β(1-x),$ then

You can see the significance of the standard deviation quite clearly in the normal distribution. As we mentioned in the previous section, $σ$ is the distance from the maximum at $µ$ to the points of inflection at $µ-σ$ and $µ +σ.$ The larger $σ$ is, the wider the bell. The following shows three normal distributions with three different standard deviations (all with $µ = 0.5$).

Again, a small standard deviation means that the values of $X$ will be close to the mean with high probability, while a large standard deviation means that the values may wander far away with high probability.

Median

The median income in the U.S. is the income $M$ such that half the population earn incomes $≤ M$ (so the other half earn incomes $≥ M$). In terms of probability, we can think of income as a random variable $X.$ Then the probability that $X  ≤  M$ is $1/2,$ and the probability that $X  ≥  M$ is also $1/2.$

Median

Let $X$ be a continuous random variable. The median of $X$ is the number $M$ such that
    $P(X  ≤  M) = 1/2.$

Then $P(M  ≤  X) = 1/2$ also.

If $f$ is the probability density function for $X$ and $f$ is defined on $(a,  b),$ then we can calculate $M$ by solving the equation

for $M.$ Graphically, the vertical line $x = M$ divides the total area under the graph of $f$ into two equal parts. (See the figure).

Question

What is the difference between the median and the mean?

Answer

Roughly speaking, the median divides the area under the distribution curve into two equal parts, while the mean is the value of $X$ at which the graph would balance. If a probability curve has as much area to the left of the mean as to the right, then the mean is equal to the median. This is true of uniform and normal distributions, which are symmetric about their means. On the other hand, the medians and means are different for the exponential distributions and most of the beta distributions, because their areas are not distributed symmetrically.


Example 4 Lines at the Post Office


Sometimes we cannot solve the equation ∫0M$f(x) dx = 1/2$  for $M$ analytically, as the next example shows.


Example 5


2. Probability Density Functions: Uniform, Exponential, Normal, and Beta Section 3 Exercises 4. You're the Expert Creating a Family Trust Calculus and Probability Main Page "Real World" Page

We would welcome comments and suggestions for improving this resource.

Mail us at:
Stefan Waner (matszw@hofstra.edu) Steven R. Costenoble (matsrc@hofstra.edu)

Last Updated: September, 1996
Copyright © 1996 StefanWaner and Steven R. Costenoble