PHY114 - Lecture - Notes - Lecture 02
PHY114 - Lecture - Notes - Lecture 02
Lecture 02
Summer 2024
Sayeda Tashnuba Jahan
Probability Distributions
Discrete Probability Distribution
Discrete random variables can only take a finite number of values. Examples include
the number obtained when throwing a die (1, 2, 3, 4, 5, or 6 ), the number of children
in each family (0, 1, 2, . . .), and the number of people killed per year in the UK in
bizarre gardening accidents (0, 1, 2, . . .). Let x be a discrete random variable which
takes values xi with probability Pi . We require that the sum of the probabilities of
every possible outcome adds up to one. This may be written
X
Pi = 1.
i
The idea is that you weight by its probability each value taken by the random vari-
able x.
Now let us actually evaluate the mean of x for a particular discrete distribution.
1 1 1
Let x take values 0,1, and 2 with probabilities , ,
2 4
and 4
respectively. Calcu-
late ⟨x⟩ and ⟨x2 ⟩.
Solution:
1 1 1
P
First check that Pi = 1. Since 2
+ 4
+ 4
= 1, this is fine. Now we can cal-
culate the averages as follows:
X
⟨x⟩ = x i Pi
i
1 1 1
=0· +1· +2·
2 4 4
3
= .
4
Again, we find that the mean ⟨x⟩ is not actually one of the possible values of x. We
can now calculate the value of ⟨x2 ⟩ as follows:
X
x2 = x2i Pi
i
1 1 1
=0· +1· +4·
2 4 4
5
=
4
Linear Transformations
Sometimes one has a random variable, and one wants to make a second random
variable by performing a linear transformation on the first one. If y is a random
variable, which is related to the random variable x by the equation
y = ax + b,
σx2 = (x − ⟨x⟩)2 .
We will call σx the standard deviation, and it is defined as the square root of the
variance: p
σx = ⟨(x − ⟨x⟩)2 ⟩.
The standard deviation represents the "root mean square" (known as the "rms")
scatter or spread in the data. The following identity is extremely useful:
σx2 = (x − ⟨x⟩)2
= x2 − 2x⟨x⟩ + ⟨x⟩2
= x2 − 2⟨x⟩⟨x⟩ + ⟨x⟩2
= x2 − ⟨x⟩2 .
Independent Variables
If u and v are independent random variables the probability that u is in the range
from u to u + du and v is in the range from v to v + dv is given by the product
Pu (u)duPv (v)dv.
Hence, the average value of the product of u and v is
ZZ
⟨uv⟩ = uvPu (u)Pv (v)du dv
Z Z
= uPu (u)du vPv (v)dv
= ⟨u⟩⟨v⟩,
because the integrals separate for independent random variables. This implies that
the average value of the product of u and v is equal to the product of their average
values.
Suppose that there are n independent random variables, Xi , each with the same
2
mean ⟨X⟩ and variance σX . Let Y be the sum of the random variables, so that
Y = X1 + X2 + · · · + Xn . Find the mean and variance of Y .
Solution:
but since all the Xi have the same mean ⟨X⟩ this can be written
⟨Y ⟩ = n⟨X⟩.
Hence the mean of Y is n times the mean of the Xi . To find the variance of Y , we
can use the formula
σY2 = Y 2 − ⟨Y ⟩2 .
Hence
Y 2 = X12 + · · · + XN2 + X1 X2 + X2 X1 + X1 X3 + · · ·
= X12 + · · · + XN2 + ⟨X1 X2 ⟩ + ⟨X2 X1 ⟩ + ⟨X1 X3 ⟩ + · · ·
There are n terms like ⟨X12 ⟩ on the right-hand side, and n(n − 1) terms like ⟨X1 X2 ⟩.
The former terms take the value ⟨X 2 ⟩ and the latter terms (because they are the
product of two independent random variables) take the value ⟨X⟩⟨X⟩ = ⟨X⟩2 .
Hence, using eqn 3.35,
Y 2 = n X 2 + n(n − 1)⟨X⟩2 ,
so that
σY2 = Y 2 − ⟨Y ⟩2
= n X 2 − n⟨X⟩2
2
= nσX .
An application is in the theory of random walks. Imagine a drunken person stagger-
ing out of a pub and attempting to walk along a narrow street (which confines him
or her to motion in one dimension). Let’s pretend that with each inebriated step, the
drunken person is equally likely to travel one step forwards or one step backwards.
The effects of intoxication are such that each step is uncorrelated with the previous
one. Thus the average distance travelled in a single step is ⟨X⟩ = 0. P After n such
steps, we would have an expected total distance travelled of ⟨Y ⟩ = ⟨Xi ⟩ = 0.
However, in this case the root mean squared distance is more revealing.√ In this case
⟨Y 2 ⟩ = n ⟨X 2 ⟩, so that the rms length of a random walk of n steps is n times the
length of a single step. This result will be useful in considering Brownian motion.
Choose the x axis to lie along the street so that x = 0 is the position of the point
of origin, the lamppost. Clearly, since each step is of length l, the location of the
man along the x axis must be of the form x = ml, where m is an integer (positive,
negative, or zero). The question of interest is then the following: After the man has
taken N steps, what is the probability of his being located at the position x = ml ?
This statistical formulation of the problem again implies that one considers a very
large number K of similar men starting from similar lampposts. (Alternatively, if
the situation does not change in time, e.g., if the man does not gradually sober up,
one could also repeat the same experiment ℜ times with the same man.) At each
step, one finds that a fraction p of the men moves to the right. One then asks what
fraction of the men will be located at the position x = ml after N steps.
One can readily generalize this one-dimensional problem to more dimensions whether
it, be two (a drunk starting out from a lamppost in the middle of a parking lot),
three, or more. One again asks for the probability that after N steps the man is
located at a given distance from the origin (although the distance is no longer of the
form ml, with m integral).
The random walk problem illustrates some very fundamental results of proba-
bility theory. The techniques used in the study of this problem are powerful and
basic, and recur again and again throughout statistical physics. It is therefore very
instructive to gain a good understanding of this problem.