Introduction to Probability Theory

Probability Theory

Probability theory is a branch of mathematics that allows us to reason about events that are inherently random. Another way to think about probability is in terms of repeatable statistical experiments. A statistical experiment is an action or occurrence that can have multiple different outcomes, all of which can be specified in but where the particular outcome that will occur cannot be specified in advance because it depends on random chance.

In general, the probability theory is applied to two broad categories of applications:-

  1. Generation/Prediction
    1. Reasoning from causes to effects : Given a known set of causes and knowledge about how they interact, what are likely/unlikely outcomes
    2. For example: if you roll a fair 6-sided die, what is the probability of getting 1.
  1. Inference.
    1. Reasoning from effects to causes: Given knowledge about possible causes and how they interact, as well as some observed outcomes, which causes are likely/unlikely
    2. For example: observe patient's symptoms ⇒ determine disease

Basic Concepts in Probability

Sample Spaces : The SAMPLE SPACE of a statistical experiment is the set of all possible outcomes
For example:

  1. Random coin toss: S = {H, T}
  2. Roll a die: S = {1, 2, 3, 4, 5, 6}
  3. Choose a random number between 0 and 1: S = [0, 1]
Event : An EVENT is a subset of the sample space.An event is a collection of possible outcomes of a random experiment. There are two special events: the whole sample space S is called the certain or the sure event. The empty set ∅ is the null event. By definition, an impossible event has probability zero, and a certain event has probability one.

Probability of an event: every event E is assigned a probability P(E). “P(E) is the probability that event E is true”
Probability of an event
The probability of an event E, which we write as P(E), is simply the number of outcomes in E divided by number of outcomes in S:
Probability of an event
For example: If you flip a fair coin once, what is the probability of getting a tail?
Here S = {H,T} thus |S|=2, which is the total possible outcomes.
E => Event of getting tails, thus |E| = 1
Thus, P(E) = ½

Conditional probability: P(B|A) = conditional probability that B is true given that A is true
Conditional probability

Random Variables and Joint Distribution

Random Variable: A random variable is a 'random number', meaning a number which is determined by the outcome of a random experiment.
Usually denoted X, Y, . . . . The range of X is the set of possible values for X.
Mathematically, X is a real-valued map on the sample space S:
Random Variable
If the range of X is finite then we say that X is a simple random variable.

Joint Distribution: Given a collection of discrete random variables X = (X1, X2, . . . , Xn), let Ri be the range of Xi . Then the range of X is the Cartesian product R1 × · · · × Rn. Their joint pmf is the collection of probabilities P(X1 = x1, . . . , Xn = xn) for every point (x1, . . . , xn) in the range of X1, X2, . . . , Xn.
Suppose that we have two random variables X and Y.If we want to know about the values that X and Y assume simultaneously during outcomes of a random experiment, we require a more complicated structure known as the joint cumulative distribution function of X and Y , defined by FXY (x, y) = P(X ≤ x, Y ≤ y)
Here, we call FX(x) and FY (y) the marginal cumulative distribution functions of FXY (x, y).

For example:
We have a box of ten balls in which four are white, three are black, and three are red. One has to select five balls out of the box without replacement and count the number of white and red balls in the sample. What is the probability one observes two white and two red balls in the sample?
Here, the total number of outcomes is 10C5=252.
Since the box has four white balls, the number of ways of choosing two white is 4C2. Of the three red balls, one wants to choose two – the number of ways of doing that is 3C2. Last, the number of ways of choosing the remaining one black ball is 3C1. So the total number of ways of choosing two white, two red, and one black ball is the product,
P(X=2, Y=2)=4C2x3C2x3C1/10C5=54/252,
Where X=number of red balls selected, Y=number of white balls selected.
Joint Distribution

Independence and Conditional Independence

Independence: Two events A and B are INDEPENDENT EVENTS if Two events A, B are independent if P(A|B) = P(A). P(A|B) = P(A) ⇐⇒ P(B|A) = P(B) ⇐⇒ P(A ∩ B) = P(A) P(B)

For Example: A box contains two coins: a regular coin and one fake two-headed coin (P(H)=1). I choose a coin at random and toss it twice. Define the following events.
A= First coin toss results in an H.
B= Second coin toss results in an H.
C= Coin 1 (regular) has been selected.

Find P(A|C),P(B|C),P(A∩B|C),P(A),P(B), and P(A∩B). Note that A and B are NOT independent, but they are conditionally independent given C.
We have P(A|C)=P(B|C)=12. Also, given that Coin 1 is selected, we have P(A∩B|C)=12.12=14. To find P(A),P(B), and P(A∩B), we use the law of total probability:
P(A)=P(A|C)P(C)+P(A|C^c)P(C^c)
=12⋅12+1⋅12
=34
Similarly, P(B)=34. For P(A∩B), we have
P(A∩B)=P(A∩B|C)P(C)+P(A∩B|Cc)P(Cc)=P(A|C)P(B|C)P(C)+P(A|Cc)P(B|Cc)P(Cc) (by conditional independence of A and B)
=12⋅12⋅12