Unit 5: Probability

Table of Contents

1. Introduction: Basic Concepts

Probability is the measure of the likelihood that an event will occur. It is quantified as a number between 0 and 1 (where 0 indicates impossibility and 1 indicates certainty).

Random Experiment

An experiment or process for which the outcome cannot be predicted with certainty, but all possible outcomes are known.

Sample Point

A single possible outcome of a random experiment.

Sample Space (S)

The set of all possible outcomes (sample points) of a random experiment.

Event (E)

A subset of the sample space. It is a collection of one or more sample points.


2. Algebra of Events and Types of Events

Events can be combined using set operations.

Algebra of Events (Set Operations)

Types of Events

1. Mutually Exclusive (or Disjoint) Events

Two events that cannot occur at the same time. They have no sample points in common.

2. Exhaustive Events

A set of events that covers the entire sample space. When the experiment is performed, at least one of these events must occur.

3. Equally Likely Events

Outcomes that have the same chance of occurring.


3. Classical (Statistical) Definition of Probability

This is the first and most intuitive definition of probability.

If a random experiment has 'n' mutually exclusive, exhaustive, and equally likely outcomes, and 'm' of these outcomes are favorable to an event A, then the probability of event A is:
P(A) = m / n
P(A) = (Number of outcomes favorable to A) / (Total number of possible outcomes)

Example:

Experiment: Drawing one card from a standard 52-card deck.

Event A: Drawing a King.

Limitations of Classical Definition:


4. Axiomatic Definition of Probability

This is the modern, mathematical definition of probability (developed by Kolmogorov). It does not say *how* to calculate probability, but states the *rules* (axioms) that any probability measure P must follow.

Given a sample space S, a probability P(A) is a number assigned to every event A that satisfies these three axioms:

Axiom 1 (Non-negativity): For any event A, P(A) ≥ 0.
(Probability can never be negative).

Axiom 2 (Certainty): The probability of the entire sample space is 1.
P(S) = 1.
(Something *must* happen).

Axiom 3 (Additivity): If A and B are two mutually exclusive events (A ∩ B = ∅), then the probability of their union is the sum of their individual probabilities.
P(A ∪ B) = P(A) + P(B)

From these 3 simple axioms, all other probability rules can be derived, such as:


5. Laws of Addition and Multiplication

1. Addition Law of Probability

This rule is used to find the probability of (A or B).

General Rule (for *any* two events):

P(A ∪ B) = P(A) + P(B) - P(A ∩ B)

(We subtract P(A ∩ B) because it was counted twice - once in P(A) and once in P(B)).

Special Rule (for *Mutually Exclusive* events):

If A and B are mutually exclusive, then P(A ∩ B) = 0. The formula simplifies to:

P(A ∪ B) = P(A) + P(B) (This is just Axiom 3)

2. Multiplication Law of Probability

This rule is used to find the probability of (A and B). It is derived from the definition of conditional probability.

P(A ∩ B) = P(A) * P(B | A)
(The probability of A, times the probability of B *given that* A has already happened)

or equivalently:
P(A ∩ B) = P(B) * P(A | B)

A special case for *Independent* events is covered below.


6. Conditional Probability

Conditional Probability, P(A | B): The probability of event A occurring, given that event B has already occurred.

It "updates" the probability of A based on new information (that B happened). We are restricting our sample space from S down to just B.

Formula:

P(A | B) = P(A ∩ B) / P(B) (provided P(B) > 0)

Example:

Experiment: Roll a fair die (S = {1, 2, 3, 4, 5, 6}).

Event A: Getting a "4". P(A) = 1/6.

Event B: Getting an "even number". B = {2, 4, 6}. P(B) = 3/6.

Question: What is the probability of getting a 4, *given that* we know the result was an even number? We want to find P(A | B).

This makes sense: if we know the outcome is even {2, 4, 6}, the chance of it being the "4" is 1 out of 3.


7. Independent Events

Independent Events: Two events A and B are independent if the occurrence of one event does not affect the probability of the other event occurring.

Formal Definition:

A and B are independent if and only if:

P(A ∩ B) = P(A) * P(B)

This also means:

Common Mistake: Do NOT confuse "Mutually Exclusive" with "Independent".
- Mutually Exclusive: If A happens, B *cannot* happen. P(A ∩ B) = 0. They are strongly *dependent*.
- Independent: If A happens, it tells you *nothing* about B. P(A ∩ B) = P(A)P(B).

Example:

Experiment: Toss a fair coin twice.

Event A: Get Heads on the 1st toss. P(A) = 1/2.

Event B: Get Heads on the 2nd toss. P(B) = 1/2.

Event (A ∩ B): Get Heads on both. S = {HH, HT, TH, TT}. P(A ∩ B) = 1/4.

Are they independent? Let's check:
Does P(A ∩ B) = P(A) * P(B)?
1/4 = (1/2) * (1/2)
1/4 = 1/4. Yes. The events are independent.


8. Theorem of Total Probability

This theorem is used to find the probability of an event (B) by considering all the possible ways it can happen. It relies on partitioning the sample space.

Let's say the sample space S is partitioned into 'k' mutually exclusive and exhaustive events (A1, A2, ..., Ak). This means one (and only one) of the Ai events must occur.

Now, to find the total probability of an event B, we sum up the probabilities of B happening *within each partition*:

P(B) = P(B ∩ A1) + P(B ∩ A2) + ... + P(B ∩ Ak)

Using the multiplication rule, this becomes:
P(B) = P(B | A1)P(A1) + P(B | A2)P(A2) + ... + P(B | Ak)P(Ak)
P(B) = Σ [ P(B | Ai) * P(Ai) ]

Example:

Two urns, A1 and A2.

We choose an urn at random (P(A1) = 0.5, P(A2) = 0.5) and then draw one ball (Event B = "Ball is White"). What is P(B)?

Using the theorem:

P(B) = P(B | A1)P(A1) + P(B | A2)P(A2)

P(B) = (2/5)(0.5) + (4/5)(0.5) = (1/5) + (2/5) = 3/5

The total probability of getting a white ball is 3/5.


9. Bayes' Theorem and Its Applications

Bayes' Theorem is one of the most important concepts in probability. It allows us to "reverse" the conditional probability.

If we know P(B | A), Bayes' Theorem helps us find P(A | B).

The Theorem

Let A1, A2, ..., Ak be a partition of the sample space. Let B be any event. We want to find the probability of a specific partition, Ai, given that B has occurred.

P(Ai | B) = [ P(B | Ai) * P(Ai) ] / P(B)

We can expand the denominator P(B) using the Theorem of Total Probability:

Bayes' Theorem (Full Form):
P(Ai | B) = [ P(B | Ai)P(Ai) ] / [ Σ P(B | Aj)P(Aj) ]

Application / Example (Classic Disease Test):

A disease (D) affects 1% of the population. P(D) = 0.01.
The other 99% are healthy (H). P(H) = 0.99.

There is a test (T).
- If you have the disease, the test is positive 95% of the time. (P(T+ | D) = 0.95)
- If you are healthy, the test is positive 2% of the time (a "false positive"). (P(T+ | H) = 0.02)

Question: You test positive (T+). What is the probability you *actually have the disease*?
We want to find P(D | T+).

Step 1: Identify the pieces.
- A1 = D (Disease), A2 = H (Healthy)
- B = T+ (Test is positive)
- P(D) = 0.01 (Prior)
- P(H) = 0.99 (Prior)
- P(T+ | D) = 0.95 (Likelihood)
- P(T+ | H) = 0.02 (Likelihood)

Step 2: Calculate the numerator: P(T+ | D) * P(D)
(0.95) * (0.01) = 0.0095

Step 3: Calculate the total denominator, P(T+), using Total Probability.
P(T+) = P(T+ | D)P(D) + P(T+ | H)P(H)
P(T+) = (0.95)(0.01) + (0.02)(0.99)
P(T+) = 0.0095 + 0.0198
P(T+) = 0.0293

Step 4: Divide Numerator by Denominator.
P(D | T+) = 0.0095 / 0.0293 ≈ 0.324

Result Interpretation: Even though you tested positive, there is only a 32.4% chance you actually have the disease. This surprising result is because the false positive rate (2%) applied to the large healthy population (99%) creates more total positive tests than the true positive rate (95%) applied to the small sick population (1%).