Visual Tools
Calculators
Tables
Mathematical Keyboard
Converters
Other Tools


Total Probability






Splitting Probability Across Cases


Some probability questions involve several different ways an outcome can occur. Instead of treating the situation as a single block, it is often clearer to break it into separate cases and look at how each case contributes to the overall probability.

This idea is closely connected to conditional probability, because each case describes the event “if this situation happens, what is the chance of the outcome?” The law of total probability combines those conditional pieces into one overall value.

This structure also sets the stage for Bayes’ theorem. Bayes reverses conditional probabilities, but it relies on the same breakdown of cases that total probability provides. Together, these ideas form one of the main building blocks of probabilistic reasoning.



When You Need Total Probability


This reasoning applies whenever a system can be in different states, and each state affects the outcome differently.

When the outcome depends on which case you're in:
A diagnostic test's accuracy differs for people who actually have the disease versus those who don't. The overall rate of positive tests depends on both groups.

When data comes from mixed sources:
Survey results combine responses from different age groups, regions, or demographics. The overall pattern reflects contributions from each subgroup.

When you don't know the underlying condition:
A machine might be operating in good condition or degraded condition. Failure rates differ in each state, but if you don't know which state it's in, you need to consider both possibilities.

When processes have multiple pathways:
A delivery can be delayed by weather, traffic, or mechanical failure. Each pathway has its own probability, and the total delay probability accounts for all routes.

The intuition is always the same: the final probability is a weighted blend of case-specific probabilities. Each scenario contributes its piece, weighted by how likely that scenario is to occur in the first place.

Splitting a Probability Across Cases


The key idea: you can divide all possible outcomes into distinct, non-overlapping scenarios. Each scenario contributes part of the final probability.

Think of it like splitting a bill at a restaurant. The total amount comes from adding what each person pays. No overlap — each dollar is counted exactly once.

With probability, the sample space gets partitioned into separate cases. An event might occur in Case 1, Case 2, or Case 3. The event's total probability comes from adding its probability within each case.

The core phenomenon:

overall probability = sum of case-specific contributions

This isn't a formula yet — it's the conceptual structure. You're breaking one probability question into smaller, manageable pieces.

Visual intuition:

Imagine a probability tree where branches split into different scenarios. Or picture blocks of the sample space, each colored differently. The event overlaps multiple blocks, and you add up how much probability it captures from each.

Each case contributes its share, and those shares sum to give the complete picture. That's the phenomenon at the heart of total probability.

Useful Notation


    Before expressing the phenomenon formally, we need notation for the key pieces.

    The event of interest:
    Call it A. This is what you're trying to find the probability of.

    The cases (partition):
    Call them B₁, B₂, ..., Bₙ. These are the distinct scenarios that divide the sample space.

    Key properties:

  • Conditional pieces:
    P(A | Bᵢ) is the probability of A given that case Bᵢ occurred. This captures how likely A is within that specific scenario.

    Weights:
    P(Bᵢ) is the probability that case Bᵢ occurs. This tells you how much that scenario contributes to the total.

    Sum notation:
    When we add contributions from all cases, we write:
    ∑ (sum over all i)

    That's the notation toolkit. Each piece has a clear meaning:

  • Next, we combine these pieces into the formal law.

The Law of Total Probability


    Now we formalize the phenomenon described in the previous sections.

    Statement in words:

    If an event A can occur through several disjoint cases B₁, B₂, ..., Bₙ that cover the entire sample space, then the probability of A is the sum of its probability within each case, weighted by the probability of that case occurring.

    The formula:

    P(A) = P(A | B₁)·P(B₁) + P(A | B₂)·P(B₂) + ... + P(A | Bₙ)·P(Bₙ)

    Or using summation notation:

    P(A) = ∑ P(A | Bᵢ)·P(Bᵢ)

    What each term means:


  • Discrete version:

    This is the form shown above, used when you have a finite or countable number of cases.

    Continuous version (conceptual):

    When the partition becomes continuous (like a range of values), the sum becomes an integral:

    P(A) = ∫ P(A | B=b)·f(b) db

    where f(b) is the probability density of the conditioning variable.

    This law expresses mathematically what we've already understood conceptually: probability distributes across cases, and the total comes from adding weighted contributions.