Visual Tools
Calculators
Tables
Mathematical Keyboard
Converters
Other Tools


Exponential Distribution






Exponential Distribution: Waiting Time Until an Event Occurs


The exponential distribution models an experiment where events occur continuously and independently at a constant average rate, and the random variable measures the waiting time until the next event occurs. There are no discrete trials and no fixed endpoint; time flows continuously until the event happens. The defining feature of this experiment is that the probability of waiting longer depends only on the current moment, not on how much time has already passed.



The Probabilistic Experiment Behind exponential distribution


The probabilistic experiment behind the exponential distribution focuses on waiting time rather than counts or measurements. The experiment observes how long it takes until a particular event happens, under the assumption that the event occurs continuously and unpredictably, but at a steady average rate.

A defining characteristic of this experiment is lack of memory. The process does not "age": the chance that the event occurs in the next instant does not depend on how much time has already passed. Waiting longer provides no advantage or disadvantage—the system resets at every moment.

This framework applies when events occur independently, one at a time, and without buildup or anticipation. Time is continuous, and probability accumulates smoothly as time passes.

The exponential distribution captures the idea that short waiting times are common, while long waits are possible but become progressively more rare.

Example:

Consider the time until the next phone call arrives at a quiet call center where calls come in at a stable average rate. Whether the last call arrived a second ago or an hour ago does not affect the likelihood of receiving the next one in the coming minute.

Notation


XExp(λ)X \sim \text{Exp}(\lambda) — distribution of the random variable (rate parameterization).

XExponential(λ)X \sim \text{Exponential}(\lambda) — alternative explicit form.

Exp(λ)\text{Exp}(\lambda) — used to denote the distribution itself (not the random variable).

XExp(β)X \sim \text{Exp}(\beta) — scale parameterization where β=1/λ\beta = 1/\lambda is the mean.

Note: Always check whether the parameter represents the rate (λ) or the scale (β = 1/λ). Statistical software may use either convention. The rate parameterization is more common in probability theory.

See All Probability Symbols and Notations

Parameters


λ (lambda): rate parameter, where λ>0\lambda > 0

The exponential distribution is fully characterized by a single parameter.
λ represents the average rate at which events occur per unit time.
Alternatively, the distribution is sometimes parameterized by β=1/λ\beta = 1/\lambda, the scale parameter, which represents the mean waiting time.
Higher λ means events occur more frequently (shorter waiting times), while lower λ means events are more rare (longer waiting times).

Probability Density Function (PDF) and Support (Range)

Probability Density Function (PDF)


The probability density function (PDF) of an exponential distribution is given by:

f(x)={λeλxif x00if x<0f(x) = \begin{cases} \lambda e^{-\lambda x} & \text{if } x \geq 0 \\ 0 & \text{if } x < 0 \end{cases}


Intuition Behind the Formula


Decreasing Density: The exponential distribution has its highest density at x = 0 and decreases exponentially as x increases. This reflects that shorter waiting times are more likely than longer ones.

Parameters:
λ\lambda: The rate parameter controls how quickly the density decreases
• Higher λ\lambda means steeper decline (events happen more frequently)
• Lower λ\lambda means gentler decline (events happen less frequently)

Support (Range of the Random Variable):
• The random variable XX can take any non-negative value: [0,)[0, \infty)
• Time cannot be negative, so the support starts at zero
• Theoretically unbounded above, though very large values have low probability
• The support is the half-infinite interval [0,)[0, \infty)

Logic Behind the Formula:
λ\lambda: The rate parameter also serves as the initial density at x = 0
eλxe^{-\lambda x}: Exponential decay function ensuring rapid decrease for larger x
• The decay rate is proportional to λ: faster decay for larger λ
• The total area under the curve equals 1:

0λeλxdx=λ[1λeλx]0=1\int_{0}^{\infty} \lambda e^{-\lambda x}\,dx = \lambda \left[-\frac{1}{\lambda}e^{-\lambda x}\right]_{0}^{\infty} = 1


Practical Example: Customer service calls arrive at a help desk at an average rate of 5 per hour, so λ=5\lambda = 5. The time XX (in hours) until the next call follows Exp(5)\text{Exp}(5). The PDF is f(x)=5e5xf(x) = 5e^{-5x} for x0x \geq 0. The probability of waiting less than 6 minutes (0.1 hours) is P(X0.1)=1e5(0.1)0.393P(X \leq 0.1) = 1 - e^{-5(0.1)} \approx 0.393, or about 39.3%.

Exponential Distribution

Time between events in a Poisson process

Explanation

The exponential distribution models the time between events in a Poisson process, where events occur continuously and independently at a constant average rate. The probability density function is f(x)=λeλxf(x) = \lambda e^{-\lambda x} for x0x \geq 0, where λ\lambda is the rate parameter. The expected value is E[X]=1λE[X] = \frac{1}{\lambda} and the variance is Var(X)=1λ2\text{Var}(X) = \frac{1}{\lambda^2}. The exponential distribution has the memoryless property, meaning the probability of an event occurring in the next interval is independent of how much time has already passed. Common applications include time until equipment failure, waiting time between customer arrivals, radioactive decay, and time until the next earthquake.


Cumulative Distribution Function (CDF)


Cumulative Distribution Function (CDF)


The cumulative distribution function (CDF) gives the probability that XX is less than or equal to a specific value:

F(x)={0if x<01eλxif x0F(x) = \begin{cases} 0 & \text{if } x < 0 \\ 1 - e^{-\lambda x} & \text{if } x \geq 0 \end{cases}


Key Properties:
F(0)=0F(0) = 0 (no probability mass below zero)
F(x)1F(x) \to 1 as xx \to \infty (all probability is non-negative)
• The CDF increases from 0 to 1, starting slowly and approaching 1 asymptotically
• For any x0x \geq 0, P(X>x)=1F(x)=eλxP(X > x) = 1 - F(x) = e^{-\lambda x} (survival function)

Memoryless Property: The exponential distribution is the only continuous distribution with the memoryless property: P(X>s+tX>s)=P(X>t)P(X > s + t \mid X > s) = P(X > t). In terms of the CDF: knowing that X>sX > s doesn't change the probability that X>s+tX > s + t relative to waiting an additional time tt from the start.

Practical Use: The exponential CDF is one of the simplest to work with. To find P(Xx)P(X \leq x) when XExp(3)X \sim \text{Exp}(3) and x=2x = 2: F(2)=1e3(2)=1e60.9975F(2) = 1 - e^{-3(2)} = 1 - e^{-6} \approx 0.9975, meaning about 99.75% of values fall below 2. The complementary probability P(X>2)=e60.0025P(X > 2) = e^{-6} \approx 0.0025.

Exponential Distribution CDF

Visualizing probability accumulation for exponential distribution

Exponential - CDF

Rapid initial rise, asymptotic approach to 1

CDF Explanation

The cumulative distribution function (CDF) of the exponential distribution is F(x)=1eλxF(x) = 1 - e^{-\lambda x} for x0x \geq 0, and F(x)=0F(x) = 0 for x<0x < 0. The CDF shows the probability that the waiting time is less than or equal to xx. The curve rises quickly at first and then asymptotically approaches 1. The memoryless property of the exponential distribution means that P(X>s+tX>s)=P(X>t)P(X > s + t | X > s) = P(X > t) for all s,t0s, t \geq 0. The median waiting time (where F(x)=0.5F(x) = 0.5) is ln(2)λ\frac{\ln(2)}{\lambda}, and about 63.2% of events occur within the first 1λ\frac{1}{\lambda} time units.

Expected Value (Mean)


Computing the mean of a continuous random variable involves integration rather than summation, weighting each point by its probability density. The continuous expected value framework applies directly to the exponential distribution:

E[X]=xf(x)dxE[X] = \int_{-\infty}^{\infty} x \cdot f(x) \, dx


Formula


E[X]=1λE[X] = \frac{1}{\lambda}


Where:
λ\lambda = the rate parameter (events per unit time)

Derivation and Intuition


The exponential distribution has PDF f(x)=λeλxf(x) = \lambda e^{-\lambda x} for x0x \geq 0. Computing the expected value:

E[X]=0xλeλxdxE[X] = \int_{0}^{\infty} x \cdot \lambda e^{-\lambda x} \, dx


Using integration by parts with u=xu = x and dv=λeλxdxdv = \lambda e^{-\lambda x} dx:

E[X]=[xeλx]0+0eλxdxE[X] = \left[-x e^{-\lambda x}\right]_{0}^{\infty} + \int_{0}^{\infty} e^{-\lambda x} \, dx


The first term vanishes at both limits. For the second term:

E[X]=0eλxdx=[1λeλx]0=0(1λ)=1λE[X] = \int_{0}^{\infty} e^{-\lambda x} \, dx = \left[-\frac{1}{\lambda} e^{-\lambda x}\right]_{0}^{\infty} = 0 - \left(-\frac{1}{\lambda}\right) = \frac{1}{\lambda}


The result E[X]=1λE[X] = \frac{1}{\lambda} captures an intuitive relationship: if events occur at rate λ\lambda per unit time, then the average waiting time until the next event is 1λ\frac{1}{\lambda} time units. A higher rate means shorter average wait times; a lower rate means longer waits.

This parallels the geometric distribution, which models waiting in discrete trials. The exponential distribution is the continuous analog, modeling waiting in continuous time.

Example


Suppose customers arrive at a store at an average rate of λ=3\lambda = 3 customers per hour:

E[X]=13 hours=20 minutesE[X] = \frac{1}{3} \text{ hours} = 20 \text{ minutes}


The expected time until the next customer arrives is 20 minutes. If the rate increases to λ=6\lambda = 6 customers per hour, the expected wait drops to 16\frac{1}{6} hours, or 10 minutes.

Variance and Standard Deviation


The variance of a continuous random variable quantifies the spread of values around the mean. For continuous distributions, it is calculated through integration:

Var(X)=E[(Xμ)2]=(xμ)2f(x)dx\mathrm{Var}(X) = \mathbb{E}[(X - \mu)^2] = \int_{-\infty}^{\infty} (x - \mu)^2 f(x) \, dx


Alternatively, using the computational formula:

Var(X)=E[X2]μ2\mathrm{Var}(X) = \mathbb{E}[X^2] - \mu^2


For the exponential distribution, this calculation reveals how the spread relates to the rate parameter.

Formula


Var(X)=1λ2\mathrm{Var}(X) = \frac{1}{\lambda^2}


Where:
λ\lambda = the rate parameter (events per unit time)

Derivation and Intuition


Starting with the computational formula, we need to calculate E[X2]\mathbb{E}[X^2].

We know from the expected value section that μ=1λ\mu = \frac{1}{\lambda}.

The exponential distribution has PDF f(x)=λeλxf(x) = \lambda e^{-\lambda x} for x0x \geq 0:

E[X2]=0x2λeλxdx\mathbb{E}[X^2] = \int_{0}^{\infty} x^2 \cdot \lambda e^{-\lambda x} \, dx


Using integration by parts twice (or the gamma function), we obtain:

E[X2]=2λ2\mathbb{E}[X^2] = \frac{2}{\lambda^2}


Applying the computational formula:

Var(X)=2λ2(1λ)2=2λ21λ2=1λ2\mathrm{Var}(X) = \frac{2}{\lambda^2} - \left(\frac{1}{\lambda}\right)^2 = \frac{2}{\lambda^2} - \frac{1}{\lambda^2} = \frac{1}{\lambda^2}


The result Var(X)=1λ2\mathrm{Var}(X) = \frac{1}{\lambda^2} shows that variance decreases rapidly as the rate increases. When events occur frequently (large λ\lambda), waiting times are consistently short with little variability. When events are rare (small λ\lambda), waiting times become both long and highly variable. The quadratic relationship means doubling the rate reduces variance by a factor of four.

Notably, for the exponential distribution, the standard deviation equals the mean—a unique property that doesn't hold for most distributions.

Standard Deviation


σ=1λ\sigma = \frac{1}{\lambda}


Example


Suppose customers arrive at a store at an average rate of λ=3\lambda = 3 customers per hour:

Var(X)=1(3)2=19 hours20.111 hours2\mathrm{Var}(X) = \frac{1}{(3)^2} = \frac{1}{9} \text{ hours}^2 \approx 0.111 \text{ hours}^2


σ=13 hours=20 minutes\sigma = \frac{1}{3} \text{ hours} = 20 \text{ minutes}


The standard deviation of 20 minutes equals the mean waiting time of 20 minutes. This indicates substantial variability: sometimes the next customer arrives in just a few minutes, other times you might wait 40+ minutes, even though the average is 20 minutes.

Mode and Median

Mode


The mode is the value where the probability density function reaches its maximum—the peak of the distribution curve.

For the exponential distribution, the mode is always:

Mode=0\text{Mode} = 0


Intuition: The exponential PDF is f(x)=λeλxf(x) = \lambda e^{-\lambda x} for x0x \geq 0, which is a monotonically decreasing function. The density is highest at x=0x = 0 and decreases exponentially as xx increases.

At x=0x = 0: f(0)=λf(0) = \lambda

For any x>0x > 0: f(x)=λeλx<λf(x) = \lambda e^{-\lambda x} < \lambda

The maximum always occurs at the left boundary of the support. This reflects the nature of waiting times—the shortest possible wait (zero) is always the most probable density value, though the probability of any exact value is zero (as with all continuous distributions).

Example:
For customer arrivals with λ=3\lambda = 3 per hour:

Mode = 0

The density is highest at zero waiting time, then decays exponentially. While you can't actually wait exactly zero time, the probability density is concentrated near zero, making very short waits most likely.

Median


The median is the value mm such that P(Xm)=0.5P(X \leq m) = 0.5—the point that divides the distribution's probability in half.

For the exponential distribution, the median can be found using the CDF F(x)=1eλxF(x) = 1 - e^{-\lambda x}:

Setting F(m)=0.5F(m) = 0.5:

1eλm=0.51 - e^{-\lambda m} = 0.5


eλm=0.5e^{-\lambda m} = 0.5


λm=ln(0.5)=ln(2)-\lambda m = \ln(0.5) = -\ln(2)


m=ln(2)λm = \frac{\ln(2)}{\lambda}


The median is:

Median=ln(2)λ0.693λ\text{Median} = \frac{\ln(2)}{\lambda} \approx \frac{0.693}{\lambda}


Example:
For customer arrivals with λ=3\lambda = 3 per hour:

Median=ln(2)30.69330.231 hours13.9 minutes\text{Median} = \frac{\ln(2)}{3} \approx \frac{0.693}{3} \approx 0.231 \text{ hours} \approx 13.9 \text{ minutes}


Half of the waiting times are less than 13.9 minutes, and half are greater.

Properties:
• The exponential distribution is right-skewed: mode < median < mean
• Mode = 0, Median = ln(2)λ0.693×1λ\frac{\ln(2)}{\lambda} \approx 0.693 \times \frac{1}{\lambda}, Mean = 1λ\frac{1}{\lambda}
• The median is always smaller than the mean: about 69.3% of the mean value
• This reflects the right skew—most observations cluster near zero, but occasional long waits pull the mean higher
• The relationship Median/Mean ≈ 0.693 holds for all exponential distributions regardless of λ\lambda

Quantiles/Percentiles


A quantile is a value that divides the distribution at a specific probability threshold. The pp-th quantile xpx_p satisfies:

P(Xxp)=pP(X \leq x_p) = p


where 0<p<10 < p < 1.

Percentiles are quantiles expressed as percentages: the kk-th percentile corresponds to the quantile at p=k/100p = k/100. For example, the 25th percentile is the 0.25 quantile, the 50th percentile is the median, and the 75th percentile is the 0.75 quantile.

Quantiles are found by inverting the CDF: if F(xp)=pF(x_p) = p, then xp=F1(p)x_p = F^{-1}(p).

Finding Quantiles for the Exponential Distribution


For an exponential distribution with rate parameter λ\lambda, the CDF is:

F(x)=1eλxF(x) = 1 - e^{-\lambda x}


To find the pp-th quantile, we solve F(xp)=pF(x_p) = p:

1eλxp=p1 - e^{-\lambda x_p} = p


eλxp=1pe^{-\lambda x_p} = 1 - p


λxp=ln(1p)-\lambda x_p = \ln(1 - p)


xp=ln(1p)λ=ln(1p)λx_p = -\frac{\ln(1 - p)}{\lambda} = \frac{-\ln(1 - p)}{\lambda}


This gives the general quantile formula:

xp=ln(1p)λx_p = \frac{-\ln(1 - p)}{\lambda}


Unlike the normal distribution, the exponential distribution has a closed-form expression for quantiles—no tables or numerical methods needed.

Common Percentiles


25th Percentile (First Quartile, Q1):

x0.25=ln(10.25)λ=ln(0.75)λ=0.288λx_{0.25} = \frac{-\ln(1 - 0.25)}{\lambda} = \frac{-\ln(0.75)}{\lambda} = \frac{0.288}{\lambda}


About 25% of waiting times are less than 0.288λ\frac{0.288}{\lambda}.

50th Percentile (Median, Q2):

x0.50=ln(10.50)λ=ln(0.50)λ=ln(2)λ0.693λx_{0.50} = \frac{-\ln(1 - 0.50)}{\lambda} = \frac{-\ln(0.50)}{\lambda} = \frac{\ln(2)}{\lambda} \approx \frac{0.693}{\lambda}


This matches the median derived earlier.

75th Percentile (Third Quartile, Q3):

x0.75=ln(10.75)λ=ln(0.25)λ=1.386λx_{0.75} = \frac{-\ln(1 - 0.75)}{\lambda} = \frac{-\ln(0.25)}{\lambda} = \frac{1.386}{\lambda}


About 75% of waiting times are less than 1.386λ\frac{1.386}{\lambda}.

Interquartile Range (IQR):

IQR=Q3Q1=1.386λ0.288λ=1.098λ\text{IQR} = Q3 - Q1 = \frac{1.386}{\lambda} - \frac{0.288}{\lambda} = \frac{1.098}{\lambda}


The IQR contains the middle 50% of the distribution.

Example


For customer arrivals with λ=3\lambda = 3 per hour:

25th percentile: 0.2883=0.096\frac{0.288}{3} = 0.096 hours 5.8\approx 5.8 minutes

25% of waits are less than 5.8 minutes.

50th percentile: 0.6933=0.231\frac{0.693}{3} = 0.231 hours 13.9\approx 13.9 minutes

Half of waits are less than 13.9 minutes (the median).

75th percentile: 1.3863=0.462\frac{1.386}{3} = 0.462 hours 27.7\approx 27.7 minutes

75% of waits are less than 27.7 minutes.

IQR: 1.0983=0.366\frac{1.098}{3} = 0.366 hours 22\approx 22 minutes

The middle 50% of waiting times span about 22 minutes.

Other Notable Percentiles


90th percentile: x0.90=ln(0.10)λ=2.303λx_{0.90} = \frac{-\ln(0.10)}{\lambda} = \frac{2.303}{\lambda} (only 10% of waits exceed this)

95th percentile: x0.95=ln(0.05)λ=2.996λx_{0.95} = \frac{-\ln(0.05)}{\lambda} = \frac{2.996}{\lambda} (only 5% of waits exceed this)

99th percentile: x0.99=ln(0.01)λ=4.605λx_{0.99} = \frac{-\ln(0.01)}{\lambda} = \frac{4.605}{\lambda} (only 1% of waits exceed this)

The exponential distribution's right skew means that while most observations cluster near zero, the upper percentiles can be quite large relative to the mean 1λ\frac{1}{\lambda}.

Real-World Examples and Common Applications


The exponential distribution models waiting times and lifetimes when events occur randomly and independently at a constant rate.

Common Applications


Reliability Engineering and Survival Analysis:
• Component failure times (electronics, mechanical parts)
• System reliability modeling
• Time until equipment breakdown
• Light bulb lifetimes

Queueing Theory and Service Systems:
• Time between customer arrivals
• Call center inter-arrival times
• Web server request intervals
• Traffic flow patterns

Physics and Natural Sciences:
• Radioactive decay intervals
• Time between cosmic ray detections
• Molecular collision intervals
• Photon emission times

Insurance and Actuarial Science:
• Time until next insurance claim
• Inter-occurrence times for accidents
• Time between policy cancellations

Why It Appears


The exponential distribution is the continuous analog of the geometric distribution. It emerges naturally from Poisson processes—when events follow a Poisson process with rate λ\lambda, the time until the next event follows an exponential distribution with the same rate parameter.

The memoryless property makes it unique: P(X>s+tX>s)=P(X>t)P(X > s + t \mid X > s) = P(X > t). Past waiting time provides no information about future waiting time.

Example Application


A web server receives requests at an average rate of 120 per hour (λ=2\lambda = 2 per minute). The time between consecutive requests follows Exp(2)\text{Exp}(2).

The median wait between requests is ln(2)20.35\frac{\ln(2)}{2} \approx 0.35 minutes (21 seconds). However, 10% of inter-arrival times exceed 2.32=1.15\frac{2.3}{2} = 1.15 minutes, requiring server capacity planning for occasional gaps.

Interactive Calculator

Exponential Distribution Calculator

Calculate probabilities and distribution properties

Average rate of occurrence


Special Cases


The exponential distribution has several special cases and limiting behaviors that reveal its connections to other distributions and its unique mathematical properties.

Unit Exponential Distribution


When λ=1\lambda = 1, we obtain the unit exponential distribution Exp(1)\text{Exp}(1):

f(x)=ex for x0f(x) = e^{-x} \text{ for } x \geq 0


This simplified form is often used in theoretical work. Any exponential random variable XExp(λ)X \sim \text{Exp}(\lambda) can be transformed to unit exponential:

Y=λXExp(1)Y = \lambda X \sim \text{Exp}(1)


As λ → 0 (Rate Decreases)


As λ0\lambda \to 0, events become increasingly rare:

• The mean 1λ\frac{1}{\lambda} \to \infty (average wait grows without bound)
• The distribution becomes more spread out
• Most probability mass shifts toward larger values
• The density near zero approaches zero

This represents the limit of extremely infrequent events.

As λ → ∞ (Rate Increases)


As λ\lambda \to \infty, events occur nearly instantaneously:

• The mean 1λ0\frac{1}{\lambda} \to 0 (average wait shrinks to zero)
• All probability concentrates near zero
• The distribution degenerates toward a point mass at zero

This represents the limit where waiting time becomes negligible.

Limiting Behavior and Connections


Relationship to Poisson:
The exponential distribution is intimately connected to the Poisson distribution. If events follow a Poisson process with rate λ\lambda, then:

• The time until the first event is Exp(λ)\text{Exp}(\lambda)
• The time between consecutive events is Exp(λ)\text{Exp}(\lambda)

The exponential is the continuous-time analog of the geometric distribution.

Minimum of Exponentials:
If X1Exp(λ1)X_1 \sim \text{Exp}(\lambda_1) and X2Exp(λ2)X_2 \sim \text{Exp}(\lambda_2) are independent, then:

min(X1,X2)Exp(λ1+λ2)\min(X_1, X_2) \sim \text{Exp}(\lambda_1 + \lambda_2)


The minimum waiting time has rate equal to the sum of rates. This property is crucial in reliability theory (system fails when first component fails).

Memoryless Property:
The exponential distribution is the only continuous distribution with the memoryless property:

P(X>s+tX>s)=P(X>t)P(X > s + t \mid X > s) = P(X > t)


This makes it the natural choice for modeling random failures and arrivals.

Practical Implications


Reliability Analysis:
When λ\lambda is very small (reliable components), failures are rare and the exponential model applies over long time spans. As λ\lambda increases (less reliable components), the model still applies but predicts more frequent failures.

Queueing Systems:
In service systems, when λ\lambda is large (high arrival rate), the system stays busy. As λ\lambda decreases (low traffic), the server experiences longer idle periods between customers.

Properties


The exponential distribution has several distinctive properties, most notably the memoryless property that uniquely characterizes it among continuous distributions.

Symmetry


The exponential distribution is not symmetric. It has:
• Support on [0,)[0, \infty) only
• Maximum density at x=0x = 0
• Long right tail extending to infinity
• Strong positive skew

Skewness


Skewness=2\text{Skewness} = 2


The skewness coefficient of 2 indicates substantial right skew. Most probability mass concentrates near zero, but the long tail pulls the mean rightward.

Kurtosis


Kurtosis=9\text{Kurtosis} = 9


Excess Kurtosis=6\text{Excess Kurtosis} = 6


High kurtosis (9 versus 3 for normal) indicates the exponential distribution has heavier tails than the normal distribution. Extreme values occur more frequently than in normal or uniform distributions.

Tail Behavior


The exponential distribution has exponentially decaying right tail:

f(x)=λeλx0 as xf(x) = \lambda e^{-\lambda x} \to 0 \text{ as } x \to \infty


The tail decreases exponentially, which is:
• Heavier than normal tails (which decay like ex2e^{-x^2})
• Lighter than polynomial tails (like power laws)
• Characteristic of many natural decay processes

Probabilities in the tail:
P(X>2λ)=e20.135P(X > \frac{2}{\lambda}) = e^{-2} \approx 0.135 (13.5%)
P(X>3λ)=e30.050P(X > \frac{3}{\lambda}) = e^{-3} \approx 0.050 (5%)
P(X>5λ)=e50.007P(X > \frac{5}{\lambda}) = e^{-5} \approx 0.007 (0.7%)

Unique Mathematical Properties


Memoryless Property (Defining Characteristic):
P(X>s+tX>s)=P(X>t) for all s,t0P(X > s + t \mid X > s) = P(X > t) \text{ for all } s, t \geq 0


The exponential is the only continuous distribution with this property. Past waiting time provides no information about future waiting time. This makes it ideal for modeling random arrivals and failures without aging effects.

Equivalently:
P(X>s+t)=P(X>s)P(X>t)P(X > s + t) = P(X > s) \cdot P(X > t)


Constant Hazard Rate:
The hazard function (instantaneous failure rate) is constant:

h(x)=f(x)1F(x)=λeλxeλx=λh(x) = \frac{f(x)}{1 - F(x)} = \frac{\lambda e^{-\lambda x}}{e^{-\lambda x}} = \lambda


The failure rate doesn't depend on age—components don't "wear out" under the exponential model.

Closure Under Minimum:
If X1Exp(λ1)X_1 \sim \text{Exp}(\lambda_1) and X2Exp(λ2)X_2 \sim \text{Exp}(\lambda_2) are independent:

min(X1,X2)Exp(λ1+λ2)\min(X_1, X_2) \sim \text{Exp}(\lambda_1 + \lambda_2)


The minimum of exponentials is exponential with rate equal to the sum of rates.

Scaling Property:
If XExp(λ)X \sim \text{Exp}(\lambda), then for any c>0c > 0:

cXExp(λ/c)cX \sim \text{Exp}(\lambda/c)


Rescaling by cc divides the rate by cc.

Useful Identities


Relationship to Poisson:
If events follow a Poisson process with rate λ\lambda, the time between events is Exp(λ)\text{Exp}(\lambda). The number of events in time tt is Poisson(λt)\text{Poisson}(\lambda t).

Mean Equals Standard Deviation:
E[X]=σ=1λE[X] = \sigma = \frac{1}{\lambda}


This unique relationship (mean = standard deviation) holds only for the exponential distribution.

Lack of Memory Formula:
E[XtX>t]=E[X]=1λE[X - t \mid X > t] = E[X] = \frac{1}{\lambda}


The expected remaining waiting time equals the original expected waiting time—knowing you've already waited tt time units doesn't change your expectation.

Survival Function:
S(x)=P(X>x)=eλxS(x) = P(X > x) = e^{-\lambda x}


Simple exponential decay makes many calculations tractable.

Sum of Exponentials:
The sum of nn independent Exp(λ)\text{Exp}(\lambda) variables follows a Gamma(n,λ)(n, \lambda) distribution, generalizing the exponential.

Related Distributions