Exponential distribution
From Academic Kids

Template:Probability distribution In probability theory and statistics, the exponential distributions are a class of continuous probability distribution. They are often used to model the time between events that happen at a constant average rate.
Contents 
Specification of the exponential distribution
Probability density function
The probability density function (pdf) of an exponential distribution has the form
 <math>
f(x) = \left\{\begin{matrix} \lambda e^{\lambda x} &,\; x \ge 0, \\ 0 &,\; x < 0. \end{matrix}\right.<math>
where λ > 0 is a parameter of the distribution, often called the rate parameter. The distribution is supported on the interval [0,∞). If a random variable X has this distribution, we write X ~ Exponential(λ).
The exponential distributions can alternatively be parameterized by a scale parameter μ = 1/λ.
Cumulative distribution function
The cumulative distribution function is given by
 <math>
F(x) = \left\{\begin{matrix} 1e^{\lambda x}&,\; x \ge 0, \\ 0 &,\; x < 0. \end{matrix}\right.<math>
Occurrence and applications
The exponential distribution is used to model Poisson processes, which are situations in which an object initially in state A can change to state B with constant probability per unit time λ. The time at which the state actually changes is described by an exponential random variable with parameter λ. Therefore, the integral from 0 to T over f is the probability that the object is in state B at time T.
The exponential distribution may be viewed as a continuous counterpart of the geometric distribution, which describes the number of Bernoulli trials necessary for a discrete process to change state. In contrast, the exponential distribution describes the time for a continuous process to change state.
In real world scenarios, the assumption of a constant rate (or probability per unit time) is rarely satisfied. For example, the rate of incoming phone calls differs according to the time of day. But if we focus on a time interval during which the rate is roughly constant, such as from 2 to 4 PM during work days, the exponential distribution can be used as a good approximate model for the time until the next phone call arrives. Similar caveats apply to the following examples which yield approximately exponentially distributed variables:
 the time until you have your next car accident;
 the time until a radioactive particle decays, or the time between beeps of a geiger counter;
 the number of dice rolls needed until you roll a six 11 times in a row;
 the time until a large meteor strike causes a mass extinction event.
Exponential variables can also be used to model situations where certain events occur with a constant probability per unit distance:
In queueing theory, the interarrival times (i.e. the times between customers entering the system) are often modeled as exponentially distributed variables. The length of a process that can be thought of as a sequence of several independent tasks is better modeled by a variable following the gamma distribution (which is a sum of several independent exponentially distributed variables).
Reliability theory and reliability engineering also make extensive use of the exponential distribution. Because of the memoryless property of this distribution, it is wellsuited to model the constant hazard rate portion of the bathtub curve used in reliability theory. It is also very convenient because it is so easy to add failure rates in a reliability model. The exponential distribution is however not appropriate to model the overall lifetime of organisms or technical devices, because the "failure rates" here are not constant: more failures occur for very young and for very old systems.
In physics, if you observe a gas at a fixed temperature and pressure in a uniform gravitational field, the heights of the various molecules also follow an approximate exponential distribution. This is a consequence of the entropy property mentioned below.
Properties
Mean and standard deviation
The mean or expected value of an exponentially distributed random variable X with rate parameter λ is given by
 <math>\operatorname{E}(X) = \frac{1}{\lambda}<math>
In light of the examples give above, this makes sense: if you receive phone calls at an average rate of 2 per hour, then you can expect to wait half an hour for every call.
The standard deviation of X is also equal to 1/λ.
Memorylessness
An important property of the exponential distribution is that it is memoryless. This means that if a random variable T is exponentially distributed, its conditional probability obeys
 <math>P(T > s + t\; \; T > t) = P(T > s) \;\; \hbox{for all}\ s, t \ge 0. <math>
This says that the conditional probability that we need to wait, for example, more than another 10 seconds before the first arrival, given that the first arrival has not yet happened after 30 seconds, is no different from the initial probability that we need to wait more than 10 seconds for the first arrival. This is often misunderstood by students taking courses on probability: the fact that P(T > 40  T > 30) = P(T > 10) does not mean that the events T > 40 and T > 10 are independent. To summarize: "memorylessness" of the probability distribution of the waiting time T until the first arrival means
 <math>\mathrm{(Right)}\ P(T>40 \mid T>30)=P(T>10).<math>
It does not mean
 <math>\mathrm{(Wrong)}\ P(T>40 \mid T>30)=P(T>40).<math>
(That would be independence. These two events are not independent.)
The exponential distributions are the only continuous memoryless probability distributions.
Quartiles
The quantile function (inverse cumulative distribution function) for Exponential(λ) is
 <math>F^{1}(p) = \frac{\ln(1p)}{\lambda}, \!<math>
for <math>0 \le p < 1<math>. The quartiles are therefore:
 first quartile
 <math>\ln(4/3)/\lambda\,<math>
 median
 <math>\ln(2)/\lambda\,<math>
 third quartile
 <math>\ln(4)/\lambda\,<math>
Entropy
Among all continuous probability distributions with support [0,∞) and mean μ, the exponential distribution with λ = 1/μ has the largest entropy.
Parameter estimation
Suppose you know that a given variable is exponentially distributed and you want to estimate the rate parameter λ.
Maximum likelihood
The likelihood function for λ, given an independent and identically distributed sample x = (x_{1}, ..., x_{n}) drawn from your variable, is
 <math> L(\lambda) = \prod_{i=1}^n \lambda \, \exp(\lambda x_i) = \lambda^n \, \exp\!\left(\!\lambda \sum_{i=1}^n x_i\right)=\lambda^n\exp\left(\lambda n \overline{x}\right) <math>
where
 <math>\overline{x}={1 \over n}\sum_{i=1}^n x_i<math>
is the sample mean.
The derivative of the likelihood function's logarithm is
 <math>\frac{\mathrm{d}}{\mathrm{d}\lambda} \ln L(\lambda) = \frac{\mathrm{d}}{\mathrm{d}\lambda} \left( n \ln(\lambda)  \lambda n\overline{x} \right) = {n \over \lambda}n\overline{x}\ \left\{\begin{matrix} > 0 & \mbox{if}\ 0 < \lambda < 1/\overline{x}, \\ \\ = 0 & \mbox{if}\ \lambda = 1/\overline{x}, \\ \\ < 0 & \mbox{if}\ \lambda > 1/\overline{x}. \end{matrix}\right. <math>
Consequently the maximum likelihood estimate for the rate parameter is
 <math>\widehat{\lambda} = \frac1{\overline{x}}<math>.
Bayesian inference
The conjugate prior for the exponential distribution is the gamma distribution (of which the exponential distribution is a special case). The following parameterization of the gamma pdf is useful:
 <math> \mathrm{Gamma}(\lambda \,\, \alpha, \beta) = \frac{\beta^{\alpha}}{\Gamma(\alpha)} \, \lambda^{\alpha1} \, \exp(\lambda\,\beta) \!<math>
The posterior distribution p can then be expressed in terms of the likelihood function defined above and a gamma prior:
<math>p(\lambda) \!\!\!\!<math>  <math>\propto L(\lambda) \times \mathrm{Gamma}(\lambda \,\, \alpha, \beta)<math> 
<math>= \lambda^n \, \exp(\lambda\,n\overline{x}) \times \frac{\beta^{\alpha}}{\Gamma(\alpha)} \, \lambda^{\alpha1} \, \exp(\lambda\,\beta)<math>  
<math>\propto \lambda^{(\alpha+n)1} \, \exp(\lambda\,(\beta + n\overline{x}))<math> 
Now the posterior density p has been specified up to a missing normalizing constant. Since it has the form of a gamma pdf, this can easily be filled in, and one obtains:
 <math> p(\lambda) = \mathrm{Gamma}(\lambda \,\, \alpha + n, \beta + n \overline{x}) <math>
Here the parameter α can be interpreted as the number of prior observations, and β as the sum of the prior observations.
Generating exponential variates
Given a random variate U drawn from the uniform distribution in the interval (0, 1], the variate
 <math>T=\frac{\ln U}{\lambda} \!<math>
has an exponential distribution with parameter λ. This follows from the form of the quantile function given above and yields a convenient way to produce exponentially distributed values using a random number generator on a computer, for instance to conduct simulation experiments.
Related distributions
 An exponential distribution is a special case of a gamma distribution if <math>\alpha = 1<math> (or <math>k=1<math> depending on the parameter set used).
 <math>Y \sim \mathrm{Weibull}(\gamma, \lambda)<math> is a Weibull distribution if <math>Y = X^{1/\gamma}\,<math> and <math>X \sim \mathrm{Exponential}(\lambda)<math>. In particular, every exponential distribution is also a Weibull distribution (γ = 1).
 <math>Y \sim \mathrm{Rayleigh}(1/\lambda)<math> is a Rayleigh distribution if <math>Y = \sqrt{2X/\lambda}<math> and <math>X \sim \mathrm{Exponential}(\lambda)<math>.
 <math>Y \sim \mathrm{Gumbel}(\mu, \beta)<math> is a Gumbel distribution if <math>Y = \mu  \beta \log(X/\lambda)\,<math> and <math>X \sim \mathrm{Exponential}(\lambda)<math>.
 <math>Y \sim \mathrm{Laplace}<math> is a Laplace distribution if <math>Y = X_1  X_2<math> for two independent exponential distributions <math>X_1<math> and <math>X_2<math>.
 <math>Y \sim \mathrm{Exponential}<math> is an exponential distribution if <math>Y = \min(X_1, X_2, \cdots, X_N)<math> for independent exponential distributions <math>X_i<math>.
 <math>Y \sim \mathrm{Gamma}<math> is a gamma distribution if <math>Y = \sum_{i} X_i\,<math> for independent exponential distributions <math>X_i\,<math>.
 <math>Y \sim \mathrm{Uniform}(0,1)<math> is a uniform distribution if <math>Y = \exp(X/\lambda)\,<math> and <math>X \sim \mathrm{Exponential}(\lambda)<math>.
 <math>X \sim \chi_2^2<math> is a chisquare distribution (with 2 degrees of freedom) if <math>X \sim \mathrm{Exponential}(\lambda = 2)<math>.de:Exponentialverteilung
es:Distribución exponencial fr:Loi exponentielle he:התפלגות מעריכית it:Variabile casuale Esponenziale Negativa sv:Exponentialfördelning