Weibull (2-parameter)
Probability density function
Cumulative distribution function
Parameters λ ( 0 , + ) {\displaystyle \lambda \in (0,+\infty )\,} {\displaystyle \lambda \in (0,+\infty )\,} scale
k ( 0 , + ) {\displaystyle k\in (0,+\infty )\,} {\displaystyle k\in (0,+\infty )\,} shape
Support x [ 0 , + ) {\displaystyle x\in [0,+\infty )\,} {\displaystyle x\in [0,+\infty )\,}
PDF f ( x ) = { k λ ( x λ ) k 1 e ( x / λ ) k , x 0 , 0 , x < 0. {\displaystyle f(x)={\begin{cases}{\frac {k}{\lambda }}\left({\frac {x}{\lambda }}\right)^{k-1}e^{-(x/\lambda )^{k}},&x\geq 0,\\0,&x<0.\end{cases}}} {\displaystyle f(x)={\begin{cases}{\frac {k}{\lambda }}\left({\frac {x}{\lambda }}\right)^{k-1}e^{-(x/\lambda )^{k}},&x\geq 0,\\0,&x<0.\end{cases}}}
CDF F ( x ) = { 1 e ( x / λ ) k , x 0 , 0 , x < 0. {\displaystyle F(x)={\begin{cases}1-e^{-(x/\lambda )^{k}},&x\geq 0,\\0,&x<0.\end{cases}}} {\displaystyle F(x)={\begin{cases}1-e^{-(x/\lambda )^{k}},&x\geq 0,\\0,&x<0.\end{cases}}}
Quantile Q ( p ) = λ ( ln ( 1 p ) ) 1 k {\displaystyle Q(p)=\lambda (-\ln(1-p))^{\frac {1}{k}}} {\displaystyle Q(p)=\lambda (-\ln(1-p))^{\frac {1}{k}}}
Mean λ Γ ( 1 + 1 / k ) {\displaystyle \lambda \,\Gamma (1+1/k)\,} {\displaystyle \lambda \,\Gamma (1+1/k)\,}
Median λ ( ln 2 ) 1 / k {\displaystyle \lambda (\ln 2)^{1/k}\,} {\displaystyle \lambda (\ln 2)^{1/k}\,}
Mode { λ ( k 1 k ) 1 / k , k > 1 , 0 , k 1. {\displaystyle {\begin{cases}\lambda \left({\frac {k-1}{k}}\right)^{1/k}\,,&k>1,\\0,&k\leq 1.\end{cases}}} {\displaystyle {\begin{cases}\lambda \left({\frac {k-1}{k}}\right)^{1/k}\,,&k>1,\\0,&k\leq 1.\end{cases}}}
Variance λ 2 [ Γ ( 1 + 2 k ) ( Γ ( 1 + 1 k ) ) 2 ] {\displaystyle \lambda ^{2}\left[\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}\right]\,} {\displaystyle \lambda ^{2}\left[\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}\right]\,}
Skewness Γ ( 1 + 3 / k ) λ 3 3 μ σ 2 μ 3 σ 3 {\displaystyle {\frac {\Gamma (1+3/k)\lambda ^{3}-3\mu \sigma ^{2}-\mu ^{3}}{\sigma ^{3}}}} {\displaystyle {\frac {\Gamma (1+3/k)\lambda ^{3}-3\mu \sigma ^{2}-\mu ^{3}}{\sigma ^{3}}}}
Excess kurtosis (see text)
Entropy γ ( 1 1 / k ) + ln ( λ / k ) + 1 {\displaystyle \gamma (1-1/k)+\ln(\lambda /k)+1\,} {\displaystyle \gamma (1-1/k)+\ln(\lambda /k)+1\,}
MGF n = 0 t n λ n n ! Γ ( 1 + n / k ) ,   k 1 {\displaystyle \sum _{n=0}^{\infty }{\frac {t^{n}\lambda ^{n}}{n!}}\Gamma (1+n/k),\ k\geq 1} {\displaystyle \sum _{n=0}^{\infty }{\frac {t^{n}\lambda ^{n}}{n!}}\Gamma (1+n/k),\ k\geq 1}
CF n = 0 ( i t ) n λ n n ! Γ ( 1 + n / k ) {\displaystyle \sum _{n=0}^{\infty }{\frac {(it)^{n}\lambda ^{n}}{n!}}\Gamma (1+n/k)} {\displaystyle \sum _{n=0}^{\infty }{\frac {(it)^{n}\lambda ^{n}}{n!}}\Gamma (1+n/k)}
Kullback–Leibler divergence see below

In probability theory and statistics, the Weibull distribution /ˈwbʊl/ is a continuous probability distribution. It models a broad range of random variables, largely in the nature of a time to failure or time between events. Examples are maximum one-day rainfalls and the time a user spends on a web page.

The distribution is named after Swedish mathematician Waloddi Weibull, who described it in detail in 1939,[1][2] although it was first identified by René Maurice Fréchet and first applied by Rosin & Rammler (1933) to describe a particle size distribution.[3]

Definition

Standard parameterization

The probability density function of a Weibull random variable is[4][5]

f ( x ; λ , k ) = { k λ ( x λ ) k 1 e ( x / λ ) k , x 0 , 0 , x < 0 , {\displaystyle f(x;\lambda ,k)={\begin{cases}{\frac {k}{\lambda }}\left({\frac {x}{\lambda }}\right)^{k-1}e^{-(x/\lambda )^{k}},&x\geq 0,\\0,&x<0,\end{cases}}} {\displaystyle f(x;\lambda ,k)={\begin{cases}{\frac {k}{\lambda }}\left({\frac {x}{\lambda }}\right)^{k-1}e^{-(x/\lambda )^{k}},&x\geq 0,\\0,&x<0,\end{cases}}}

where k > 0 is the shape parameter and λ > 0 is the scale parameter of the distribution. Its complementary cumulative distribution function is a stretched exponential function. The Weibull distribution is related to a number of other probability distributions; in particular, it interpolates between the exponential distribution (k = 1) and the Rayleigh distribution (k = 2 and λ = 2 σ {\displaystyle \lambda ={\sqrt {2}}\sigma } {\displaystyle \lambda ={\sqrt {2}}\sigma }).[6]

If the quantity, x, is a "time-to-failure", the Weibull distribution gives a distribution for which the failure rate is proportional to a power of time. The shape parameter, k, is that power plus one, and so this parameter can be interpreted directly as follows:[7]

In the field of materials science, the shape parameter k of a distribution of strengths is known as the Weibull modulus. In the context of diffusion of innovations, the Weibull distribution is a "pure" imitation/rejection model.

Optional parameterizations

First option

Applications in medical statistics and econometrics often adopt a different parameterization.[9][10] The shape parameter k is the same as above, while the scale parameter is b = λ k {\displaystyle b=\lambda ^{-k}} {\displaystyle b=\lambda ^{-k}}. In this case, for x ≥ 0, the probability density function is

f ( x ; k , b ) = b k x k 1 e b x k , {\displaystyle f(x;k,b)=bkx^{k-1}e^{-bx^{k}},} {\displaystyle f(x;k,b)=bkx^{k-1}e^{-bx^{k}},}

the cumulative distribution function is

F ( x ; k , b ) = 1 e b x k , {\displaystyle F(x;k,b)=1-e^{-bx^{k}},} {\displaystyle F(x;k,b)=1-e^{-bx^{k}},}

the quantile function is

Q ( p ; k , b ) = ( 1 b ln ( 1 p ) ) 1 k , {\displaystyle Q(p;k,b)=\left(-{\frac {1}{b}}\ln(1-p)\right)^{\frac {1}{k}},} {\displaystyle Q(p;k,b)=\left(-{\frac {1}{b}}\ln(1-p)\right)^{\frac {1}{k}},}

the hazard function is

h ( x ; k , b ) = b k x k 1 , {\displaystyle h(x;k,b)=bkx^{k-1},} {\displaystyle h(x;k,b)=bkx^{k-1},}

and the mean is

b 1 / k Γ ( 1 + 1 / k ) . {\displaystyle b^{-1/k}\Gamma (1+1/k).} {\displaystyle b^{-1/k}\Gamma (1+1/k).}

Second option

A second parameterization option can also be found.[11][12] The shape parameter k is the same as in the standard case, while the scale parameter λ is replaced with a rate parameter β = 1/λ. Then, for x ≥ 0, the probability density function is

f ( x ; k , β ) = β k ( β x ) k 1 e ( β x ) k {\displaystyle f(x;k,\beta )=\beta k({\beta x})^{k-1}e^{-(\beta x)^{k}}} {\displaystyle f(x;k,\beta )=\beta k({\beta x})^{k-1}e^{-(\beta x)^{k}}}

the cumulative distribution function is

F ( x ; k , β ) = 1 e ( β x ) k , {\displaystyle F(x;k,\beta )=1-e^{-(\beta x)^{k}},} {\displaystyle F(x;k,\beta )=1-e^{-(\beta x)^{k}},}

the quantile function is

Q ( p ; k , β ) = 1 β ( ln ( 1 p ) ) 1 k , {\displaystyle Q(p;k,\beta )={\frac {1}{\beta }}(-\ln(1-p))^{\frac {1}{k}},} {\displaystyle Q(p;k,\beta )={\frac {1}{\beta }}(-\ln(1-p))^{\frac {1}{k}},}

and the hazard function is

h ( x ; k , β ) = β k ( β x ) k 1 . {\displaystyle h(x;k,\beta )=\beta k({\beta x})^{k-1}.} {\displaystyle h(x;k,\beta )=\beta k({\beta x})^{k-1}.}

In all three parameterizations, the hazard is decreasing for k < 1, increasing for k > 1 and constant for k = 1, in which case the Weibull distribution reduces to an exponential distribution.

Properties

Density function

The form of the density function of the Weibull distribution changes drastically with the value of k. For 0 < k < 1, the density function tends to ∞ as x approaches zero from above and is strictly decreasing. For k = 1, the density function tends to 1/λ as x approaches zero from above and is strictly decreasing. For k > 1, the density function tends to zero as x approaches zero from above, increases until its mode and decreases after it. The density function has infinite negative slope at x = 0 if 0 < k < 1, infinite positive slope at x = 0 if 1 < k < 2 and null slope at x = 0 if k > 2. For k = 1 the density has a finite negative slope at x = 0. For k = 2 the density has a finite positive slope at x = 0. As k goes to infinity, the Weibull distribution converges to a Dirac delta distribution centered at x = λ. Moreover, the skewness and coefficient of variation depend only on the shape parameter. A generalization of the Weibull distribution is the hyperbolastic distribution of type III.

Cumulative distribution function

The cumulative distribution function for the Weibull distribution is

F ( x ; k , λ ) = 1 e ( x / λ ) k {\displaystyle F(x;k,\lambda )=1-e^{-(x/\lambda )^{k}}\,} {\displaystyle F(x;k,\lambda )=1-e^{-(x/\lambda )^{k}}\,}

for x ≥ 0, and F(x; k; λ) = 0 for x < 0.

If x = λ then F(x; k; λ) = 1 − e−1 ≈ 0.632 for all values of k. Vice versa: at F(x; k; λ) = 0.632 the value of x ≈ λ.

The quantile (inverse cumulative distribution) function for the Weibull distribution is

Q ( p ; k , λ ) = λ ( ln ( 1 p ) ) 1 / k {\displaystyle Q(p;k,\lambda )=\lambda (-\ln(1-p))^{1/k}} {\displaystyle Q(p;k,\lambda )=\lambda (-\ln(1-p))^{1/k}}

for 0 ≤ p < 1.

The failure rate h (or hazard function) is given by

h ( x ; k , λ ) = k λ ( x λ ) k 1 . {\displaystyle h(x;k,\lambda )={k \over \lambda }\left({x \over \lambda }\right)^{k-1}.} {\displaystyle h(x;k,\lambda )={k \over \lambda }\left({x \over \lambda }\right)^{k-1}.}

The Mean time between failures MTBF is

MTBF ( k , λ ) = λ Γ ( 1 + 1 / k ) . {\displaystyle {\text{MTBF}}(k,\lambda )=\lambda \Gamma (1+1/k).} {\displaystyle {\text{MTBF}}(k,\lambda )=\lambda \Gamma (1+1/k).}

Moments

The moment generating function of the logarithm of a Weibull distributed random variable is given by[13]

E [ e t log X ] = λ t Γ ( t k + 1 ) {\displaystyle \operatorname {E} \left[e^{t\log X}\right]=\lambda ^{t}\Gamma \left({\frac {t}{k}}+1\right)} {\displaystyle \operatorname {E} \left[e^{t\log X}\right]=\lambda ^{t}\Gamma \left({\frac {t}{k}}+1\right)}

where Γ is the gamma function. Similarly, the characteristic function of log X is given by

E [ e i t log X ] = λ i t Γ ( i t k + 1 ) . {\displaystyle \operatorname {E} \left[e^{it\log X}\right]=\lambda ^{it}\Gamma \left({\frac {it}{k}}+1\right).} {\displaystyle \operatorname {E} \left[e^{it\log X}\right]=\lambda ^{it}\Gamma \left({\frac {it}{k}}+1\right).}

In particular, the nth raw moment of X is given by

m n = λ n Γ ( 1 + n k ) . {\displaystyle m_{n}=\lambda ^{n}\Gamma \left(1+{\frac {n}{k}}\right).} {\displaystyle m_{n}=\lambda ^{n}\Gamma \left(1+{\frac {n}{k}}\right).}

The mean and variance of a Weibull random variable can be expressed as

E ( X ) = λ Γ ( 1 + 1 k ) {\displaystyle \operatorname {E} (X)=\lambda \Gamma \left(1+{\frac {1}{k}}\right)\,} {\displaystyle \operatorname {E} (X)=\lambda \Gamma \left(1+{\frac {1}{k}}\right)\,}

and

var ( X ) = λ 2 [ Γ ( 1 + 2 k ) ( Γ ( 1 + 1 k ) ) 2 ] . {\displaystyle \operatorname {var} (X)=\lambda ^{2}\left[\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}\right]\,.} {\displaystyle \operatorname {var} (X)=\lambda ^{2}\left[\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}\right]\,.}

The skewness is given by

γ 1 = 2 Γ 1 3 3 Γ 1 Γ 2 + Γ 3 [ Γ 2 Γ 1 2 ] 3 / 2 {\displaystyle \gamma _{1}={\frac {2\Gamma _{1}^{3}-3\Gamma _{1}\Gamma _{2}+\Gamma _{3}}{[\Gamma _{2}-\Gamma _{1}^{2}]^{3/2}}}} {\displaystyle \gamma _{1}={\frac {2\Gamma _{1}^{3}-3\Gamma _{1}\Gamma _{2}+\Gamma _{3}}{[\Gamma _{2}-\Gamma _{1}^{2}]^{3/2}}}}

where Γ i = Γ ( 1 + i / k ) {\displaystyle \Gamma _{i}=\Gamma (1+i/k)} {\displaystyle \Gamma _{i}=\Gamma (1+i/k)}, which may also be written as

γ 1 = Γ ( 1 + 3 k ) λ 3 3 μ σ 2 μ 3 σ 3 {\displaystyle \gamma _{1}={\frac {\Gamma \left(1+{\frac {3}{k}}\right)\lambda ^{3}-3\mu \sigma ^{2}-\mu ^{3}}{\sigma ^{3}}}} {\displaystyle \gamma _{1}={\frac {\Gamma \left(1+{\frac {3}{k}}\right)\lambda ^{3}-3\mu \sigma ^{2}-\mu ^{3}}{\sigma ^{3}}}}

where the mean is denoted by μ and the standard deviation is denoted by σ.

The excess kurtosis is given by

γ 2 = 6 Γ 1 4 + 12 Γ 1 2 Γ 2 3 Γ 2 2 4 Γ 1 Γ 3 + Γ 4 [ Γ 2 Γ 1 2 ] 2 {\displaystyle \gamma _{2}={\frac {-6\Gamma _{1}^{4}+12\Gamma _{1}^{2}\Gamma _{2}-3\Gamma _{2}^{2}-4\Gamma _{1}\Gamma _{3}+\Gamma _{4}}{[\Gamma _{2}-\Gamma _{1}^{2}]^{2}}}} {\displaystyle \gamma _{2}={\frac {-6\Gamma _{1}^{4}+12\Gamma _{1}^{2}\Gamma _{2}-3\Gamma _{2}^{2}-4\Gamma _{1}\Gamma _{3}+\Gamma _{4}}{[\Gamma _{2}-\Gamma _{1}^{2}]^{2}}}}

where Γ i = Γ ( 1 + i / k ) {\displaystyle \Gamma _{i}=\Gamma (1+i/k)} {\displaystyle \Gamma _{i}=\Gamma (1+i/k)}. The kurtosis excess may also be written as:

γ 2 = λ 4 Γ ( 1 + 4 k ) 4 γ 1 σ 3 μ 6 μ 2 σ 2 μ 4 σ 4 3. {\displaystyle \gamma _{2}={\frac {\lambda ^{4}\Gamma (1+{\frac {4}{k}})-4\gamma _{1}\sigma ^{3}\mu -6\mu ^{2}\sigma ^{2}-\mu ^{4}}{\sigma ^{4}}}-3.} {\displaystyle \gamma _{2}={\frac {\lambda ^{4}\Gamma (1+{\frac {4}{k}})-4\gamma _{1}\sigma ^{3}\mu -6\mu ^{2}\sigma ^{2}-\mu ^{4}}{\sigma ^{4}}}-3.}

Moment generating function

A variety of expressions are available for the moment generating function of X itself. As a power series, since the raw moments are already known, one has

E [ e t X ] = n = 0 t n λ n n ! Γ ( 1 + n k ) . {\displaystyle \operatorname {E} \left[e^{tX}\right]=\sum _{n=0}^{\infty }{\frac {t^{n}\lambda ^{n}}{n!}}\Gamma \left(1+{\frac {n}{k}}\right).} {\displaystyle \operatorname {E} \left[e^{tX}\right]=\sum _{n=0}^{\infty }{\frac {t^{n}\lambda ^{n}}{n!}}\Gamma \left(1+{\frac {n}{k}}\right).}

Alternatively, one can attempt to deal directly with the integral

E [ e t X ] = 0 e t x k λ ( x λ ) k 1 e ( x / λ ) k d x . {\displaystyle \operatorname {E} \left[e^{tX}\right]=\int _{0}^{\infty }e^{tx}{\frac {k}{\lambda }}\left({\frac {x}{\lambda }}\right)^{k-1}e^{-(x/\lambda )^{k}}\,dx.} {\displaystyle \operatorname {E} \left[e^{tX}\right]=\int _{0}^{\infty }e^{tx}{\frac {k}{\lambda }}\left({\frac {x}{\lambda }}\right)^{k-1}e^{-(x/\lambda )^{k}}\,dx.}

If the parameter k is assumed to be a rational number, expressed as k = p/q where p and q are integers, then this integral can be evaluated analytically.[a] With t replaced by −t, one finds

E [ e t X ] = 1 λ k t k p k q / p ( 2 π ) q + p 2 G p , q q , p ( 1 k p , 2 k p , , p k p 0 q , 1 q , , q 1 q | p p ( q λ k t k ) q ) {\displaystyle \operatorname {E} \left[e^{-tX}\right]={\frac {1}{\lambda ^{k}\,t^{k}}}\,{\frac {p^{k}\,{\sqrt {q/p}}}{({\sqrt {2\pi }})^{q+p-2}}}\,G_{p,q}^{\,q,p}\!\left(\left.{\begin{matrix}{\frac {1-k}{p}},{\frac {2-k}{p}},\dots ,{\frac {p-k}{p}}\\{\frac {0}{q}},{\frac {1}{q}},\dots ,{\frac {q-1}{q}}\end{matrix}}\;\right|\,{\frac {p^{p}}{\left(q\,\lambda ^{k}\,t^{k}\right)^{q}}}\right)} {\displaystyle \operatorname {E} \left[e^{-tX}\right]={\frac {1}{\lambda ^{k}\,t^{k}}}\,{\frac {p^{k}\,{\sqrt {q/p}}}{({\sqrt {2\pi }})^{q+p-2}}}\,G_{p,q}^{\,q,p}\!\left(\left.{\begin{matrix}{\frac {1-k}{p}},{\frac {2-k}{p}},\dots ,{\frac {p-k}{p}}\\{\frac {0}{q}},{\frac {1}{q}},\dots ,{\frac {q-1}{q}}\end{matrix}}\;\right|\,{\frac {p^{p}}{\left(q\,\lambda ^{k}\,t^{k}\right)^{q}}}\right)}

where G is the Meijer G-function.

The characteristic function has also been obtained by Muraleedharan et al. (2007)[16]

Minima

Let X 1 , X 2 , , X n {\displaystyle X_{1},X_{2},\ldots ,X_{n}} {\displaystyle X_{1},X_{2},\ldots ,X_{n}} be independent and identically distributed Weibull random variables with scale parameter λ {\displaystyle \lambda } {\displaystyle \lambda } and shape parameter k {\displaystyle k} {\displaystyle k}. If the minimum of these n {\displaystyle n} {\displaystyle n} random variables is Z = min ( X 1 , X 2 , , X n ) {\displaystyle Z=\min(X_{1},X_{2},\ldots ,X_{n})} {\displaystyle Z=\min(X_{1},X_{2},\ldots ,X_{n})}, then the cumulative probability distribution of Z {\displaystyle Z} {\displaystyle Z} is given by

F ( z ) = 1 e n ( z / λ ) k . {\displaystyle F(z)=1-e^{-n(z/\lambda )^{k}}.} {\displaystyle F(z)=1-e^{-n(z/\lambda )^{k}}.}

That is, Z {\displaystyle Z} {\displaystyle Z} will also be Weibull distributed with scale parameter n 1 / k λ {\displaystyle n^{-1/k}\lambda } {\displaystyle n^{-1/k}\lambda } and with shape parameter k {\displaystyle k} {\displaystyle k}.

Reparametrization tricks

Fix some α > 0 {\displaystyle \alpha >0} {\displaystyle \alpha >0}. Let ( π 1 , . . . , π n ) {\displaystyle (\pi _{1},...,\pi _{n})} {\displaystyle (\pi _{1},...,\pi _{n})} be nonnegative, and not all zero, and let g 1 , . . . , g n {\displaystyle g_{1},...,g_{n}} {\displaystyle g_{1},...,g_{n}} be independent samples of Weibull ( 1 , α 1 ) {\displaystyle {\text{Weibull}}(1,\alpha ^{-1})} {\displaystyle {\text{Weibull}}(1,\alpha ^{-1})}, then[17]

Shannon entropy

The information entropy is given by[18]

H ( λ , k ) = γ ( 1 1 k ) + ln ( λ k ) + 1 {\displaystyle H(\lambda ,k)=\gamma \left(1-{\frac {1}{k}}\right)+\ln \left({\frac {\lambda }{k}}\right)+1} {\displaystyle H(\lambda ,k)=\gamma \left(1-{\frac {1}{k}}\right)+\ln \left({\frac {\lambda }{k}}\right)+1}

where γ {\displaystyle \gamma } {\displaystyle \gamma } is the Euler–Mascheroni constant. The Weibull distribution is the maximum entropy distribution for a non-negative real random variate with a fixed expected value of xk equal to λk and a fixed expected value of ln(xk) equal to ln(λk) −  γ {\displaystyle \gamma } {\displaystyle \gamma }.

Kullback–Leibler divergence

The Kullback–Leibler divergence between two Weibull distributions is given by[19]

D KL ( W e i b 1 W e i b 2 ) = log k 1 λ 1 k 1 log k 2 λ 2 k 2 + ( k 1 k 2 ) [ log λ 1 γ k 1 ] + ( λ 1 λ 2 ) k 2 Γ ( k 2 k 1 + 1 ) 1 {\displaystyle D_{\text{KL}}(\mathrm {Weib} _{1}\parallel \mathrm {Weib} _{2})=\log {\frac {k_{1}}{\lambda _{1}^{k_{1}}}}-\log {\frac {k_{2}}{\lambda _{2}^{k_{2}}}}+(k_{1}-k_{2})\left[\log \lambda _{1}-{\frac {\gamma }{k_{1}}}\right]+\left({\frac {\lambda _{1}}{\lambda _{2}}}\right)^{k_{2}}\Gamma \left({\frac {k_{2}}{k_{1}}}+1\right)-1} {\displaystyle D_{\text{KL}}(\mathrm {Weib} _{1}\parallel \mathrm {Weib} _{2})=\log {\frac {k_{1}}{\lambda _{1}^{k_{1}}}}-\log {\frac {k_{2}}{\lambda _{2}^{k_{2}}}}+(k_{1}-k_{2})\left[\log \lambda _{1}-{\frac {\gamma }{k_{1}}}\right]+\left({\frac {\lambda _{1}}{\lambda _{2}}}\right)^{k_{2}}\Gamma \left({\frac {k_{2}}{k_{1}}}+1\right)-1}

Parameter estimation

Ordinary least square using Weibull plot

The fit of a Weibull distribution to data can be visually assessed using a Weibull plot.[20] The Weibull plot is a plot of the empirical cumulative distribution function F ^ ( x ) {\displaystyle {\widehat {F}}(x)} {\displaystyle {\widehat {F}}(x)} of data on special axes in a type of Q–Q plot. The axes are ln ( ln ( 1 F ^ ( x ) ) ) {\displaystyle \ln(-\ln(1-{\widehat {F}}(x)))} {\displaystyle \ln(-\ln(1-{\widehat {F}}(x)))} versus ln ( x ) {\displaystyle \ln(x)} {\displaystyle \ln(x)}. The reason for this change of variables is the cumulative distribution function can be linearized:

F ( x ) = 1 e ( x / λ ) k ln ( 1 F ( x ) ) = ( x / λ ) k ln ( ln ( 1 F ( x ) ) ) 'y' = k ln x 'mx' k ln λ 'c' {\displaystyle {\begin{aligned}F(x)&=1-e^{-(x/\lambda )^{k}}\\[4pt]-\ln(1-F(x))&=(x/\lambda )^{k}\\[4pt]\underbrace {\ln(-\ln(1-F(x)))} _{\textrm {'y'}}&=\underbrace {k\ln x} _{\textrm {'mx'}}-\underbrace {k\ln \lambda } _{\textrm {'c'}}\end{aligned}}} {\displaystyle {\begin{aligned}F(x)&=1-e^{-(x/\lambda )^{k}}\\[4pt]-\ln(1-F(x))&=(x/\lambda )^{k}\\[4pt]\underbrace {\ln(-\ln(1-F(x)))} _{\textrm {'y'}}&=\underbrace {k\ln x} _{\textrm {'mx'}}-\underbrace {k\ln \lambda } _{\textrm {'c'}}\end{aligned}}}

which can be seen to be in the standard form of a straight line. Therefore, if the data came from a Weibull distribution then a straight line is expected on a Weibull plot.

There are various approaches to obtaining the empirical distribution function from data. One method is to obtain the vertical coordinate for each point using

F ^ = i 0.3 n + 0.4 {\displaystyle {\widehat {F}}={\frac {i-0.3}{n+0.4}}} {\displaystyle {\widehat {F}}={\frac {i-0.3}{n+0.4}}},

where i {\displaystyle i} {\displaystyle i} is the rank of the data point and n {\displaystyle n} {\displaystyle n} is the number of data points.[21][22] Another common estimator[23] is

F ^ = i 0.5 n {\displaystyle {\widehat {F}}={\frac {i-0.5}{n}}} {\displaystyle {\widehat {F}}={\frac {i-0.5}{n}}}.

Linear regression can also be used to numerically assess goodness of fit and estimate the parameters of the Weibull distribution. The gradient informs one directly about the shape parameter k {\displaystyle k} {\displaystyle k} and the scale parameter λ {\displaystyle \lambda } {\displaystyle \lambda } can also be inferred.

Method of moments

The coefficient of variation of Weibull distribution depends only on the shape parameter:[24]

C V 2 = σ 2 μ 2 = Γ ( 1 + 2 k ) ( Γ ( 1 + 1 k ) ) 2 ( Γ ( 1 + 1 k ) ) 2 . {\displaystyle CV^{2}={\frac {\sigma ^{2}}{\mu ^{2}}}={\frac {\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}{\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}}.} {\displaystyle CV^{2}={\frac {\sigma ^{2}}{\mu ^{2}}}={\frac {\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}{\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}}.}

Equating the sample quantities s 2 / x ¯ 2 {\displaystyle s^{2}/{\bar {x}}^{2}} {\displaystyle s^{2}/{\bar {x}}^{2}} to σ 2 / μ 2 {\displaystyle \sigma ^{2}/\mu ^{2}} {\displaystyle \sigma ^{2}/\mu ^{2}}, the moment estimate of the shape parameter k {\displaystyle k} {\displaystyle k} can be read off either from a look up table or a graph of C V 2 {\displaystyle CV^{2}} {\displaystyle CV^{2}} versus k {\displaystyle k} {\displaystyle k}. A more accurate estimate of k ^ {\displaystyle {\hat {k}}} {\displaystyle {\hat {k}}} can be found using a root finding algorithm to solve

Γ ( 1 + 2 k ) ( Γ ( 1 + 1 k ) ) 2 ( Γ ( 1 + 1 k ) ) 2 = s 2 x ¯ 2 . {\displaystyle {\frac {\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}{\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}}={\frac {s^{2}}{{\bar {x}}^{2}}}.} {\displaystyle {\frac {\Gamma \left(1+{\frac {2}{k}}\right)-\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}{\left(\Gamma \left(1+{\frac {1}{k}}\right)\right)^{2}}}={\frac {s^{2}}{{\bar {x}}^{2}}}.}

The moment estimate of the scale parameter can then be found using the first moment equation as

λ ^ = x ¯ Γ ( 1 + 1 k ^ ) . {\displaystyle {\hat {\lambda }}={\frac {\bar {x}}{\Gamma \left(1+{\frac {1}{\hat {k}}}\right)}}.} {\displaystyle {\hat {\lambda }}={\frac {\bar {x}}{\Gamma \left(1+{\frac {1}{\hat {k}}}\right)}}.}

Maximum likelihood

The maximum likelihood estimator for the λ {\displaystyle \lambda } {\displaystyle \lambda } parameter given k {\displaystyle k} {\displaystyle k} is[24]

λ ^ = ( 1 n i = 1 n x i k ) 1 k {\displaystyle {\widehat {\lambda }}=\left({\frac {1}{n}}\sum _{i=1}^{n}x_{i}^{k}\right)^{\frac {1}{k}}} {\displaystyle {\widehat {\lambda }}=\left({\frac {1}{n}}\sum _{i=1}^{n}x_{i}^{k}\right)^{\frac {1}{k}}}

The maximum likelihood estimator for k {\displaystyle k} {\displaystyle k} is the solution for k of the following equation[25]

0 = i = 1 n x i k ln x i i = 1 n x i k 1 k 1 n i = 1 n ln x i {\displaystyle 0={\frac {\sum _{i=1}^{n}x_{i}^{k}\ln x_{i}}{\sum _{i=1}^{n}x_{i}^{k}}}-{\frac {1}{k}}-{\frac {1}{n}}\sum _{i=1}^{n}\ln x_{i}} {\displaystyle 0={\frac {\sum _{i=1}^{n}x_{i}^{k}\ln x_{i}}{\sum _{i=1}^{n}x_{i}^{k}}}-{\frac {1}{k}}-{\frac {1}{n}}\sum _{i=1}^{n}\ln x_{i}}

This equation defines k ^ {\displaystyle {\widehat {k}}} {\displaystyle {\widehat {k}}} only implicitly, one must generally solve for k {\displaystyle k} {\displaystyle k} by numerical means.

When x 1 > x 2 > > x N {\displaystyle x_{1}>x_{2}>\cdots >x_{N}} {\displaystyle x_{1}>x_{2}>\cdots >x_{N}} are the N {\displaystyle N} {\displaystyle N} largest observed samples from a dataset of more than N {\displaystyle N} {\displaystyle N} samples, then the maximum likelihood estimator for the λ {\displaystyle \lambda } {\displaystyle \lambda } parameter given k {\displaystyle k} {\displaystyle k} is[25]

λ ^ k = 1 N i = 1 N ( x i k x N k ) {\displaystyle {\widehat {\lambda }}^{k}={\frac {1}{N}}\sum _{i=1}^{N}(x_{i}^{k}-x_{N}^{k})} {\displaystyle {\widehat {\lambda }}^{k}={\frac {1}{N}}\sum _{i=1}^{N}(x_{i}^{k}-x_{N}^{k})}

Also given that condition, the maximum likelihood estimator for k {\displaystyle k} {\displaystyle k} is

0 = i = 1 N ( x i k ln x i x N k ln x N ) i = 1 N ( x i k x N k ) 1 N i = 1 N ln x i {\displaystyle 0={\frac {\sum _{i=1}^{N}(x_{i}^{k}\ln x_{i}-x_{N}^{k}\ln x_{N})}{\sum _{i=1}^{N}(x_{i}^{k}-x_{N}^{k})}}-{\frac {1}{N}}\sum _{i=1}^{N}\ln x_{i}} {\displaystyle 0={\frac {\sum _{i=1}^{N}(x_{i}^{k}\ln x_{i}-x_{N}^{k}\ln x_{N})}{\sum _{i=1}^{N}(x_{i}^{k}-x_{N}^{k})}}-{\frac {1}{N}}\sum _{i=1}^{N}\ln x_{i}}

Again, this being an implicit function, one must generally solve for k {\displaystyle k} {\displaystyle k} by numerical means.

Applications

The Weibull distribution is used

Prediction

See also

Notes

  1. See Cheng, Tellambura & Beaulieu (2004) [14] for the case when k is an integer, and Sagias & Karagiannidis (2005)[15] for the rational case.

References

  1. W. Weibull (1939). "The Statistical Theory of the Strength of Materials". Ingeniors Vetenskaps Academy Handlingar (151). Stockholm: Generalstabens Litografiska Anstalts Förlag: 1–45.
  2. Bowers, et. al. (1997) Actuarial Mathematics, 2nd ed. Society of Actuaries.
  3. Rosin, P.; Rammler, E. (1933). "The law governing the fineness of powdered coal". Journal of the Institute of Fuel. 7: 29–36, 109–122.
  4. Papoulis, Athanasios Papoulis; Pillai, S. Unnikrishna (2002). Probability, Random Variables, and Stochastic Processes (4th ed.). Boston: McGraw-Hill. ISBN 0-07-366011-6.
  5. Kizilersu, Ayse; Kreer, Markus; Thomas, Anthony W. (2018). "The Weibull distribution". Significance. 15 (2): 10–11. doi:10.1111/j.1740-9713.2018.01123.x.
  6. "Rayleigh Distribution – MATLAB & Simulink – MathWorks Australia". www.mathworks.com.au. Archived from the original on 12 October 2014.
  7. Jiang, R.; Murthy, D.N.P. (2011). "A study of Weibull shape parameter: Properties and significance". Reliability Engineering & System Safety. 96 (12): 1619–26. doi:10.1016/j.ress.2011.09.003.
  8. Eliazar, Iddo (November 2017). "Lindy's Law". Physica A: Statistical Mechanics and Its Applications. 486: 797–805. Bibcode:2017PhyA..486..797E. doi:10.1016/j.physa.2017.05.077. S2CID 125349686.
  9. Collett, David (2015). Modelling survival data in medical research (3rd ed.). Boca Raton: Chapman and Hall / CRC. ISBN 978-1439856789.
  10. Cameron, A. C.; Trivedi, P. K. (2005). Microeconometrics : methods and applications. Cambridge University Press. p. 584. ISBN 978-0-521-84805-3.
  11. Kalbfleisch, J. D.; Prentice, R. L. (2002). The statistical analysis of failure time data (2nd ed.). Hoboken, N.J.: J. Wiley. ISBN 978-0-471-36357-6. OCLC 50124320.
  12. Therneau, T. (2020). "A Package for Survival Analysis in R." R package version 3.1.
  13. Johnson, Kotz & Balakrishnan 1994 harvnb error: no target: CITEREFJohnsonKotzBalakrishnan1994 (help)
  14. Cheng, J.; Tellambura, C.; Beaulieu, N.C. (August 2004). "Performance of digital linear modulations on Weibull slow-fading channels". IEEE Transactions on Communications. 52 (8): 1265–1268. Bibcode:2004ITCom..52.1265C. doi:10.1109/TCOMM.2004.833015.
  15. Sagias, N.C.; Karagiannidis, G.K. (October 2005). "Gaussian class multivariate Weibull distributions: theory and applications in fading channels". IEEE Transactions on Information Theory. 51 (10): 3608–3619. Bibcode:2005ITIT...51.3608S. doi:10.1109/TIT.2005.855598.
  16. Muraleedharan, G.; Rao, A. D.; Kurup, P. G.; Nair, N. Unnikrishnan; Sinha, Mourani (1 August 2007). "Modified Weibull distribution for maximum and significant wave height simulation and prediction". Coastal Engineering. 54 (8): 630–638. Bibcode:2007CoasE..54..630M. doi:10.1016/j.coastaleng.2007.05.001. ISSN 0378-3839.
  17. Balog, Matej; Tripuraneni, Nilesh; Ghahramani, Zoubin; Weller, Adrian (17 July 2017). "Lost Relatives of the Gumbel Trick". International Conference on Machine Learning. PMLR: 371–379. arXiv:1706.04161.
  18. Cho, Youngseuk; Sun, Hokeun; Lee, Kyeongjun (5 January 2015). "Estimating the Entropy of a Weibull Distribution under Generalized Progressive Hybrid Censoring". Entropy. 17 (1): 102–122. Bibcode:2015Entrp..17..102C. doi:10.3390/e17010102. ISSN 1099-4300.
  19. Bauckhage, Christian (2013). "Computing the Kullback-Leibler Divergence between two Weibull Distributions". arXiv:1310.3713 [cs.IT].
  20. "1.3.3.30. Weibull Plot". www.itl.nist.gov.
  21. Wayne Nelson (2004) Applied Life Data Analysis. Wiley-Blackwell ISBN 0-471-64462-5
  22. Barnett, V. (1975). "Probability Plotting Methods and Order Statistics". Journal of the Royal Statistical Society. Series C (Applied Statistics). 24 (1): 95–108. doi:10.2307/2346708. ISSN 0035-9254. JSTOR 2346708.
  23. ISO 20501:2019 – Fine ceramics (advanced ceramics, advanced technical ceramics) – Weibull statistics for strength data.
  24. Cohen, A. Clifford (November 1965). "Maximum Likelihood Estimation in the Weibull Distribution Based on Complete and on Censored Samples" (PDF). Technometrics. 7 (4): 579–588. doi:10.1080/00401706.1965.10490300.
  25. Sornette, D. (2004). Critical Phenomena in Natural Science: Chaos, Fractals, Self-organization, and Disorder..
  26. Lee, Se Yoon; Mallick, Bani (2021). "Bayesian Hierarchical Modeling: Application Towards Production Results in the Eagle Ford Shale of South Texas". Sankhya B. 84: 1–43. doi:10.1007/s13571-020-00245-8.
  27. "Wind Speed Distribution Weibull – REUK.co.uk". www.reuk.co.uk.
  28. Liu, Chao; White, Ryen W.; Dumais, Susan (19 July 2010). Understanding web browsing behaviors through Weibull analysis of dwell time. ACM. pp. 379–386. doi:10.1145/1835449.1835513. ISBN 9781450301534. S2CID 12186028.
  29. Sharif, M.Nawaz; Islam, M.Nazrul (1980). "The Weibull distribution as a general model for forecasting technological change". Technological Forecasting and Social Change. 18 (3): 247–56. doi:10.1016/0040-1625(80)90026-8.
  30. Computational Optimization of Internal Combustion Engine page 49
  31. Austin, L. G.; Klimpel, R. R.; Luckie, P. T. (1984). Process Engineering of Size Reduction. Hoboken, NJ: Guinn Printing Inc. ISBN 0-89520-421-5.
  32. Chandrashekar, S. (1943). "Stochastic Problems in Physics and Astronomy". Reviews of Modern Physics. 15 (1): 86. Bibcode:1943RvMP...15....1C. doi:10.1103/RevModPhys.15.1.
  33. ECSS-E-ST-10-12C – Methods for the calculation of radiation received and its effects, and a policy for design margins (Report). European Cooperation for Space Standardization. 15 November 2008.
  34. L. D. Edmonds; C. E. Barnes; L. Z. Scheick (May 2000). "8.3 Curve Fitting". An Introduction to Space Radiation Effects on Microelectronics (PDF) (Report). NASA Jet Propulsion Laboratory, California Institute of Technology. pp. 75–76.
  35. Jewson, Stephen; Sweeting, Trevor; Jewson, Lynne (20 February 2025). "Reducing reliability bias in assessments of extreme weather risk using calibrating priors". Advances in Statistical Climatology, Meteorology and Oceanography. 11 (1): 1–22. Bibcode:2025ASCMO..11....1J. doi:10.5194/ascmo-11-1-2025. ISSN 2364-3579.
  36. Severini, Thomas A.; Mukerjee, Rahul; Ghosh, Malay (1 December 2002). "On an exact probability matching property of right-invariant priors". Biometrika. 89 (4): 952–957. doi:10.1093/biomet/89.4.952. ISSN 0006-3444.
  37. Gerrard, R.; Tsanakas, A. (2011). "Failure Probability Under Parameter Uncertainty". Risk Analysis. 31 (5): 727–744. Bibcode:2011RiskA..31..727G. doi:10.1111/j.1539-6924.2010.01549.x. ISSN 1539-6924. PMID 21175720.
  38. "System evolution and reliability of systems". Sysev (Belgium). 1 January 2010.
  39. Montgomery, Douglas (19 June 2012). Introduction to statistical quality control. [S.l.]: John Wiley. p. 95. ISBN 9781118146811.
  40. Chatfield, C.; Goodhardt, G.J. (1973). "A Consumer Purchasing Model with Erlang Interpurchase Times". Journal of the American Statistical Association. 68 (344): 828–835. doi:10.1080/01621459.1973.10481432.

Further reading