The Weibull Distribution

The Weibull distribution is one of the most widely used lifetime distributions in reliability engineering. It is a versatile distribution that can take on the characteristics of other types of distributions, based on the value of the shape parameter, $$ {\beta} \,\!$$. This chapter provides a brief background on the Weibull distribution, presents and derives most of the applicable equations and presents examples calculated both manually and by using ReliaSoft's Weibull++.

The 2-Parameter Weibull
The 2-parameter Weibull pdf is obtained by setting $$ \gamma=0 \,\!$$, and is given by:


 * $$ f(t)={ \frac{\beta }{\eta }}\left( {\frac{t}{\eta }}\right) ^{\beta -1}e^{-\left( { \frac{t}{\eta }}\right) ^{\beta }} \,\!$$

The 1-Parameter Weibull
The 1-parameter Weibull pdf is obtained by again setting $$\gamma=0 \,\!$$ and assuming $$\beta=C=Constant \,\!$$ assumed value or:

$$ f(t)={ \frac{C}{\eta }}\left( {\frac{t}{\eta }}\right) ^{C-1}e^{-\left( {\frac{t}{ \eta }}\right) ^{C}} \,\!$$

where the only unknown parameter is the scale parameter, $$\eta\,\!$$.

Note that in the formulation of the 1-parameter Weibull, we assume that the shape parameter $$\beta \,\!$$ is known a priori from past experience with identical or similar products. The advantage of doing this is that data sets with few or no failures can be analyzed.

Estimation of the Weibull Parameters
The estimates of the parameters of the Weibull distribution can be found graphically via probability plotting paper, or analytically, using either least squares (rank regression) or maximum likelihood estimation (MLE).

Probability Plotting
One method of calculating the parameters of the Weibull distribution is by using probability plotting. To better illustrate this procedure, consider the following example from Kececioglu [20].

Assume that six identical units are being reliability tested at the same application and operation stress levels. All of these units fail during the test after operating the following number of hours: 93, 34, 16, 120, 53 and 75. Estimate the values of the parameters for a 2-parameter Weibull distribution and determine the reliability of the units at a time of 15 hours.

Solution

The steps for determining the parameters of the Weibull representing the data, using probability plotting, are outlined in the following instructions. First, rank the times-to-failure in ascending order as shown next.

Obtain their median rank plotting positions. Median rank positions are used instead of other ranking methods because median ranks are at a specific confidence level (50%). Median ranks can be found tabulated in many reliability books. They can also be estimated using the following equation:


 * $$ MR \sim { \frac{i-0.3}{N+0.4}}\cdot 100 $$

where $$i$$ is the failure order number and $$N$$ is the total sample size. The exact median ranks are found in Weibull++ by solving:


 * $$\sum_{k=i}^N{\binom{N}{k}}{MR^k}{(1-MR)^{N-k}}=0.5=50%

$$

for $$MR$$, where $$N$$ is the sample size and $$i$$ the order number. The times-to-failure, with their corresponding median ranks, are shown next.

On a Weibull probability paper, plot the times and their corresponding ranks. A sample of a Weibull probability paper is given in the following figure.



The points of the data in the example are shown in the figure below. Draw the best possible straight line through these points, as shown below, then obtain the slope of this line by drawing a line, parallel to the one just obtained, through the slope indicator. This value is the estimate of the shape parameter $$ \hat{\beta } $$, in this case $$ \hat{\beta }=1.4 $$.



At the $$ Q(t)=63.2%\,\! $$ ordinate point, draw a straight horizontal line until this line intersects the fitted straight line. Draw a vertical line through this intersection until it crosses the abscissa. The value at the intersection of the abscissa is the estimate of $$ \hat{\eta } $$. For this case, $$ \hat{\eta }=76 $$ hours. This is always at 63.2% since:


 * $$ Q(t)=1-e^{-(\frac{t}{\eta })^{\beta }}=1-e^{-1}=0.632=63.2% $$

Now any reliability value for any mission time $$t$$ can be obtained. For example, the reliability for a mission of 15 hours, or any other time, can now be obtained either from the plot or analytically. To obtain the value from the plot, draw a vertical line from the abscissa, at hours, to the fitted line. Draw a horizontal line from this intersection to the ordinate and read $$ Q(t)\,\! $$, in this case $$ Q(t)=9.8% \,\! $$. Thus, $$ R(t)=1-Q(t)=90.2%\,\! $$. This can also be obtained analytically from the Weibull reliability function since the estimates of both of the parameters are known or:


 * $$ R(t=15)=e^{-\left( \frac{15}{\eta }\right) ^{\beta }}=e^{-\left( \frac{15}{76 }\right) ^{1.4}}=90.2% $$

Probability Plotting for the Location Parameter, Gamma
The third parameter of the Weibull distribution is utilized when the data do not fall on a straight line, but fall on either a concave up or down curve. The following statements can be made regarding the value of $$\gamma \,\!$$:


 * Case 1: If the curve for MR versus $${{t}_{j}}\,\!$$ is concave down and the curve for MR versus $${({t}_{j}-{t}_{1})}\,\!$$ is concave up, then there exists a $$\gamma \,\!$$ such that $$0< \gamma < t_{1}\,\!$$, or $$\gamma \,\!$$ has a positive value.


 * Case 2: If the curves for MR versus $${{t}_{j}}\,\!$$ and MR versus $${({t}_{j}-{t}_{1})}\,\!$$ are both concave up, then there exists a negative $$\gamma \,\!$$ which will straighten out the curve of MR versus $${{t}_{j}}\,\!$$.


 * Case 3: If neither one of the previous two cases prevails, then either reject the Weibull as one capable of representing the data, or proceed with the multiple population (mixed Weibull) analysis. To obtain the location parameter, $$\gamma \,\!$$:


 * Subtract the same arbitrary value, $$\gamma \,\!$$, from all the times to failure and replot the data.
 * If the initial curve is concave up, subtract a negative $$\gamma \,\!$$ from each failure time.
 * If the initial curve is concave down, subtract a positive $$\gamma \,\!$$ from each failure time.
 * Repeat until the data plots on an acceptable straight line.
 * The value of $$\gamma \,\!$$ is the subtracted (positive or negative) value that places the points in an acceptable straight line.

The other two parameters are then obtained using the techniques previously described. Also, it is important to note that we used the term subtract a positive or negative gamma, where subtracting a negative gamma is equivalent to adding it. Note that when adjusting for gamma, the x-axis scale for the straight line becomes $${({t}-\gamma)}\,\!$$.

Rank Regression on Y
Performing rank regression on Y requires that a straight line mathematically be fitted to a set of data points such that the sum of the squares of the vertical deviations from the points to the line is minimized. This is in essence the same methodology as the probability plotting method, except that we use the principle of least squares to determine the line through the points, as opposed to just eyeballing it. The first step is to bring our function into a linear form. For the two-parameter Weibull distribution, the (cumulative density function) is:


 * $$ F(t)=1-e^{-\left( \frac{t}{\eta }\right) ^{\beta }} $$

Taking the natural logarithm of both sides of the equation yields:


 * $$\ln[ 1-F(t)] =-( \frac{t}{\eta }) ^{\beta } $$


 * $$ \ln{ -\ln[ 1-F(t)]} =\beta \ln ( \frac{t}{ \eta }) $$

or:


 * $$\begin{align}

\ln \{ -\ln[ 1-F(t)]\} =-\beta \ln (\eta )+\beta \ln (t) \end{align}$$

Now let:


 * $$\begin{align}

y = \ln \{ -\ln[ 1-F(t)]\} \end{align}$$


 * $$\begin{align}

a = - ßln(\eta) \end{align}$$

and:


 * $$\begin{align}

b= \beta \end{align}$$

which results in the linear equation of:


 * $$\begin{align}

y=a+bx \end{align}$$

The least squares parameter estimation method (also known as regression analysis) was discussed in Parameter Estimation, and the following equations for regression on Y were derived:


 * $$ \hat{a}=\frac{\sum\limits_{i=1}^{N}y_{i}}{N}-\hat{b}\frac{ \sum\limits_{i=1}^{N}x_{i}}{N}=\bar{y}-\hat{b}\bar{x} $$

and:


 * $$ \hat{b}={\frac{\sum\limits_{i=1}^{N}x_{i}y_{i}-\frac{\sum \limits_{i=1}^{N}x_{i}\sum\limits_{i=1}^{N}y_{i}}{N}}{\sum \limits_{i=1}^{N}x_{i}^{2}-\frac{\left( \sum\limits_{i=1}^{N}x_{i}\right) ^{2}}{N}}} $$

In this case the equations for yi and xi are:


 * $$ y_{i}=\ln \left\{ -\ln [1-F(t_{i})]\right\} $$

and:
 * $$\begin{align}

x_{i}=ln(t_{i}) \end{align}$$

The $$ F(t_{i}) \,\! $$ values are estimated from the median ranks.

Once $$ \hat{a} $$ and $$ \hat{b} $$ are obtained, then $$ \hat{\beta } $$ and $$ \hat{\eta } $$ can easily be obtained from previous equations.

The Correlation Coefficient

The correlation coefficient is defined as follows:


 * $$ \rho ={\frac{\sigma _{xy}}{\sigma _{x}\sigma _{y}}} $$

where $$\sigma_{xy}\,\!$$ = covariance of $$x\,\!$$ and $$y\,\!$$, $$\sigma_{x}\,\!$$ = standard deviation of $$x\,\!$$, and $$\sigma_{y}\,\!$$ = standard deviation of $$y\,\!$$. The estimator of $$\rho\,\!$$ is the sample correlation coefficient, $$ \hat{\rho} $$, given by:


 * $$ \hat{\rho}=\frac{\sum\limits_{i=1}^{N}(x_{i}-\overline{x})(y_{i}-\overline{y} )}{\sqrt{\sum\limits_{i=1}^{N}(x_{i}-\overline{x})^{2}\cdot \sum\limits_{i=1}^{N}(y_{i}-\overline{y})^{2}}}$$

RRY Example
Consider the same data set from the probability plotting example given above (with six failures at 16, 34, 53, 75, 93 and 120 hours). Estimate the parameters and the correlation coefficient using rank regression on Y, assuming that the data follow the 2-parameter Weibull distribution.

Solution

Construct a table as shown next.

Utilizing the values from the table, calculate $$ \hat{a} $$ and $$ \hat{b} $$ using the following equations:
 * $$ \hat{b} =\frac{\sum\limits_{i=1}^{6}(\ln t_{i})y_{i}-(\sum\limits_{i=1}^{6}\ln t_{i})(\sum\limits_{i=1}^{6}y_{i})/6}{ \sum\limits_{i=1}^{6}(\ln t_{i})^{2}-(\sum\limits_{i=1}^{6}\ln t_{i})^{2}/6}

$$


 * $$ \hat{b}=\frac{-8.0699-(23.9068)(-3.0070)/6}{97.9909-(23.9068)^{2}/6} $$

or:


 * $$ \hat{b}=1.4301 $$

and:


 * $$ \hat{a}=\overline{y}-\hat{b}\overline{T}=\frac{\sum \limits_{i=1}^{N}y_{i}}{N}-\hat{b}\frac{\sum\limits_{i=1}^{N}\ln t_{i}}{N } $$

or:


 * $$ \hat{a}=\frac{(-3.0070)}{6}-(1.4301)\frac{23.9068}{6}=-6.19935 $$

Therefore:


 * $$ \hat{\beta }=\hat{b}=1.4301 $$

and:


 * $$ \hat{\eta }=e^{-\frac{\hat{a}}{\hat{b}}}=e^{-\frac{(-6.19935)}{ 1.4301}} $$

or:


 * $$ \hat{\eta }=76.318\text{ hr} $$

The correlation coefficient can be estimated as:


 * $$ \hat{\rho }=0.9956 $$

This example can be repeated in the Weibull++ software. The following plot shows the Weibull probability plot for the data set (with 90% two-sided confidence bounds).



If desired, the Weibull $$pdf$$ representing the data set can be written as:


 * $$ f(t)={\frac{\beta }{\eta }}\left( {\frac{t}{\eta }}\right) ^{\beta -1}e^{-\left( {\frac{t}{\eta }}\right) ^{\beta }} $$

or:


 * $$ f(t)={\frac{1.4302}{76.317}}\left( {\frac{t}{76.317}}\right) ^{0.4302}e^{-\left( {\frac{t}{76.317}}\right) ^{1.4302}} $$

You can also plot this result in Weibull++, as shown next. From this point on, different results, reports and plots can be obtained.



Rank Regression on X
Performing a rank regression on X is similar to the process for rank regression on Y, with the difference being that the horizontal deviations from the points to the line are minimized rather than the vertical. Again, the first task is to bring the reliability function into a linear form. This step is exactly the same as in the regression on Y analysis and all the equations apply in this case too. The derivation from the previous analysis begins on the least squares fit part, where in this case we treat as the dependent variable and as the independent variable. The best-fitting straight line to the data, for regression on X (see Parameter Estimation), is the straight line:


 * $$ x= \hat{a}+\hat{b}y $$

The corresponding equations for $$ \hat{a} $$ and $$ \hat{b} $$ are:


 * $$ \hat{a}=\overline{x}-\hat{b}\overline{y}=\frac{\sum\limits_{i=1}^{N}x_{i}}{N} -\hat{b}\frac{\sum\limits_{i=1}^{N}y_{i}}{N} $$

and:


 * $$ \hat{b}={\frac{\sum\limits_{i=1}^{N}x_{i}y_{i}-\frac{\sum \limits_{i=1}^{N}x_{i}\sum\limits_{i=1}^{N}y_{i}}{N}}{\sum \limits_{i=1}^{N}y_{i}^{2}-\frac{\left( \sum\limits_{i=1}^{N}y_{i}\right) ^{2}}{N}}} $$

where:


 * $$ y_{i}=\ln \left\{ -\ln [1-F(t_{i})]\right\} $$

and:


 * $$\begin{align}

x_{i}=\ln (t_{i}) \end{align}$$

and the F(ti) values are again obtained from the median ranks.

Once $$ \hat{a} $$ and $$ \hat{b} $$ are obtained, solve the linear equation for y, which corresponds to:


 * $$ y=-\frac{\hat{a}}{\hat{b}}+\frac{1}{\hat{b}}x $$ Solving for the parameters from above equations, we get:


 * $$ a=-\frac{\hat{a}}{\hat{b}}=-\beta \ln (\eta )$$

and


 * $$ b=\frac{1}{\hat{b}}=\beta$$

The correlation coefficient is evaluated as before.

RRX Example
Again using the same data set from the probability plotting and RRY examples (with six failures at 16, 34, 53, 75, 93 and 120 hours), calculate the parameters using rank regression on X.

Solution

The same table constructed above for the RRY example can also be applied for RRX.

Using the values from this table we get:


 * $$ \hat{b} ={\frac{\sum\limits_{i=1}^{6}(\ln T_{i})y_{i}-\frac{ \sum\limits_{i=1}^{6}\ln T_{i}\sum\limits_{i=1}^{6}y_{i}}{6}}{ \sum\limits_{i=1}^{6}y_{i}^{2}-\frac{\left( \sum\limits_{i=1}^{6}y_{i}\right) ^{2}}{6}}}

$$


 * $$\hat{b} =\frac{-8.0699-(23.9068)(-3.0070)/6}{7.1502-(-3.0070)^{2}/6} $$

or:


 * $$ \hat{b}=0.6931 $$

and:


 * $$ \hat{a}=\overline{x}-\hat{b}\overline{y}=\frac{\sum\limits_{i=1}^{6}\ln T_{i} }{6}-\hat{b}\frac{\sum\limits_{i=1}^{6}y_{i}}{6} $$

or:


 * $$ \hat{a}=\frac{23.9068}{6}-(0.6931)\frac{(-3.0070)}{6}=4.3318 $$

Therefore:


 * $$ \hat{\beta }=\frac{1}{\hat{b}}=\frac{1}{0.6931}=1.4428 $$

and:


 * $$ \hat{\eta }=e^{\frac{\hat{a}}{\hat{b}}\cdot \frac{1}{\hat{ \beta }}}=e^{\frac{4.3318}{0.6931}\cdot \frac{1}{1.4428}}=76.0811\text{ hr} $$

The correlation coefficient is:


 * $$ \hat{\rho }=0.9956 $$

The results and the associated graph using Weibull++ are shown next. Note that the slight variation in the results is due to the number of significant figures used in the estimation of the median ranks. Weibull++ by default uses double precision accuracy when computing the median ranks.



3-Parameter Weibull Regression
When the MR versus tj points plotted on the Weibull probability paper do not fall on a satisfactory straight line and the points fall on a curve, then a location parameter, $$\gamma\,\!$$, might exist which may straighten out these points. The goal in this case is to fit a curve, instead of a line, through the data points using nonlinear regression. The Gauss-Newton method can be used to solve for the parameters, $$\beta\,\!$$>, $$\eta\,\!$$ and $$\gamma\,\!$$, by performing a Taylor series expansion on $$F(t{_{i}};\beta ,\eta, \gamma )\,\!$$. Then the nonlinear model is approximated with linear terms and ordinary least squares are employed to estimate the parameters. This procedure is iterated until a satisfactory solution is reached.

(Note that other shapes, particularly shapes, might suggest the existence of more than one population. In these cases, the multiple population mixed Weibull distribution, may be more appropriate.)

When you use the 3-parameter Weibull distribution, Weibull++ calculates the value of $$\gamma\,\!$$ by utilizing an optimized Nelder-Mead algorithm and adjusts the points by this value of $$\gamma\,\!$$ such that they fall on a straight line, and then plots both the adjusted and the original unadjusted points. To draw a curve through the original unadjusted points, if so desired, select Weibull 3P Line Unadjusted for Gamma from the Show Plot Line submenu under the Plot Options menu. The returned estimations of the parameters are the same when selecting RRX or RRY. To display the unadjusted data points and line along with the adjusted data points and line, select Show/Hide Items under the Plot Options menu and include the unadjusted data points and line as follows:





The results and the associated graph for the previous example using the 3-parameter Weibull case are shown next:



Maximum Likelihood Estimation
As outlined in Parameter Estimation, maximum likelihood estimation works by developing a likelihood function based on the available data and finding the values of the parameter estimates that maximize the likelihood function. This can be achieved by using iterative methods to determine the parameter estimate values that maximize the likelihood function, but this can be rather difficult and time-consuming, particularly when dealing with the three-parameter distribution. Another method of finding the parameter estimates involves taking the partial derivatives of the likelihood function with respect to the parameters, setting the resulting equations equal to zero and solving simultaneously to determine the values of the parameter estimates. ( Note that MLE asymptotic properties do not hold when estimating $$\gamma\,\!$$ using MLE [27].) The log-likelihood functions and associated partial derivatives used to determine maximum likelihood estimates for the Weibull distribution are covered in Appendix D.

MLE Example
One last time, use the same data set from the probability plotting, RRY and RRX examples (with six failures at 16, 34, 53, 75, 93 and 120 hours) and calculate the parameters using MLE.

Solution

In this case, we have non-grouped data with no suspensions or intervals, (i.e., complete data). The equations for the partial derivatives of the log-likelihood function are derived in an appendix and given next:
 * $$ \frac{\partial \Lambda }{\partial \beta }=\frac{6}{\beta } +\sum_{i=1}^{6}\ln \left( \frac{T_{i}}{\eta }\right) -\sum_{i=1}^{6}\left( \frac{T_{i}}{\eta }\right) ^{\beta }\ln \left( \frac{T_{i}}{\eta }\right) =0

$$

And:


 * $$ \frac{\partial \Lambda }{\partial \eta }=\frac{-\beta }{\eta }\cdot 6+\frac{ \beta }{\eta }\sum\limits_{i=1}^{6}\left( \frac{T_{i}}{\eta }\right) ^{\beta }=0 $$

Solving the above equations simultaneously we get:


 * $$ \hat{\beta }=1.933,$$ $$\hat{\eta }=73.526 $$

The variance/covariance matrix is found to be:


 * $$ \left[ \begin{array}{ccc} \hat{Var}\left( \hat{\beta }\right) =0.4211 & \hat{Cov}( \hat{\beta },\hat{\eta })=3.272 \\

\hat{Cov}(\hat{\beta },\hat{\eta })=3.272 & \hat{Var} \left( \hat{\eta }\right) =266.646 \end{array} \right] $$

The results and the associated plot using Weibull++ (MLE) are shown next.



You can view the variance/covariance matrix directly by clicking the Analysis Summary table in the control panel. Note that the decimal accuracy displayed and used is based on your individual Application Setup.



Fisher Matrix Confidence Bounds
One of the methods used by the application in estimating the different types of confidence bounds for Weibull data, the Fisher matrix method, is presented in this section. The complete derivations were presented in detail (for a general function) in Confidence Bounds.

Bounds on the Parameters
One of the properties of maximum likelihood estimators is that they are asymptotically normal, meaning that for large samples they are normally distributed. Additionally, since both the shape parameter estimate, $$ \hat{\beta } $$, and the scale parameter estimate, $$ \hat{\eta }, $$ must be positive, thus $$ln\beta \,\!$$ and $$ln\eta \,\!$$ are treated as being normally distributed as well. The lower and upper bounds on the parameters are estimated from [30]:


 * $$ \beta _{U} =\hat{\beta }\cdot e^{\frac{K_{\alpha }\sqrt{Var(\hat{ \beta })}}{\hat{\beta }}}\text{ (upper bound)} $$


 * $$ \beta _{L} =\frac{\hat{\beta }}{e^{\frac{K_{\alpha }\sqrt{Var(\hat{ \beta })}}{\hat{\beta }}}} \text{ (lower bound)}

$$

and:


 * $$ \eta _{U} =\hat{\eta }\cdot e^{\frac{K_{\alpha }\sqrt{Var(\hat{ \eta })}}{\hat{\eta }}}\text{ (upper bound)}

$$


 * $$ \eta _{L} =\frac{\hat{\eta }}{e^{\frac{K_{\alpha }\sqrt{Var(\hat{ \eta })}}{\hat{\eta }}}}\text{ (lower bound)} $$

where $$ K_{\alpha}\,\!$$ is defined by:


 * $$ \alpha =\frac{1}{\sqrt{2\pi }}\int_{K_{\alpha }}^{\infty }e^{-\frac{t^{2}}{2} }dt=1-\Phi (K_{\alpha }) $$

If d is the confidence level, then $$ \alpha =\frac{1-\delta }{2} $$ for the two-sided bounds and a = 1 - d for the one-sided bounds. The variances and covariances of $$ \hat{\beta } $$ and $$ \hat{\eta } $$ are estimated from the inverse local Fisher matrix, as follows:
 * $$ \left( \begin{array}{cc} \hat{Var}\left( \hat{\beta }\right) & \hat{Cov}\left( \hat{ \beta },\hat{\eta }\right)

\\ \hat{Cov}\left( \hat{\beta },\hat{\eta }\right) & \hat{Var} \left( \hat{\eta }\right) \end{array} \right) =\left( \begin{array}{cc} -\frac{\partial ^{2}\Lambda }{\partial \beta ^{2}} & -\frac{\partial ^{2}\Lambda }{\partial \beta \partial \eta } \\

-\frac{\partial ^{2}\Lambda }{\partial \beta \partial \eta } & -\frac{ \partial ^{2}\Lambda }{\partial \eta ^{2}} \end{array} \right) _{\beta =\hat{\beta },\text{ }\eta =\hat{\eta }}^{-1} $$

Fisher Matrix Confidence Bounds and Regression Analysis

Note that the variance and covariance of the parameters are obtained from the inverse Fisher information matrix as described in this section. The local Fisher information matrix is obtained from the second partials of the likelihood function, by substituting the solved parameter estimates into the particular functions. This method is based on maximum likelihood theory and is derived from the fact that the parameter estimates were computed using maximum likelihood estimation methods. When one uses least squares or regression analysis for the parameter estimates, this methodology is theoretically then not applicable. However, if one assumes that the variance and covariance of the parameters will be similar ( One also assumes similar properties for both estimators.) regardless of the underlying solution method, then the above methodology can also be used in regression analysis.

The Fisher matrix is one of the methodologies that Weibull++ uses for both MLE and regression analysis. Specifically, Weibull++ uses the likelihood function and computes the local Fisher information matrix based on the estimates of the parameters and the current data. This gives consistent confidence bounds regardless of the underlying method of solution, (i.e., MLE or regression). In addition, Weibull++ checks this assumption and proceeds with it if it considers it to be acceptable. In some instances, Weibull++ will prompt you with an "Unable to Compute Confidence Bounds" message when using regression analysis. This is an indication that these assumptions were violated.

Bounds on Reliability
The bounds on reliability can easily be derived by first looking at the general extreme value distribution (EVD). Its reliability function is given by:


 * $$ R(t)=e^{-e^{\left( \frac{t-p_{1}}{p_{2}}\right) }} $$

By transforming t = lnt and converting $$ p=\ln({\eta})$$, $$ p_{2}=\frac{1}{ \beta } $$, the above equation becomes the Weibull reliability function:


 * $$ R(t)=e^{-e^{\beta \left( \ln t-\ln \eta \right) }}=e^{-e^{\ln \left( \frac{t }{\eta }\right) ^{\beta }}}=e^{-\left( \frac{t}{\eta }\right) ^{\beta }} $$

with:


 * $$ R(T)=e^{-e^{\beta \left( \ln t-\ln \eta \right) }}$$

set:


 * $$ u=\beta \left( \ln t-\ln \eta \right) $$

The reliability function now becomes:


 * $$ R(T)=e^{-e^{u}} $$

The next step is to find the upper and lower bounds on u. Using the equations derived in Confidence Bounds, the bounds on are then estimated from [30]:


 * $$ u_{U} =\hat{u}+K_{\alpha }\sqrt{Var(\hat{u})}

$$


 * $$ u_{L} =\hat{u}-K_{\alpha }\sqrt{Var(\hat{u})}

$$

where:


 * $$ Var(\hat{u}) =\left( \frac{\partial u}{\partial \beta }\right) ^{2}Var( \hat{\beta })+\left( \frac{\partial u}{\partial \eta }\right) ^{2}Var( \hat{\eta }) +2\left( \frac{\partial u}{\partial \beta }\right) \left( \frac{\partial u }{\partial \eta }\right) Cov\left( \hat{\beta },\hat{\eta }\right) $$

or:


 * $$ Var(\hat{u}) =\frac{\hat{u}^{2}}{\hat{\beta }^{2}}Var(\hat{ \beta })+\frac{\hat{\beta }^{2}}{\hat{\eta }^{2}}Var(\hat{\eta }) -\left( \frac{2u}{\hat{\eta }}\right) Cov\left( \hat{\beta }, \hat{\eta }\right). $$

The upper and lower bounds on reliability are:


 * $$ R_{U} =e^{-e^{u_{L}}}\text{ (upper bound)}$$


 * $$ R_{L} =e^{-e^{u_{U}}}\text{ (lower bound)}$$

Other Weibull Forms

Weibull++ makes the following assumptions/substitutions when using the three-parameter or one-parameter forms:


 * For the 3-parameter case, substitute $$ t=\ln (t-\hat{\gamma }) $$ (and by definition $$\gamma\,\!$$ &lt; t ), instead of lnt. (Note that this is an approximation since it eliminates the third parameter and assumes that $$ Var( \hat{\gamma })=0. $$)
 * For the 1-parameter, $$ Var(\hat{\beta })=0, $$ thus:


 * $$ Var(\hat{u})=\left( \frac{\partial u}{\partial \eta }\right) ^{2}Var( \hat{\eta })=\left( \frac{\hat{\beta }}{\hat{\eta }}\right) ^{2}Var(\hat{\eta }) $$

Also note that the time axis (x-axis) in the three-parameter Weibull plot in Weibull++ is not but t - $$\gamma\,\!$$. This means that one must be cautious when obtaining confidence bounds from the plot. If one desires to estimate the confidence bounds on reliability for a given time t0 from the adjusted plotted line, then these bounds should be obtained for a t0 - $$\gamma\,\!$$ entry on the time axis.

Bounds on Time
The bounds around the time estimate or reliable life estimate, for a given Weibull percentile (unreliability), are estimated by first solving the reliability equation with respect to time, as follows [24, 30]:


 * $$ \ln R =-\left( \frac{t}{\eta }\right) ^{\beta }

$$


 * $$ \ln (-\ln R) =\beta \ln \left( \frac{t}{\eta }\right) $$


 * $$\begin{align}

\ln (-\ln R) =\beta (\ln t-\ln \eta ) \end{align}$$

or:


 * $$ u=\frac{1}{\beta }\ln (-\ln R)+\ln \eta $$

where u = lnt.

The upper and lower bounds on are estimated from:


 * $$ u_{U} =\hat{u}+K_{\alpha }\sqrt{Var(\hat{u})} $$


 * $$ u_{L} =\hat{u}-K_{\alpha }\sqrt{Var(\hat{u})} $$

where:


 * $$ Var(\hat{u})=\left( \frac{\partial u}{\partial \beta }\right) ^{2}Var( \hat{\beta })+\left( \frac{\partial u}{\partial \eta }\right) ^{2}Var( \hat{\eta })+2\left( \frac{\partial u}{\partial \beta }\right) \left( \frac{\partial u}{\partial \eta }\right) Cov\left( \hat{\beta },\hat{ \eta }\right) $$

or:
 * $$ Var(\hat{u}) =\frac{1}{\hat{\beta }^{4}}\left[ \ln (-\ln R)\right] ^{2}Var(\hat{\beta })+\frac{1}{\hat{\eta }^{2}}Var(\hat{\eta })+2\left( -\frac{1}{\hat{\beta }^{2}}\right) \left( \frac{\ln (-\ln R)}{ \hat{\eta }}\right) Cov\left( \hat{\beta },\hat{\eta }\right) $$

The upper and lower bounds are then found by:


 * $$ T_{U} =e^{u_{U}}\text{ (upper bound)} $$


 * $$ T_{L} =e^{u_{L}}\text{ (lower bound)} $$

Likelihood Ratio Confidence Bounds
As covered in Confidence Bounds, the likelihood confidence bounds are calculated by finding values for ?1 and ?2 that satisfy:


 * $$ -2\cdot \text{ln}\left( \frac{L(\theta _{1},\theta _{2})}{L(\hat{\theta }_{1}, \hat{\theta }_{2})}\right) =\chi _{\alpha ;1}^{2} $$

This equation can be rewritten as:


 * $$ L(\theta _{1},\theta _{2})=L(\hat{\theta }_{1},\hat{\theta } _{2})\cdot e^{\frac{-\chi _{\alpha ;1}^{2}}{2}} $$

For complete data, the likelihood function for the Weibull distribution is given by:
 * $$ L(\beta ,\eta )=\prod_{i=1}^{N}f(x_{i};\beta ,\eta )=\prod_{i=1}^{N}\frac{ \beta }{\eta }\cdot \left( \frac{x_{i}}{\eta }\right) ^{\beta -1}\cdot e^{-\left( \frac{x_{i}}{\eta }\right) ^{\beta }} $$

For a given value of a, values for ß and ? can be found which represent the maximum and minimum values that satisfy the above equation. These represent the confidence bounds for the parameters at a confidence level d, where a = d for two-sided bounds and a = 2d - 1 for one-sided.

Similarly, the bounds on time and reliability can be found by substituting the Weibull reliability equation into the likelihood function so that it is in terms of ß and time or reliability, as discussed in Confidence Bounds. The likelihood ratio equation used to solve for bounds on time (Type 1) is:


 * $$ L(\beta ,t)=\prod_{i=1}^{N}\frac{\beta }{\left( \frac{t}{(-\text{ln}(R))^{ \frac{1}{\beta }}}\right) }\cdot \left( \frac{x_{i}}{\left( \frac{t}{(-\text{ ln}(R))^{\frac{1}{\beta }}}\right) }\right) ^{\beta -1}\cdot \text{exp}\left[ -\left( \frac{x_{i}}{\left( \frac{t}{(-\text{ln}(R))^{\frac{1}{\beta }}} \right) }\right) ^{\beta }\right] $$

The likelihood ratio equation used to solve for bounds on reliability (Type 2) is:


 * $$ L(\beta ,R)=\prod_{i=1}^{N}\frac{\beta }{\left( \frac{t}{(-\text{ln}(R))^{ \frac{1}{\beta }}}\right) }\cdot \left( \frac{x_{i}}{\left( \frac{t}{(-\text{ ln}(R))^{\frac{1}{\beta }}}\right) }\right) ^{\beta -1}\cdot \text{exp}\left[ -\left( \frac{x_{i}}{\left( \frac{t}{(-\text{ln}(R))^{\frac{1}{\beta }}} \right) }\right) ^{\beta }\right] $$

Bounds on Parameters
Bayesian Bounds use non-informative prior distributions for both parameters. From Confidence Bounds, we know that if the prior distribution of ? and ß are independent, the posterior joint distribution of ? and ß can be written as:


 * $$ f(\eta ,\beta |Data)= \dfrac{L(Data|\eta ,\beta )\varphi (\eta )\varphi (\beta )}{\int_{0}^{\infty }\int_{0}^{\infty }L(Data|\eta ,\beta )\varphi (\eta )\varphi (\beta )d\eta d\beta } $$

The marginal distribution of ? is:


 * $$ f(\eta |Data) =\int_{0}^{\infty }f(\eta ,\beta |Data)d\beta =

\dfrac{\int_{0}^{\infty }L(Data|\eta ,\beta )\varphi (\eta )\varphi (\beta )d\beta }{\int_{0}^{\infty }\int_{-\infty }^{\infty }L(Data|\eta ,\beta )\varphi (\eta )\varphi (\beta )d\eta d\beta } $$

where: $$ \varphi (\beta )=\frac{1}{\beta } $$ is the non-informative prior of ß. $$ \varphi (\eta )=\frac{1}{\eta } $$ is the non-informative prior of ? . Using these non-informative prior distributions, $$f(\eta|Data)$$ can be rewritten as:


 * $$ f(\eta |Data)=\dfrac{\int_{0}^{\infty }L(Data|\eta ,\beta )\frac{1}{\beta } \frac{1}{\eta }d\beta }{\int_{0}^{\infty }\int_{0}^{\infty }L(Data|\eta ,\beta )\frac{1}{\beta }\frac{1}{\eta }d\eta d\beta } $$

The one-sided upper bounds of ? is:


 * $$ CL=P(\eta \leq \eta _{U})=\int_{0}^{\eta _{U}}f(\eta |Data)d\eta $$

The one-sided lower bounds of ? is:


 * $$ 1-CL=P(\eta \leq \eta _{L})=\int_{0}^{\eta _{L}}f(\eta |Data)d\eta $$

The two-sided bounds of ? is:


 * $$ CL=P(\eta _{L}\leq \eta \leq \eta _{U})=\int_{\eta _{L}}^{\eta _{U}}f(\eta |Data)d\eta $$

Same method is used to obtain the bounds of ß.

Bounds on Reliability

 * $$ CL=\Pr (R\leq R_{U})=\Pr (\eta \leq T\exp (-\frac{\ln (-\ln R_{U})}{\beta })) $$

From the posterior distribution of ?, we have:


 * $$ CL=\dfrac{\int\nolimits_{0}^{\infty }\int\nolimits_{0}^{T\exp (-\dfrac{\ln (-\ln R_{U})}{\beta })}L(\beta ,\eta )\frac{1}{\beta }\frac{1}{\eta }d\eta d\beta }{\int\nolimits_{0}^{\infty }\int\nolimits_{0}^{\infty }L(\beta ,\eta )\frac{1}{\beta }\frac{1}{\eta }d\eta d\beta } $$

The above equation is solved numerically for RU. The same method can be used to calculate the one sided lower bounds and two-sided bounds on reliability.

Bounds on Time
From Confidence Bounds, we know that:


 * $$ CL=\Pr (T\leq T_{U})=\Pr (\eta \leq T_{U}\exp (-\frac{\ln (-\ln R)}{\beta })) $$

From the posterior distribution of ? , we have:


 * $$ CL=\dfrac{\int\nolimits_{0}^{\infty }\int\nolimits_{0}^{T_{U}\exp (-\dfrac{ \ln (-\ln R)}{\beta })}L(\beta ,\eta )\frac{1}{\beta }\frac{1}{\eta }d\eta d\beta }{\int\nolimits_{0}^{\infty }\int\nolimits_{0}^{\infty }L(\beta ,\eta )\frac{1}{\beta }\frac{1}{\eta }d\eta d\beta } $$

The above equation is solved numerically for TU. The same method can be applied to calculate one sided lower bounds and two-sided bounds on time.