Template:Maximum Likelihood Parameter Estimation: Difference between revisions

From ReliaWiki
Jump to navigation Jump to search
Line 114: Line 114:
*[[The Loglogistic Log-Likelihood Functions and their Partials]]
*[[The Loglogistic Log-Likelihood Functions and their Partials]]
*[[The Gumbel Log-Likelihood Functions and their Partials]]
*[[The Gumbel Log-Likelihood Functions and their Partials]]
 
*[[The Gamma Log-Likelihood Functions and their Partials]]
 
 
 
 
 
 
 
 
===  The Gamma Log-Likelihood Functions and their Partials===
This log-likelihood function is composed of three summation portions:
 
::<math>\begin{align}
  \ln (L)= & \Lambda =\underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}\ln \left( \frac{{{e}^{k(\ln ({{T}_{i}})-\mu )-{{e}^{{{e}^{\ln ({{T}_{i}})-\mu }}}}}}}{{{T}_{i}}\Gamma (k)} \right) \\
  & +\underset{i=1}{\mathop{\overset{S}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime }}\ln \left( 1-\Gamma \left( _{1}k;{{e}^{\ln (T_{i}^{^{\prime }})-\mu )}} \right) \right) \\
  & +\underset{i=1}{\mathop{\overset{{{F}_{I}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime \prime }}\ln \left( {{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }} \right)-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }} \right) \right) 
\end{align}</math>
 
:or:
 
::<math>\begin{align}
  \Lambda = & \underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{-\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}\ln ({{T}_{i}})\underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{-\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}\ln (\Gamma (k))+k\underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}(\ln ({{T}_{i}})-\mu ) \\
  & \underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{-\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}{{e}^{\ln ({{T}_{i}})-\mu }} \\
  & +\underset{i=1}{\mathop{\overset{S}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime }}\ln \left( 1-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{i}^{^{\prime }})-\mu }} \right) \right) \\
  & +\underset{i=1}{\mathop{\overset{{{F}_{I}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime \prime }}\ln \left( {{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu )}} \right)-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu )}} \right) \right) 
\end{align}</math>
 
 
:where:
::• <math>{{F}_{e}}</math> is the number of groups of times-to-failure data points
::• <math>{{N}_{i}}</math> is the number of times-to-failure in the <math>{{i}^{th}}</math> time-to-failure data group
::• <math>\mu </math> is the gamma shape parameter (unknown a priori, the first of two parameters to be found)
::• <math>k</math> is the gamma scale parameter (unknown a priori, the second of two parameters to be found)
::• <math>{{T}_{i}}</math> is the time of the <math>{{i}^{th}}</math> group of time-to-failure data
::• <math>S</math> is the number of groups of suspension data points
::• .. is the number of suspensions in <math>{{i}^{th}}</math> group of suspension data points
::• <math>T_{i}^{\prime }</math> is the time of the <math>{{i}^{th}}</math> suspension data group
::• <math>FI</math> is the number of interval failure data groups
::• <math>N_{i}^{\prime \prime }</math> is the number of intervals in <math>{{i}^{th}}</math> group of data intervals
::• <math>T_{Li}^{\prime \prime }</math> is the beginning of the <math>{{i}^{th}}</math> interval
::• and <math>T_{Ri}^{\prime \prime }</math> is the ending of the <math>{{i}^{th}}</math> interval
 
For the purposes of MLE, left censored data will be considered to be intervals with <math>T_{Li}^{\prime \prime }=0.</math>
 
The solution of the maximum log-likelihood function is found by solving for (<math>\widehat{\mu },\widehat{\sigma })</math> so that <math>\tfrac{\partial \Lambda }{\partial \mu }=0,\tfrac{\partial \Lambda }{\partial k}=0.</math>
 
::<math>\begin{align}
  \frac{\partial \Lambda }{\partial \mu }= & -k\underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}+\underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}{{e}^{\ln ({{T}_{i}})-\mu }} \\
  & +\frac{1}{\Gamma (k)}\underset{i=1}{\mathop{\overset{S}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime }}\frac{{{e}^{k\left( \ln (T_{i}^{^{\prime }})-\mu )-{{e}^{\ln (T_{i}^{^{\prime }})-\mu )}} \right)}}}{1-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{i}^{^{\prime }})-\mu }} \right)} \\
  & +\frac{1}{\Gamma (k)}\underset{i=1}{\mathop{\overset{{{F}_{I}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime \prime }}\{\frac{{{e}^{k{{e}^{{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }}}}-{{e}^{{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }}}}}}}{{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }} \right)-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }} \right)} \\
  & -\frac{{{e}^{k{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }}-{{e}^{{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }}}}}}}{{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }} \right)-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }} \right)}\} 
\end{align}</math>
 
::<math>\begin{align}
  \frac{\partial \Lambda }{\partial k}= & \underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,{{N}_{i}}(\ln ({{T}_{i}})-\mu )-\frac{{{\Gamma }^{^{\prime }}}(k)\underset{i=1}{\mathop{\overset{{{F}_{e}}}{\mathop{\mathop{}_{}^{}}}\,}}\,{{N}_{i}}}{\Gamma (k)} \\
  & -\underset{i=1}{\mathop{\overset{S}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime }}\frac{\tfrac{\partial {{\Gamma }_{1}}(k;{{e}^{\ln (T_{i}^{^{\prime }})-\mu }})}{\partial k}}{1-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{i}^{^{\prime }})-\mu }} \right)} \\
  & +\underset{i=1}{\mathop{\overset{{{F}_{I}}}{\mathop{\underset{}{\overset{}{\mathop \sum }}\,}}\,}}\,N_{i}^{^{\prime \prime }}\left( \frac{\tfrac{\partial {{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }} \right)}{\partial k}-\tfrac{\partial {{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }} \right)}{\partial k}}{{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{R}_{i}}}^{^{\prime \prime }})-\mu }} \right)-{{\Gamma }_{1}}\left( k;{{e}^{\ln (T_{{{L}_{i}}}^{^{\prime \prime }})-\mu }}) \right)} \right) 
\end{align}</math>
<br>

Revision as of 13:21, 29 October 2011

New format available! This reference is now available in a new format that offers faster page load, improved display for calculations and images, more targeted search and the latest content available as a PDF. As of September 2023, this Reliawiki page will not continue to be updated. Please update all links and bookmarks to the latest reference at help.reliasoft.com/reference/life_data_analysis

Chapter 4B: Maximum Likelihood Parameter Estimation


Weibullbox.png

Chapter 4B  
Maximum Likelihood Parameter Estimation  

Synthesis-icon.png

Available Software:
Weibull++

Examples icon.png

More Resources:
Weibull++ Examples Collection


Maximum Likelihood Estimation

If [math]\displaystyle{ x }[/math] is a continuous random variable with [math]\displaystyle{ pdf\ \ : }[/math]

[math]\displaystyle{ f(x;{{\theta }_{1}},{{\theta }_{2}},...,{{\theta }_{k}}), }[/math]

where [math]\displaystyle{ {{\theta }_{1}}, }[/math] [math]\displaystyle{ {{\theta }_{2}}, }[/math][math]\displaystyle{ ..., }[/math] [math]\displaystyle{ {{\theta }_{k}} }[/math] are [math]\displaystyle{ k }[/math] unknown constant parameters that need to be estimated, conduct an experiment and obtain [math]\displaystyle{ N }[/math] independent observations, [math]\displaystyle{ {{x}_{1}}, }[/math] [math]\displaystyle{ {{x}_{2}}, }[/math][math]\displaystyle{ ..., }[/math] [math]\displaystyle{ {{x}_{N}} }[/math], which correspond in the case of life data analysis to failure times. The likelihood function (for complete data) is given by:

[math]\displaystyle{ L({{x}_{1}},{{x}_{2}},...,{{x}_{N}}|{{\theta }_{1}},{{\theta }_{2}},...,{{\theta }_{k}})=L=\underset{i=1}{\overset{N}{\mathop \prod }}\,f({{x}_{i}};{{\theta }_{1}},{{\theta }_{2}},...,{{\theta }_{k}}) }[/math]
[math]\displaystyle{ i=1,2,...,N }[/math]


The logarithmic likelihood function is:

[math]\displaystyle{ \Lambda =\ln L=\underset{i=1}{\overset{N}{\mathop \sum }}\,\ln f({{x}_{i}};{{\theta }_{1}},{{\theta }_{2}},...,{{\theta }_{k}}) }[/math]

The maximum likelihood estimators (MLE) of [math]\displaystyle{ {{\theta }_{1}},{{\theta }_{2}},...,{{\theta }_{k}}, }[/math] are obtained by maximizing [math]\displaystyle{ L }[/math] or [math]\displaystyle{ \Lambda . }[/math]
By maximizing [math]\displaystyle{ \Lambda , }[/math] which is much easier to work with than [math]\displaystyle{ L }[/math], the maximum likelihood estimators (MLE) of [math]\displaystyle{ {{\theta }_{1}},{{\theta }_{2}},...,{{\theta }_{k}} }[/math] are the simultaneous solutions of [math]\displaystyle{ k }[/math] equations such that:

[math]\displaystyle{ \frac{\partial (\Lambda )}{\partial {{\theta }_{j}}}=0,j=1,2,...,k }[/math]


Even though it is common practice to plot the MLE solutions using median ranks (points are plotted according to median ranks and the line according to the MLE solutions), this is not completely accurate. As it can be seen from the equations above, the MLE method is independent of any kind of ranks. For this reason, many times the MLE solution appears not to track the data on the probability plot. This is perfectly acceptable since the two methods are independent of each other, and in no way suggests that the solution is wrong.

Illustrating the MLE Method Using the Exponential Distribution

• To estimate [math]\displaystyle{ \widehat{\lambda } }[/math] for a sample of [math]\displaystyle{ n }[/math] units (all tested to failure), first obtain the likelihood function:


[math]\displaystyle{ \begin{align} L(\lambda |{{t}_{1}},{{t}_{2}},...,{{t}_{n}})= & \underset{i=1}{\overset{n}{\mathop \prod }}\,f({{t}_{i}}) \\ = & \underset{i=1}{\overset{n}{\mathop \prod }}\,\lambda {{e}^{-\lambda {{t}_{i}}}} \\ = & {{\lambda }^{n}}\cdot {{e}^{-\lambda \underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{t}_{i}}}} \end{align} }[/math]


• Take the natural log of both sides:
[math]\displaystyle{ \Lambda =\ln (L)=n\ln (\lambda )-\lambda \underset{i=1}{\overset{n}{\mathop \sum }}\,{{t}_{i}}. }[/math]
• Obtain [math]\displaystyle{ \tfrac{\partial \Lambda }{\partial \lambda } }[/math], and set it equal to zero:
[math]\displaystyle{ \frac{\partial \Lambda }{\partial \lambda }=\frac{n}{\lambda }-\underset{i=1}{\overset{n}{\mathop \sum }}\,{{t}_{i}}=0 }[/math]
• Solve for [math]\displaystyle{ \widehat{\lambda } }[/math] or:
[math]\displaystyle{ \hat{\lambda }=\frac{n}{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{t}_{i}}} }[/math]


Notes About [math]\displaystyle{ \widehat{\lambda } }[/math]

Note that the value of [math]\displaystyle{ \widehat{\lambda } }[/math] is an estimate because if we obtain another sample from the same population and re-estimate [math]\displaystyle{ \lambda }[/math], the new value would differ from the one previously calculated. In plain language, [math]\displaystyle{ \hat{\lambda } }[/math] is an estimate of the true value of ... How close is the value of our estimate to the true value? To answer this question, one must first determine the distribution of the parameter, in this case [math]\displaystyle{ \lambda }[/math]. This methodology introduces a new term, confidence bound, which allows us to specify a range for our estimate with a certain confidence level. The treatment of confidence bounds is integral to reliability engineering, and to all of statistics. (Confidence bounds are covered in Chapter 5.)

Illustrating the MLE Method Using Normal Distribution

To obtain the MLE estimates for the mean, [math]\displaystyle{ \bar{T}, }[/math] and standard deviation, [math]\displaystyle{ {{\sigma }_{T}}, }[/math] for the normal distribution, start with the [math]\displaystyle{ pdf }[/math] of the normal distribution which is given by:

[math]\displaystyle{ f(T)=\frac{1}{{{\sigma }_{T}}\sqrt{2\pi }}{{e}^{-\tfrac{1}{2}{{\left( \tfrac{T-\bar{T}}{{{\sigma }_{T}}} \right)}^{2}}}} }[/math]

If [math]\displaystyle{ {{T}_{1}},{{T}_{2}},...,{{T}_{N}} }[/math] are known times-to-failure (and with no suspensions), then the likelihood function is given by:

[math]\displaystyle{ L({{T}_{1}},{{T}_{2}},...,{{T}_{N}}|\bar{T},{{\sigma }_{T}})=L=\underset{i=1}{\overset{N}{\mathop \prod }}\,\left[ \frac{1}{{{\sigma }_{T}}\sqrt{2\pi }}{{e}^{-\tfrac{1}{2}{{\left( \tfrac{{{T}_{i}}-\bar{T}}{{{\sigma }_{T}}} \right)}^{2}}}} \right] }[/math]
[math]\displaystyle{ L=\frac{1}{{{({{\sigma }_{T}}\sqrt{2\pi })}^{N}}}{{e}^{-\tfrac{1}{2}\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{\left( \tfrac{{{T}_{i}}-\bar{T}}{{{\sigma }_{T}}} \right)}^{2}}}} }[/math]


then:


[math]\displaystyle{ \Lambda =\ln L=-\frac{N}{2}\ln (2\pi )-N\ln {{\sigma }_{T}}-\frac{1}{2}\underset{i=1}{\overset{N}{\mathop \sum }}\,\left( \frac{{{T}_{i}}-\bar{T}}{{{\sigma }_{T}}} \right)_{}^{2} }[/math]

Then taking the partial derivatives of [math]\displaystyle{ \Lambda }[/math] with respect to each one of the parameters and setting them equal to zero yields:

[math]\displaystyle{ \frac{\partial (\Lambda )}{\partial \bar{T}}=\frac{1}{\sigma _{T}^{2}}\underset{i=1}{\overset{N}{\mathop \sum }}\,({{T}_{i}}-\bar{T})=0 }[/math]
and:


[math]\displaystyle{ \frac{\partial (\Lambda )}{\partial {{\sigma }_{T}}}=-\frac{N}{{{\sigma }_{T}}}+\frac{1}{\sigma _{T}^{3}}\underset{i=1}{\overset{N}{\mathop \sum }}\,{{({{T}_{i}}-\bar{T})}^{2}}=0 }[/math]

Solving Eqns. (dldt) and (dlds) simultaneously yields:

[math]\displaystyle{ \bar{T}=\frac{1}{N}\underset{i=1}{\overset{N}{\mathop \sum }}\,{{T}_{i}} }[/math]
and:
[math]\displaystyle{ \begin{align} & \hat{\sigma }_{T}^{2}= & \frac{1}{N}\underset{i=1}{\overset{N}{\mathop \sum }}\,{{({{T}_{i}}-\bar{T})}^{2}} \\ & & \\ & {{{\hat{\sigma }}}_{T}}= & \sqrt{\frac{1}{N}\underset{i=1}{\overset{N}{\mathop \sum }}\,{{({{T}_{i}}-\bar{T})}^{2}}} \end{align} }[/math]

It should be noted that these solutions are only valid for data with no suspensions, i.e. all units are tested to failure. In the case where suspensions are present or all units are not tested to failure, the methodology changes and the problem becomes much more complicated.

Illustrating with an Example of the Normal Distribution

If we had five units that failed at 10, 20, 30, 40 and 50 hours, the mean would be:

[math]\displaystyle{ \begin{align} \bar{T}= & \frac{1}{N}\underset{i=1}{\overset{N}{\mathop \sum }}\,{{T}_{i}} \\ = & \frac{10+20+30+40+50}{5} \\ = & 30 \end{align} }[/math]

The standard deviation estimate then would be:

[math]\displaystyle{ \begin{align} {{{\hat{\sigma }}}_{T}}= & \sqrt{\frac{1}{N}\underset{i=1}{\overset{N}{\mathop \sum }}\,{{({{T}_{i}}-\bar{T})}^{2}}} \\ = & \sqrt{\frac{{{(10-30)}^{2}}+{{(20-30)}^{2}}+{{(30-30)}^{2}}+{{(40-30)}^{2}}+{{(50-30)}^{2}}}{5}}, \\ = & 14.1421 \end{align} }[/math]

A look at the likelihood function surface plot in Figure A-1 reveals that both of these values are the maximum values of the function.

LdaA.1.gif

This three-dimensional plot represents the likelihood function. As can be seen from the plot, the maximum likelihood estimates for the two parameters correspond with the peak or maximum of the likelihood function surface.

Distribution Log-Likelihood Equations

The following list provides links to the log-likelihood functions and their associated partial derivatives for most of the distributions available in Weibull++.