<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.reliawiki.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Pengying+niu</id>
	<title>ReliaWiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://www.reliawiki.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Pengying+niu"/>
	<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php/Special:Contributions/Pengying_niu"/>
	<updated>2026-04-06T06:22:51Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.44.0</generator>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33823</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33823"/>
		<updated>2012-08-24T06:36:11Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Residual Analysis */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}:{{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})&amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}}) &amp;amp;= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{R}^{2}} &amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   R_{adj}^{2} &amp;amp;= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{d}_{i}}&amp;amp;= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{r}_{i}} &amp;amp;= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33822</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33822"/>
		<updated>2012-08-24T06:35:30Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Coefficient of Multiple Determination,  {{R}^{2}} */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}:{{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})&amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}}) &amp;amp;= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{R}^{2}} &amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   R_{adj}^{2} &amp;amp;= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33821</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33821"/>
		<updated>2012-08-24T06:34:49Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Sequential Sum of Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}:{{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})&amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}}) &amp;amp;= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33820</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33820"/>
		<updated>2012-08-24T06:33:21Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Partial Sum of Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}) &amp;amp;= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}} &amp;amp;= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33819</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33819"/>
		<updated>2012-08-24T06:31:51Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Partial Sum of Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33818</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33818"/>
		<updated>2012-08-24T06:30:18Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Calculation of the Statistic  {{F}_{0}} */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33817</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33817"/>
		<updated>2012-08-24T06:28:42Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Test on Individual Regression Coefficients ( t  Test) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp; = &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33816</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33816"/>
		<updated>2012-08-24T06:27:38Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Test on Individual Regression Coefficients ( t  Test) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\sigma }}}^{2}} &amp;amp;= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   C &amp;amp;= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   se({{{\hat{\beta }}}_{1}}) &amp;amp;= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
  se({{{\hat{\beta }}}_{2}})&amp;amp; = &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
  {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &amp;amp;= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{t}_{\alpha /2,n-(k+1)}} &amp;amp;= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
  -{{t}_{\alpha /2,n-(k+1)}} &amp;amp; = &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33815</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33815"/>
		<updated>2012-08-24T06:26:01Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Test for Significance of Regression */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}:&amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}:&amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   H &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp; = &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{R}}&amp;amp; = &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}} &amp;amp;= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{E}} &amp;amp;= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{0}}&amp;amp; = &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{f}_{\alpha ,k,n-(k+1)}} &amp;amp;= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33814</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33814"/>
		<updated>2012-08-24T06:17:40Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Estimating Regression Models Using Least Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y}(47,31)&amp;amp; = &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; H= &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{R}}= &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{E}}= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{\alpha ,k,n-(k+1)}}= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33813</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33813"/>
		<updated>2012-08-24T06:16:38Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Estimating Regression Models Using Least Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{e}_{i}} &amp;amp; = &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
  {{e}_{5}}&amp;amp; = &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}(47,31)= &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; H= &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{R}}= &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{E}}= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{\alpha ,k,n-(k+1)}}= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33812</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33812"/>
		<updated>2012-08-24T06:15:49Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Estimating Regression Models Using Least Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{i}} &amp;amp;= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
  {{{\hat{y}}}_{5}} &amp;amp; = &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{e}_{i}}= &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
 &amp;amp; {{e}_{5}}= &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}(47,31)= &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; H= &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{R}}= &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{E}}= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{\alpha ,k,n-(k+1)}}= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33811</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33811"/>
		<updated>2012-08-24T06:14:47Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Estimating Regression Models Using Least Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp;= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{i}}= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
 &amp;amp; {{{\hat{y}}}_{5}}= &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{e}_{i}}= &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
 &amp;amp; {{e}_{5}}= &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}(47,31)= &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; H= &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{R}}= &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{E}}= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{\alpha ,k,n-(k+1)}}= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33810</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33810"/>
		<updated>2012-08-24T06:14:15Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Estimating Regression Models Using Least Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{y} &amp;amp; = &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{i}}= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
 &amp;amp; {{{\hat{y}}}_{5}}= &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{e}_{i}}= &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
 &amp;amp; {{e}_{5}}= &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}(47,31)= &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; H= &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{R}}= &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{E}}= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{\alpha ,k,n-(k+1)}}= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33809</id>
		<title>Multiple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Multiple_Linear_Regression_Analysis&amp;diff=33809"/>
		<updated>2012-08-24T06:13:30Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Estimating Regression Models Using Least Squares */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|4}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in DOE++ are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Chapter 9. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. &lt;br /&gt;
ANOVA models are discussed in Chapter 6.&lt;br /&gt;
&lt;br /&gt;
==Multiple Linear Regression Model==&lt;br /&gt;
&lt;br /&gt;
A linear regression model that contains more than one predictor variable is called a multiple linear regression model. The following model is a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The model is linear because it is linear in the parameters  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The model describes a plane in the three dimensional space of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The parameter  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is the intercept of this plane. Parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are referred to as partial regression coefficients. Parameter  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is held constant. Parameter  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  represents the change in the mean response corresponding to a unit change in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is held constant.  &lt;br /&gt;
Consider the following example of a multiple linear regression model with two predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is one. The regression plane corresponding to this model is shown in Figure TrueRegrPlane. Also shown is an observed data point and the corresponding random error,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; . The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Section 5.MatrixApproach.&lt;br /&gt;
 &lt;br /&gt;
Figure ContourPlot1 shows the contour plot for the regression model of Eqn. (FirstOrderModelExample). The contour plot shows lines of constant mean response values as a function of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.&lt;br /&gt;
 &lt;br /&gt;
A linear regression model may also take the following form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.1.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
A cross-product term,  &amp;lt;math&amp;gt;{{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt; , is included in the model. This term represents an interaction effect between the two variables  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation  &amp;lt;math&amp;gt;Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression plane and contour plot for this model are shown in Figures RegrPlaneWInteraction and ContourPlotWInteraction, respectively.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.2.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+\epsilon&amp;lt;/math&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
 &lt;br /&gt;
Now consider the regression model shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is also a linear regression model and is referred to as a polynomial regression model. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.3.png|thumb|center|300px|Regression plane for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.4.png|thumb|center|300px|Countour plot for the model &amp;lt;math&amp;gt;Y=30+5 x_1+7 x_2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This model is a second order model because the maximum power of the terms in the model is two. The regression surface for this model is shown in Figure PolynomialRegrSurface. Such regression models are used in RSM to find the optimum value of the response,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (for details see Chapter 9). Notice that, although the shape of the regression surface is curvilinear, the regression model of Eqn. (SecondOrderModelEx) is still linear because the model is linear in the parameters. The contour plot for this model is shown in Figure ContourPlotPolynomialRegr.&lt;br /&gt;
All multiple linear regression models can be expressed in the following general form:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  denotes the number of terms in the model. For example, the model of Eqn. (SecondOrderModelEx) can be written in the general form using  &amp;lt;math&amp;gt;{{x}_{3}}=x_{1}^{2}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{4}}=x_{2}^{3}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{5}}={{x}_{1}}{{x}_{2}}&amp;lt;/math&amp;gt;  as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Estimating Regression Models Using Least Squares==&lt;br /&gt;
&lt;br /&gt;
Consider a multiple linear regression model with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Let each of the  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; , have  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. Then  &amp;lt;math&amp;gt;{{x}_{ij}}&amp;lt;/math&amp;gt;  represents the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; . For example,  &amp;lt;math&amp;gt;{{x}_{51}}&amp;lt;/math&amp;gt;  represents the fifth level of the first predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , while  &amp;lt;math&amp;gt;{{x}_{19}}&amp;lt;/math&amp;gt;  represents the first level of the ninth predictor variable,  &amp;lt;math&amp;gt;{{x}_{9}}&amp;lt;/math&amp;gt; . Observations,  &amp;lt;math&amp;gt;{{y}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{y}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{y}_{n}}&amp;lt;/math&amp;gt; , recorded for each of these  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels can be expressed in the following way:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{y}_{1}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ &lt;br /&gt;
 &amp;amp; {{y}_{2}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{i}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; .. \\ &lt;br /&gt;
 &amp;amp; {{y}_{n}}= &amp;amp; {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.5.png|thumb|center|400px|Regression surface for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.6.png|thumb|center|400px|Contour plot for the model &amp;lt;math&amp;gt;500+5 x_1+7 x_2-3 x_1^2-5 x_2^2+3 x_1 x_2+\epsilon &amp;lt;/math&amp;gt;.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  equations shown previously can be represented in matrix notation as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:where&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;y=\left[ \begin{matrix}&lt;br /&gt;
   {{y}_{1}}  \\&lt;br /&gt;
   {{y}_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{y}_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{      }X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; {{x}_{11}} &amp;amp; {{x}_{12}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{1n}}  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{21}} &amp;amp; {{x}_{22}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{2n}}  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; {} &amp;amp; {} &amp;amp; {} &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; {{x}_{n1}} &amp;amp; {{x}_{n2}} &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; {{x}_{nn}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{0}}  \\&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\beta }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]\text{    and   }\epsilon =\left[ \begin{matrix}&lt;br /&gt;
   {{\epsilon }_{1}}  \\&lt;br /&gt;
   {{\epsilon }_{2}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{\epsilon }_{n}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  in Eqn. (TrueModelMatrixNotation) is referred to as the design matrix. It contains information about the levels of the predictor variables at which the observations are obtained.  The vector  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  contains all the regression coefficients. To obtain the regression model,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  should be known.  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  is estimated using least square estimates. The following equation is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;^{\prime }&amp;lt;/math&amp;gt;  represents the transpose of the matrix while  &amp;lt;math&amp;gt;^{-1}&amp;lt;/math&amp;gt;  represents the matrix inverse. Knowing the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , the multiple linear regression model can now be estimated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=X\hat{\beta }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated regression model is also referred to as the fitted model. The observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , may be different from the fitted values  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  obtained from this model. The difference between these two values is the residual,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The vector of residuals,  &amp;lt;math&amp;gt;e&amp;lt;/math&amp;gt; , is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e=y-\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The fitted model of Eqn. (FittedValueMatrixNotation) can also be written as follows, using  &amp;lt;math&amp;gt;\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y&amp;lt;/math&amp;gt;  from Eqn. (LeastSquareEstimate):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}= &amp;amp; X\hat{\beta } \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; Hy  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}&amp;lt;/math&amp;gt; . The matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , is referred to as the hat matrix. It transforms the vector of the observed response values,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; , to the vector of fitted values,  &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
An analyst studying a chemical process expects the yield to be affected by the levels of two factors,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . Observations recorded for various levels of the two factors are shown in Table 5.1. The analyst wants to fit a first order regression model to the data. Interaction between  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.1.png|thumb|center|400px|Observed yield data for various levels of two factors.]]&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The data of Table 5.1 can be entered into DOE++ using the Multiple Regression tool as shown in Figure MLRTDataEntrySshot. A scatter plot for the data in Table 5.1 is shown in Figure ThreedScatterPlot. The first order regression model applicable to this data set having two predictor variables is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where the dependent variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , represents the yield and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , represent the two factors respectively. The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the data can be obtained as:  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;X=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 43.4 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }y=\left[ \begin{matrix}&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   251.3  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   349.0  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.7.png|thumb|center|400px|Multiple Regression tool in DOE++ with the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.8.png|thumb|center|400px|Three dimensional scatter plot for the observed data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , can now be obtained:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   \hat{\beta } &amp;amp;= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; {{\left[ \begin{matrix}&lt;br /&gt;
   17 &amp;amp; 941 &amp;amp; 525.3  \\&lt;br /&gt;
   941 &amp;amp; 54270 &amp;amp; 29286  \\&lt;br /&gt;
   525.3 &amp;amp; 29286 &amp;amp; 16254  \\&lt;br /&gt;
\end{matrix} \right]}^{-1}}\left[ \begin{matrix}&lt;br /&gt;
   4902.8  \\&lt;br /&gt;
   276610  \\&lt;br /&gt;
   152020  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{\beta }=\left[ \begin{matrix}&lt;br /&gt;
   {{{\hat{\beta }}}_{0}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{1}}  \\&lt;br /&gt;
   {{{\hat{\beta }}}_{2}}  \\&lt;br /&gt;
\end{matrix} \right]=\left[ \begin{matrix}&lt;br /&gt;
   -153.51  \\&lt;br /&gt;
   1.24  \\&lt;br /&gt;
   12.08  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
and the estimated regression coefficients are  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}=-153.51&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}=1.24&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}=12.08&amp;lt;/math&amp;gt; . The fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24{{x}_{1}}+12.08{{x}_{2}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, the fitted regression model can be viewed using the Show Analysis Summary icon in the Control Panel. The model is shown in Figure EquationScreenshot.&lt;br /&gt;
&lt;br /&gt;
A plot of the fitted regression plane is shown in Figure FittedRegrModel. The fitted regression model can be used to obtain fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to an observed response value,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; . For example, the fitted value corresponding to the fifth observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.9.png|thumb|center|400px|Equation of the fitted regression model for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.10.png|thumb|center|400px|Fitted regression plane &amp;lt;math&amp;gt;\hat{y}=-153.5+1.24 x_1+12.08 x_2 &amp;lt;/math&amp;gt; for the data of Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{i}}= &amp;amp; -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ &lt;br /&gt;
 &amp;amp; {{{\hat{y}}}_{5}}= &amp;amp; -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; -153.5+1.24(47.3)+12.08(29.9) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 266.3  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed fifth response value is  &amp;lt;math&amp;gt;{{y}_{5}}=273.0&amp;lt;/math&amp;gt; . The residual corresponding to this value is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{e}_{i}}= &amp;amp; {{y}_{i}}-{{{\hat{y}}}_{i}} \\ &lt;br /&gt;
 &amp;amp; {{e}_{5}}= &amp;amp; {{y}_{5}}-{{{\hat{y}}}_{5}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 273.0-266.3 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6.7  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure DiagnosticSshot. The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and 31 units of  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , the value is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{y}(47,31)= &amp;amp; -153.5+1.24(47)+12.08(31) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 279.26  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
===Properties of the Least Square Estimators,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt;===&lt;br /&gt;
The least square estimates,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\hat{\beta }}_{k}}&amp;lt;/math&amp;gt; , are unbiased estimators of  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{\beta }_{k}}&amp;lt;/math&amp;gt; , provided that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed. The variances of the  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; s are obtained using the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.11.png|thumb|center|400px|Fitted values and residuals for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is a symmetric matrix whose diagonal elements,  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt; , represent the variance of the estimated  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The off-diagonal elements,  &amp;lt;math&amp;gt;{{C}_{ij}}&amp;lt;/math&amp;gt; , represent the covariance between the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th and  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th estimated regression coefficients,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; . The value of  &amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}&amp;lt;/math&amp;gt;  is obtained using the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , which can be calculated as discussed in Section 5.MANOVA. The variance-covariance matrix for the data in Table 5.1 is shown in Figure VarCovMatrixSshot. It is available in DOE++ using the Show Analysis Summary icon in the Control Panel. Calculations to obtain the matrix are given in Example 3 in Section 5.tTest. The positive square root of  &amp;lt;math&amp;gt;{{C}_{jj}}&amp;lt;/math&amp;gt;  represents the estimated standard deviation of the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th regression coefficient,  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt; , and is called the estimated standard error of  &amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}&amp;lt;/math&amp;gt;  (abbreviated  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.12.png|thumb|center|400px|The variance-covariance matrix for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are normally and independently distributed with a mean of zero and variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Three types of hypothesis tests can be carried out for multiple linear regression models:&lt;br /&gt;
:•	Test for significance of regression&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of the whole regression model.&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test checks the significance of individual regression coefficients.&lt;br /&gt;
&lt;br /&gt;
:•	Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test&lt;br /&gt;
&lt;br /&gt;
This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.&lt;br /&gt;
&lt;br /&gt;
===Test for Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0\text{     for at least one }j  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test for  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is carried out using the following statistic:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  is the regression mean square and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square. If the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is true then the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;n-&amp;lt;/math&amp;gt; ( &amp;lt;math&amp;gt;k+1&amp;lt;/math&amp;gt; ) degrees of freedom in the denominator.  The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,k,n-(k+1)}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
To calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , the mean squares  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  must be known. As explained in Chapter 4, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  is the total sum of squares and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  is the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . In multiple linear regression, the following equation is used to calculate  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; : &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations (that was defined in Section 5.MatrixApproach),  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ). Knowing  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt;  the total mean square,  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
The regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , is obtained by dividing the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations,  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix (that was defined in Section 5.MatrixApproach) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  represents an  &amp;lt;math&amp;gt;n\times n&amp;lt;/math&amp;gt;  square matrix of ones. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt;  the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
The error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is obtained by dividing the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , by the respective degrees of freedom,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , is calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}={{y}^{\prime }}(I-H)y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  is the vector of observations,  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix of order  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is the hat matrix. The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the total number of observations and  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  is the number of predictor variables in the model. Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be calculated. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=M{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression, for the regression model obtained for the data in Table 5.1, is illustrated in this example. The null hypothesis for the model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To calculate  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  to carry out the significance test.&lt;br /&gt;
The regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  is calculated as follows using the design matrix  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from Example 1:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; H= &amp;amp; X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   0.27552 &amp;amp; 0.25154 &amp;amp; . &amp;amp; . &amp;amp; -0.04030  \\&lt;br /&gt;
   0.25154 &amp;amp; 0.23021 &amp;amp; . &amp;amp; . &amp;amp; -0.029120  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; . &amp;amp; . &amp;amp; . &amp;amp; .  \\&lt;br /&gt;
   -0.04030 &amp;amp; -0.02920 &amp;amp; . &amp;amp; . &amp;amp; 0.30115  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt; , the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , can be calculated:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , which equals to a value of two since there are two predictor variables in the data in Table 5.1. Therefore, the regression mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{R}}= &amp;amp; \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12816.35}{2} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 6408.17  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly to calculate the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; {{y}^{\prime }}\left[ I-H \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 423.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;n-(k+1)&amp;lt;/math&amp;gt; . Therefore, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{E}}= &amp;amp; \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{E}}}{(n-(k+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{423.37}{(17-(2+1))} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The statistic to test the significance of regression can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{6408.17}{423.37/(17-3)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 211.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical value for this test, corresponding to a significance level of 0.1, is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{\alpha ,k,n-(k+1)}}= &amp;amp; {{f}_{0.1,2,14}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{0.1,2,14}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\beta }_{1}}={{\beta }_{2}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that at least one coefficient out of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in Table 5.1. The analysis of variance is summarized in Table 5.2.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.2.png|thumb|center|300px|ANOVA table for the significance of regression test in Example 2.]]&lt;br /&gt;
&lt;br /&gt;
===Test on Individual Regression Coefficients ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{j}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{j}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution (and is similar to the one used in the case of simple linear regression models in Chapter 4):&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where the standard error,  &amp;lt;math&amp;gt;se({{\hat{\beta }}_{j}})&amp;lt;/math&amp;gt; , is obtained from Eqn. (StandardErrorBetaJ). The analyst would fail to reject the null hypothesis if the test statistic, calculated using Eqn. (TtestStatistic), lies in the acceptance region:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha /2,n-2}}&amp;lt;{{T}_{0}}&amp;lt;{{t}_{\alpha /2,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This test measures the contribution of a variable while the remaining variables are included in the model. For the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; , if the test is carried out for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , then the test will check the significance of including the variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  in the model that contains  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{3}}&amp;lt;/math&amp;gt;  (i.e. the model  &amp;lt;math&amp;gt;\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}&amp;lt;/math&amp;gt; ). Hence the test is also referred to as partial or marginal test. In DOE++, this test is displayed in the Regression Information table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The test to check the significance of the estimated regression coefficients for the data in Table 5.1 is illustrated in this example. The null hypothesis to test the coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner. To calculate the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , we need to calculate the standard error using Eqn. (StandardErrorBetaJ).&lt;br /&gt;
In Example 2, the value of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , was obtained as 30.24. The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . &lt;br /&gt;
&lt;br /&gt;
:Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\sigma }}}^{2}}= &amp;amp; M{{S}_{E}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance-covariance matrix of the estimated regression coefficients is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; C= &amp;amp; {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 30.24\left[ \begin{matrix}&lt;br /&gt;
   336.5 &amp;amp; 1.2 &amp;amp; -13.1  \\&lt;br /&gt;
   1.2 &amp;amp; 0.005 &amp;amp; -0.049  \\&lt;br /&gt;
   -13.1 &amp;amp; -0.049 &amp;amp; 0.5  \\&lt;br /&gt;
\end{matrix} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   10176.75 &amp;amp; 37.145 &amp;amp; -395.83  \\&lt;br /&gt;
   37.145 &amp;amp; 0.1557 &amp;amp; -1.481  \\&lt;br /&gt;
   -395.83 &amp;amp; -1.481 &amp;amp; 15.463  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
From the diagonal elements of  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , the estimated standard error for  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; se({{{\hat{\beta }}}_{1}})= &amp;amp; \sqrt{0.1557}=0.3946 \\ &lt;br /&gt;
 &amp;amp; se({{{\hat{\beta }}}_{2}})= &amp;amp; \sqrt{15.463}=3.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding test statistics for these coefficients are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ &lt;br /&gt;
 &amp;amp; {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}= &amp;amp; \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The critical values for the present  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test at a significance of 0.1 are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{t}_{\alpha /2,n-(k+1)}}= &amp;amp; {{t}_{0.05,14}}=1.761 \\ &lt;br /&gt;
 &amp;amp; -{{t}_{\alpha /2,n-(k+1)}}= &amp;amp; -{{t}_{0.05,14}}=-1.761  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Considering  &amp;lt;math&amp;gt;{{\hat{\beta }}_{2}}&amp;lt;/math&amp;gt; , it can be seen that  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}&amp;lt;/math&amp;gt;  does not lie in the acceptance region of  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}&amp;lt;{{t}_{0}}&amp;lt;{{t}_{0.05,14}}&amp;lt;/math&amp;gt; . The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{2}}=0&amp;lt;/math&amp;gt; , is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant at  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; . This conclusion can also be arrived at using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value noting that the hypothesis is two-sided. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}=&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;3.0726&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with 14 degrees of freedom is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 2\times (1-P(T\le |{{t}_{0}}|) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2\times (1-0.9959) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0083  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is less than the significance,  &amp;lt;math&amp;gt;\alpha =0.1&amp;lt;/math&amp;gt; , it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is significant. The hypothesis test on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner.&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, in DOE++, the information related to the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test is displayed in the Regression Information table as shown in Figure RegrInfoSshot. In this table, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is displayed in the row for the term Factor 2 because  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test and the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test, respectively. These values have been calculated for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated using Eqn. (LeastSquareEstimate) as shown in Example &lt;br /&gt;
&lt;br /&gt;
:1. The Effect column represents values obtained by multiplying the coefficients by a factor of &lt;br /&gt;
:2. This value is useful in the case of two factor experiments and is explained in Chapter 7. &lt;br /&gt;
&lt;br /&gt;
Columns labeled Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Section 5.RegrCoeffCI. The Variance Inflation Factor column displays values that give a measure of multicollinearity. This is explained in &lt;br /&gt;
Section 5.MultiCollinearity. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.13.png|thumb|center|400px|Regression results for the data in Table 5.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Test on Subsets of Regression Coefficients (Partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Test)===&lt;br /&gt;
&lt;br /&gt;
This test can be considered to be the general form of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; . The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the extra sum of squares. &lt;br /&gt;
Assume that the vector of the regression coefficients,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; , for the multiple linear regression model,  &amp;lt;math&amp;gt;y=X\beta +\epsilon &amp;lt;/math&amp;gt; , is partitioned into two vectors with the second vector,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; , containing the last  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  regression coefficients, and the first vector,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , containing the first ( &amp;lt;math&amp;gt;k+1-r&amp;lt;/math&amp;gt; ) coefficients as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\beta =\left[ \begin{matrix}&lt;br /&gt;
   {{\beta }_{1}}  \\&lt;br /&gt;
   {{\beta }_{2}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:with:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}&#039;\text{ and }{{\beta }_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}&#039;\text{    }&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hypothesis statements to test the significance of adding the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  to a model containing the regression coefficients in  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  may be written as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\beta }_{2}}=0 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\beta }_{2}}\ne 0  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for this test follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution and can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  is the the increase in the regression sum of squares when the variables corresponding to the coefficients in  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  are added to a model already containing  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is obtained from Eqn. (ErrorMeanSquare). The value of the extra sum of squares is obtained as explained in the next section.&lt;br /&gt;
&lt;br /&gt;
The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is rejected if  &amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,r,n-(k+1)}}&amp;lt;/math&amp;gt; . Rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  leads to the conclusion that at least one of the variables in  &amp;lt;math&amp;gt;{{x}_{k-r+1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{k-r+2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt;  contributes significantly to the regression model.  In DOE++, the results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Types of Extra Sum of Squares===&lt;br /&gt;
The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic of Eqn. (PartialFtest). In DOE++, selection for the type of extra sum of squares is available in the Options tab of the Control Panel as shown in Figure SSselectionSshot. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.14.png|thumb|center|500px|Selection of the type of extra sum of squared in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
====Partial Sum of Squares====&lt;br /&gt;
The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that we need to know the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is the increase in the regression sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  is added to the model. This increase is the difference in the regression sum of squares for the full model of Eqn. (PartialSSFullModel) and the model that includes all terms except  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; . These terms are  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt; . The model that contains these terms is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be represented as  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})&amp;lt;/math&amp;gt;  and is calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{. () }-S{{S}_{R}}\text{ for Eqn}\text{. ()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{2}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the partial sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients other than the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
DOE++ has the partial sum of squares as the default selection. This is because the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test explained in Section 5.tTest is a partial test, i.e. the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test are displayed in the Regression Information table. The results from the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table.&lt;br /&gt;
The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the partial sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the full model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model excluding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; . The regression sum of squares for the full model can be obtained using Eqn. (TotalSumofSquares) and has been calculated in Example 2 as  &amp;lt;math&amp;gt;12816.35&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the second column in the design matrix of the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The second column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  which is no longer in the model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt; , can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12518.32  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the partial sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12816.35-12518.32 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.03  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the partial sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.03/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 9.855  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.9928 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.0072  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. In the results obtained from DOE++, the calculations for this test are displayed in the ANOVA table as shown in Figure AnovaTableSshot. Note that the conclusion obtained in this example can also be obtained using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test as explained in Example 3 in Section 5.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the variables included in the multiple linear regression model.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
====Sequential Sum of Squares====&lt;br /&gt;
The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is the increase in the sum of squares when  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added to the model observing the sequence of Eqn. (SeqSSEqn). Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added and the regression sum of squares for the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  was added to the model. The model after  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added is as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.15.png|thumb|center|500px|ANOVA results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
This is because to maintain the sequence of Eqn. (SeqSSEqn) all coefficients preceding  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  must be included in the model. These are the coefficients  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\beta }_{12}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{3}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
Similarly the model before  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  is added must contain all coefficients of Eqn. (SeqSSEqnafter) except  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt; . This model can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{13}}&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})= &amp;amp; S{{S}_{R}}\text{ for Eqn}\text{.()}-S{{S}_{R}}\text{ for Eqn}\text{.()} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the present case,  &amp;lt;math&amp;gt;{{\beta }_{2}}=[{{\beta }_{13}}{]}&#039;&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\beta }_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}&#039;&amp;lt;/math&amp;gt; . It can be noted that for the sequential sum of squares  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  contains all coefficients proceeding the coefficient being tested.&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This example illustrates the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test using the sequential sum of squares. The test is conducted for the coefficient  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  corresponding to the predictor variable  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  for the data in Table 5.1. The regression model used for this data set in Example 1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The null hypothesis to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statistic to test this hypothesis is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  represents the sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;  represents the number of degrees of freedom for  &amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})&amp;lt;/math&amp;gt;  (which is one because there is just one coefficient,  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; , being tested) and  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  is the error mean square that can obtained using Eqn. (ErrorMeanSquare) and has been calculated in Example 2 as 30.24. &lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is the difference between the regression sum of squares for the model after adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , and the regression sum of squares for the model before adding  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt; .&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  is obtained as shown next. First the design matrix for this model,  &amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; , is obtained by dropping the third column in the design matrix for the full model,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  (the full design matrix,  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; , was obtained in Example 1). The third column of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  corresponds to coefficient  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  which is no longer used in the present model. Therefore, the design matrix for the model,  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 41.9  \\&lt;br /&gt;
   1 &amp;amp; 43.4  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The hat matrix corresponding to this design matrix is  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt; . It can be calculated using  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }&amp;lt;/math&amp;gt; . Once  &amp;lt;math&amp;gt;{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}&amp;lt;/math&amp;gt;  is known, the regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon &amp;lt;/math&amp;gt;  can be calculated using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})= &amp;amp; {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.16.png|thumb|center|500px|Sequential sum of squares for the data in Table 5.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for the model  &amp;lt;math&amp;gt;Y={{\beta }_{0}}+\epsilon &amp;lt;/math&amp;gt;  is equal to zero since this model does not contain any variables. Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}({{\beta }_{0}})=0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sequential sum of squares for  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})= &amp;amp; S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85-0 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12530.85  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sequential sum of squares, the statistic to test the significance of  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{1}})/r}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12530.85/1}{30.24} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 414.366  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to this statistic based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 8.46\times {{10}^{-12}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
       &lt;br /&gt;
Assuming that the desired significance is 0.1, since  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it can be concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is significant. The test for  &amp;lt;math&amp;gt;{{\beta }_{2}}&amp;lt;/math&amp;gt;  can be carried out in a similar manner. This result is shown in Figure SequentialSshot.&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Chapter 4.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on the regression coefficient,  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low CI and High CI columns as shown in Figure RegrInfoSshot.&lt;br /&gt;
Confidence Interval on Fitted Values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; &lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is given by:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{i}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{i1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{ik}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In Example 1 (Section 5.MatrixApproach), the fitted value corresponding to the fifth observation was calculated as  &amp;lt;math&amp;gt;{{\hat{y}}_{5}}=266.3&amp;lt;/math&amp;gt; . The 90% confidence interval on this value can be obtained as shown in Figure CIfittedvalueSshot. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation in Table 5.1. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.17.png|thumb|center|500px|Confidence interval for the fitted value corresponding to the fifth observation in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
As explained in Chapter 4, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{p}}=\left[ \begin{matrix}&lt;br /&gt;
   1  \\&lt;br /&gt;
   {{x}_{p1}}  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   {{x}_{pk}}  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{x}_{p1}}&amp;lt;/math&amp;gt; ,...,  &amp;lt;math&amp;gt;{{x}_{pk}}&amp;lt;/math&amp;gt;  are the levels of the predictor variables at which the new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , needs to be obtained.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.18.png|thumb|center|400px|Predicted values and region of model application in multiple linear regression.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider Figure JointRegion where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of first predictor variable,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , and  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; th level of the second predictor variable,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; , does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Multiple Determination,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;===&lt;br /&gt;
&lt;br /&gt;
The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}}{S{{S}_{T}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicates the amount of total variability explained by the regression model. The positive square root of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is called the multiple correlation coefficient and measures the linear association between  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  and the predictor variables,  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{x}_{k}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  statistic defined as follows: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{adj}^{2}= &amp;amp; 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The adjusted  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
In DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; and  &amp;lt;math&amp;gt;R_{adj}^{2}&amp;lt;/math&amp;gt;  values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Chapter 4, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents the &amp;quot;standard error of the model.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , in Eqn. (ErrorSumofSquares). The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , for a particular observation,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained by fitting the regression model to the remaining observations. Then the value for a new observation,  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , corresponding to the observation in question,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , is obtained based on the new regression model. The difference between  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  gives  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; . The PRESS residual,  &amp;lt;math&amp;gt;{{e}_{(i)}}&amp;lt;/math&amp;gt; , can also be obtained using  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; , the diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; , as follows:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
R-sq(pred), also referred to as prediction  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; , is obtained using PRESS as shown next:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. Figure RSqadjSshot. shows these values for the data in Table 5.1. The values indicate that the regression model fits the data well and also predicts well.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
&lt;br /&gt;
Plots of residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , similar to the ones discussed in the previous chapter for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals,  &amp;lt;math&amp;gt;{{d}_{i}}&amp;lt;/math&amp;gt; , are obtained using the following equation: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{d}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.19.png|thumb|center|400px|Coefficient of multiple determination and related results for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals,  &amp;lt;math&amp;gt;{{r}_{i}}&amp;lt;/math&amp;gt; , are used in their place. Studentized residuals are calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{i}}= &amp;amp; \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th diagonal element of the hat matrix,  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; . External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Section 5.Rsquare. The reason for using the external studentized residuals is that if the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier, it may influence the fitted model. In this case, the residual  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt;  will be small and may not disclose that  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation is an outlier. The external studentized residual for the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th observation,  &amp;lt;math&amp;gt;{{t}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Residual values for the data of Table 5.1 are shown in Figure ResidualSshot. These values are available using the Diagnostics icon in the Control Panel. Standardized residual plots for the data are shown in Figures Res1NPP to ResVsRuns. DOE++ compares the residual values to the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution for studentized and external studentized residuals. For other residuals the normal distribution is used. For example, for the data in Table 5.1, the critical values on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution at a significance of 0.1 are  &amp;lt;math&amp;gt;{{t}_{0.05,14}}=1.761&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;-{{t}_{0.05,14}}=-1.761&amp;lt;/math&amp;gt;  (as calculated in Example 3, Section 5.tTest). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in Figures ResVsFitted and ResVsRuns.&lt;br /&gt;
&lt;br /&gt;
===Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  Observations===&lt;br /&gt;
&lt;br /&gt;
Residuals help to identify outlying  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  observations. Outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix,  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt; . The  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  values always lie between 0 and 1. Values of  &amp;lt;math&amp;gt;{{h}_{ii}}&amp;lt;/math&amp;gt;  greater than  &amp;lt;math&amp;gt;2(k+1)/n&amp;lt;/math&amp;gt;  are considered to be indicators of outlying  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Influential Observations Detection===&lt;br /&gt;
&lt;br /&gt;
Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook&#039;s distance measure which is computed as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To use Cook&#039;s distance measure, the  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are compared to percentile values on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(k+1,n-(k+1))&amp;lt;/math&amp;gt;  degrees of freedom. If the percentile value is less than 10 or 20 percent, then the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case is influential, and fitted values with and without the  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th case will differ substantially.[Kutner]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 6&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data in Table 5.1. The remaining values along with the leverage values are shown in Figure CookSshot.&lt;br /&gt;
The standardized residual corresponding to the first observation is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.20.png|thumb|center|400px|Residual values for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.21.png|thumb|center|400px|Residual probability plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.22.png|thumb|center|400px|Residual versus fitted values plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.23.png|thumb|center|400px|Residual versus run order plot for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{r}_{1}}= &amp;amp; \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.2804  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Cook&#039;s distance measure for the first observation can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{D}_{1}}= &amp;amp; \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.01  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The 50th percentile value for  &amp;lt;math&amp;gt;{{F}_{3,14}}&amp;lt;/math&amp;gt;  is 0.83. Since all  &amp;lt;math&amp;gt;{{D}_{i}}&amp;lt;/math&amp;gt;  values are less than this value there are no influential observations. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.24.png|thumb|center|400px|Leverage and Cook&#039;s distance measure for the data in Table 5.1.]]&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit test for simple linear regression discussed in Chapter 4 may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  replicates may be collected as follows for all  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of the predictor variables:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{     }m\text{ repeated observations at the first level } \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{     }m\text{ repeated observations at the second level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{       }m\text{ repeated observations at the }i\text{th level} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{    }m\text{ repeated observations at the }n\text{th level }  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , can be obtained as discussed in the previous chapter as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; , can be obtained as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  are:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-(k+1)-(nm-n)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{F}_{0}}= &amp;amp; \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Other Topics in Multiple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
===Polynomial Regression Models===&lt;br /&gt;
&lt;br /&gt;
Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The reason for using coded predictor variables is that many times  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}^{2}}&amp;lt;/math&amp;gt;  are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the  &amp;lt;math&amp;gt;{{({{X}^{\prime }}X)}^{-1}}&amp;lt;/math&amp;gt;  matrix to obtain the estimates,  &amp;lt;math&amp;gt;\hat{\beta }&amp;lt;/math&amp;gt; , of the regression coefficients using Eqn. (LeastSquareEstimate).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Qualitative Factors===&lt;br /&gt;
&lt;br /&gt;
The multiple linear regression model also supports the use of qualitative factors.  For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{1}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   1\text{      Female}  \\&lt;br /&gt;
   0\text{      Male}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In general ( &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; ) indicator variables are required to represent a qualitative factor with  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{     Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{     Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=0\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{x}_{1}}= &amp;amp; 1,\text{   }{{x}_{2}}=0\text{           Machine Type I} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; 0,\text{   }{{x}_{2}}=1\text{           Machine Type II} \\ &lt;br /&gt;
 &amp;amp; {{x}_{1}}= &amp;amp; -1,\text{   }{{x}_{2}}=-1\text{     Machine Type III}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Indicator variables are also referred to as dummy variables or binary variables.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 7&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Consider data from two types of reactors of a chemical process shown in Table 5.3 where the yield values are recorded for various levels of factor  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; . Assuming there are no interactions between the reactor type and  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; , a regression model can be fitted to this data as shown next.&lt;br /&gt;
Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{x}_{2}}=\{\begin{array}{*{35}{l}}&lt;br /&gt;
   0\text{      Reactor Type I}  \\&lt;br /&gt;
   1\text{      Reactor Type II}  \\&lt;br /&gt;
\end{array}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet5.3.png|thumb|center|400px|Yield data from the two types of reactors for a chemical process.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Data entry in DOE++ for this example is shown in Figure IndiVarDesignSshot. The regression model for this data is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;  matrices for the given data are:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.25.png|thumb|center|400px|Data from Table 5.3 as entered in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The estimated regression coefficients for the model can be obtained using Eqn. (LeastSquareEstimate) as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; \hat{\beta }= &amp;amp; {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \left[ \begin{matrix}&lt;br /&gt;
   153.7  \\&lt;br /&gt;
   2.4  \\&lt;br /&gt;
   -27.5  \\&lt;br /&gt;
\end{matrix} \right]  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Note that since  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  in a two dimensional space (because the resulting surface plot will be meaningless for the dimension in  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ). To illustrate this, a scatter plot of the data in Table 5.3 against  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  is shown in Figure IndiVarScatterPlot. It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test of Section 5.FtestPartial (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table of Figure IndiVarResultsSshot. The results show that  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  (reactor type) contributes significantly to the fitted regression model.&lt;br /&gt;
&lt;br /&gt;
===Multicollinearity===&lt;br /&gt;
&lt;br /&gt;
At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables.&lt;br /&gt;
Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test on the individual coefficient or looking at the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.26.png|thumb|center|400px|Scatter plot of the observed yield values in Table 5.3 against &amp;lt;math&amp;gt;x_2 &amp;lt;/math&amp;gt; (reactor type)]]&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.27.png|thumb|center|400px|DOE++ results for the data in Table 5.3.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Multicollinearity can be detected using the variance inflation factor (abbreviated  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt; ).  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  for a coefficient  &amp;lt;math&amp;gt;{{\beta }_{j}}&amp;lt;/math&amp;gt;  is defined as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;VIF=\frac{1}{(1-R_{j}^{2})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;R_{j}^{2}&amp;lt;/math&amp;gt;  is the coefficient of multiple determination resulting from regressing the  &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; th predictor variable,  &amp;lt;math&amp;gt;{{x}_{j}}&amp;lt;/math&amp;gt; , on the remaining  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -1 predictor variables. Mean values of  &amp;lt;math&amp;gt;VIF&amp;lt;/math&amp;gt;  considerably greater than 1 indicate multicollinearity problems.&lt;br /&gt;
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 8&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Variance inflation factors can be obtained for the data in Table 5.1. To calculate the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  has to be calculated.  &amp;lt;math&amp;gt;R_{1}^{2}&amp;lt;/math&amp;gt;  is the coefficient of determination for the model when  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is regressed on the remaining variables. In the case of this example there is just one remaining variable which is  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; . If a regression model is fit to the data, taking  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  as the response variable and  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt;  as the predictor variable, then the design matrix and the vector of observations are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{X}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   1 &amp;amp; 29.1  \\&lt;br /&gt;
   1 &amp;amp; 29.3  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   . &amp;amp; .  \\&lt;br /&gt;
   1 &amp;amp; 32.9  \\&lt;br /&gt;
\end{matrix} \right]\text{     }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix}&lt;br /&gt;
   41.9  \\&lt;br /&gt;
   43.4  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   .  \\&lt;br /&gt;
   77.8  \\&lt;br /&gt;
\end{matrix} \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The regression sum of squares for this model can be obtained using Eqn. (RegressionSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1988.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}&amp;lt;/math&amp;gt;  is the hat matrix (and is calculated using  &amp;lt;math&amp;gt;{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }&amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;J&amp;lt;/math&amp;gt;  is the matrix of ones. The total sum of squares for the model can be calculated using Eqn. (TotalSumofSquares) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2182.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt;  is the identity matrix. Therefore: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; R_{1}^{2}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1988.6}{2182.9} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.911  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then the variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{1}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; VI{{F}_{1}}= &amp;amp; \frac{1}{(1-R_{1}^{2})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{1}{1-0.911} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 11.2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The variance inflation factor for  &amp;lt;math&amp;gt;{{x}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;VI{{F}_{2}}&amp;lt;/math&amp;gt; , can be obtained in a similar manner. In DOE++, the variance inflation factors are displayed in the VIF column of the Regression Information Table as shown in Figure VIFSshot. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data in Table 5.1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe5.28.png|thumb|center|400px|Variance inflation factors for the data in Table 5.1.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33478</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33478"/>
		<updated>2012-08-23T06:04:49Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Lack-of-Fit Test */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\beta }}}_{1}} &amp;amp;= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align} &lt;br /&gt;
  {{{\hat{\beta }}}_{0}} &amp;amp;= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{T}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \end{align} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align} &lt;br /&gt;
  S{{S}_{R}} &amp;amp;= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \end{align}&lt;br /&gt;
  &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{E}} &amp;amp;= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  \end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{PE}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   dof(S{{S}_{PE}}) &amp;amp; = &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{PE}} &amp;amp; = &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33477</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33477"/>
		<updated>2012-08-23T06:03:13Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Lack-of-Fit Test */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\beta }}}_{1}} &amp;amp;= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align} &lt;br /&gt;
  {{{\hat{\beta }}}_{0}} &amp;amp;= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{T}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \end{align} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align} &lt;br /&gt;
  S{{S}_{R}} &amp;amp;= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  \end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{PE}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   dof(S{{S}_{PE}}) &amp;amp; = &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{PE}} &amp;amp; = &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33475</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33475"/>
		<updated>2012-08-23T05:58:46Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Lack-of-Fit Test */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\beta }}}_{1}} &amp;amp;= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align} &lt;br /&gt;
  {{{\hat{\beta }}}_{0}} &amp;amp;= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{T}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{PE}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   dof(S{{S}_{PE}}) &amp;amp; = &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{PE}} &amp;amp; = &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33472</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33472"/>
		<updated>2012-08-23T05:52:06Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Lack-of-Fit Test */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\beta }}}_{1}} &amp;amp;= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  S{{S}_{T}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{PE}} &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   dof(S{{S}_{PE}}) &amp;amp; = &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   M{{S}_{PE}} &amp;amp; = &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33470</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33470"/>
		<updated>2012-08-23T05:50:36Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Lack-of-Fit Test */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{\beta }}}_{1}} &amp;amp;= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33148</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=33148"/>
		<updated>2012-08-22T23:47:10Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Fitted Regression Line */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32953</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32953"/>
		<updated>2012-08-22T03:43:19Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Confidence Interval on New Observations */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   {{{\hat{y}}}_{p}}&amp;amp;=&amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32952</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32952"/>
		<updated>2012-08-22T03:42:49Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* F Test */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{T}}&amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{R}} &amp;amp;=&amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   S{{S}_{E}}&amp;amp;= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  {{f}_{0}}&amp;amp;=&amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp;=&amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
   p\text{ }value &amp;amp;=&amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32889</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32889"/>
		<updated>2012-08-22T00:11:08Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* t  Tests */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32853</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32853"/>
		<updated>2012-08-21T05:57:09Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* t  Tests */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;(6) &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
(7)&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0)= \sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1)= \sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2\times (1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 \times (1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32852</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32852"/>
		<updated>2012-08-21T05:52:17Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Calculation of the Fitted Line Using Least Square Estimates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21}= y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;(6) &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
(7)&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32851</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32851"/>
		<updated>2012-08-21T05:51:27Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Calculation of the Fitted Line Using Least Square Estimates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21} &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;(6) &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
(7)&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32850</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32850"/>
		<updated>2012-08-21T05:50:36Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Calculation of the Fitted Line Using Least Square Estimates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
:: &amp;lt;math&amp;gt;= 17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21} &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;(6) &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
(7)&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32849</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32849"/>
		<updated>2012-08-21T05:49:49Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Calculation of the Fitted Line Using Least Square Estimates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
::= &amp;lt;math&amp;gt;17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21} &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;(6) &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
(7)&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32848</id>
		<title>Simple Linear Regression Analysis</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Simple_Linear_Regression_Analysis&amp;diff=32848"/>
		<updated>2012-08-21T05:48:55Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Calculation of the Fitted Line Using Least Square Estimates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|3}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
&lt;br /&gt;
Regression analysis is a statistical technique that attempts to explore and model the relationship between two or more variables. For example, an analyst may want to know if there is a relationship between road accidents and the age of the driver. Regression analysis forms an important part of the statistical analysis of the data obtained from designed experiments and is discussed briefly in this chapter. Every experiment analyzed in DOE++ includes regression results for each of the responses. These results, along with the results from the analysis of variance (explained in our &amp;quot;Analysis of Experiments&amp;quot; discussion), provide information that is useful to identify significant factors in an experiment and explore the nature of the relationship between these factors and the response. Regression analysis forms the basis for all DOE++ calculations related to the sum of squares used in the analysis of variance. The reason for this is explained in the last section of Chapter 6, Use of Regression to Calculate Sum of Squares. Additionally, DOE++ also includes a regression tool to see if two or more variables are related, and to explore the nature of the relationship between them. This chapter discusses simple linear regression analysis while Chapter 5 focuses on multiple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
==Simple Linear Regression Analysis== &lt;br /&gt;
&lt;br /&gt;
A linear regression model attempts to explain the relationship between two or more variables using a straight line. Consider the data obtained from a chemical process where the yield of the process is thought to be related to the reaction temperature (see Table 4.1). This data can be entered in DOE++ as shown in Figure 4.1 and a scatter plot can be obtained as shown in Figure 4.2. [Note] In the scatter plot yield, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt; is plotted for different temperature values, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; . It is clear that no line can be found to pass through all points of the plot. Thus no functional relation exists between the two variables &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;. [Note] However, the scatter plot does give an indication that a straight line may exist such that all the points on the plot are scattered randomly around this line. A statistical relation is said to exist in this case. The statistical relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; may be expressed as follows:&lt;br /&gt;
(1)&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=\beta_0+\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
[[Image:doet4.1.png|thumb|center|300px|Yield data observations of a chemical process at different values of reaction temperature.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.1.png|thumb|center|300px|Data entry in DOE++ for the observations in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.2.png|thumb|center|300px|Scatter plot for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
Eqn. (1) is the linear regression model that can be used to explain the relation between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; that is seen on the scatter plot above. In this model, the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; (abbreviated as &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;) is assumed to follow the linear relation &amp;lt;math&amp;gt;\beta_0=\beta_1{x} &amp;lt;/math&amp;gt;:  &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;E(Y)=\beta_0+\beta_1{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The actual values of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;, (which are observed as yield from the chemical process from time to time and are random in nature), are assumed to be the sum of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt; , and a random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Y=E(Y)+\epsilon &amp;lt;/math&amp;gt; &lt;br /&gt;
::&amp;lt;math&amp;gt;=\beta_0=\beta_1{x}+\epsilon &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The regression model here is called a &#039;&#039;simple&#039;&#039; linear regression model because there is just one independent variable, &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; , in the model. In regression models, the independent variables are also referred to as regressors or predictor variables. The dependent variable, &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; , is also referred to as the response. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , and the intercept, &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; , of the line &amp;lt;math&amp;gt;E(Y)=\beta_0=\beta_1{x} &amp;lt;/math&amp;gt; are called regression coefficients. The slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , can be interpreted as the change in the mean value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; for a unit change in &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
The random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is assumed to follow the normal distribution with a mean of 0 and variance of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Since &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is the sum of this random term and the mean value, &amp;lt;math&amp;gt;E(Y)&amp;lt;/math&amp;gt; , (which is a constant), the variance of  &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; is also &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. Therefore, at any given value of &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;, say &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt;, the dependent variable &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; follows a normal distribution with a mean of &amp;lt;math&amp;gt;\beta_0+\beta_1{x_i} &amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma^2 &amp;lt;/math&amp;gt;. This is illustrated in the following figure.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.3.png|thumb|center|400px|The normal distribution of  for two values of . Also shown is the true regression line and the values of the random error term, , corresponding to the two  values. The true regression line and  are usually not known.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
===Fitted Regression Line===&lt;br /&gt;
The true regression line corresponding to Eqn. (1) is usually never known. However, the regression line can be estimated by estimating the coefficients &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\beta_0 &amp;lt;/math&amp;gt; for an observed data set. The estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are calculated using least squares. (For details on least square estimates refer to [19]). The estimated regression line, obtained using the values of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, is called the fitted line. The least square estimates, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt;, are obtained using the following equations:(2) &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;=&amp;lt;math&amp;gt;\frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0=\bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
(3)&lt;br /&gt;
 &lt;br /&gt;
where &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is the mean of all the observed values and &amp;lt;math&amp;gt;\bar{x} &amp;lt;/math&amp;gt; is the mean of all values of the predictor variable at which the observations were taken. &amp;lt;math&amp;gt;\bar{y} &amp;lt;/math&amp;gt; is calculated using  &amp;lt;math&amp;gt;\bar{y}=(1/n)\sum)_{i=1}^n y_i &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\bar{x}=(1/n)\sum)_{i=1}^n x_i &amp;lt;/math&amp;gt; is calculated using .&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Once &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; are known, the fitted regression line can be written as:&lt;br /&gt;
(4)&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{y} &amp;lt;/math&amp;gt; is the fitted or estimated value based on the fitted regression model. It is an estimate of the mean value, &amp;lt;math&amp;gt;E(Y) &amp;lt;/math&amp;gt;. The fitted value,&amp;lt;math&amp;gt;\widehat{y}_i &amp;lt;/math&amp;gt; , for a given value of the predictor variable, &amp;lt;math&amp;gt;x_i &amp;lt;/math&amp;gt; , may be different from the corresponding observed value, &amp;lt;math&amp;gt;y_i &amp;lt;/math&amp;gt;. The difference between the two values is called the residual, &amp;lt;math&amp;gt;e_i &amp;lt;/math&amp;gt;: (5)&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_i=y_i-\widehat{y}_i &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Calculation of the Fitted Line Using Least Square Estimates====&lt;br /&gt;
The least square estimates of the regression coefficients can be obtained for the data in Table 4.1 using the Eqns. (2) and (3) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_1 = \frac{\sum_{i=1}^n y_i x_i- \frac{(\sum_{i=1}^n y_i) (\sum_{i=1}^n x_i)}{n}}{\sum_{i=1}^n (x_i-\bar{x})^2} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{322516-\frac{4158 x 1871}{25}}{5697.36} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=1.9952 \approx 2.00 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{\beta}_0 =  \bar{y}-\widehat{\beta}_1 \bar{x} &amp;lt;/math&amp;gt;&lt;br /&gt;
::= &amp;lt;math&amp;gt;166.32 - 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;74.84 &amp;lt;/math&amp;gt;&lt;br /&gt;
::= &amp;lt;math&amp;gt;17.0016 \approx 17.00&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing &amp;lt;math&amp;gt;\widehat{\beta}_0 &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, the fitted regression line is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}=\widehat{\beta}_0+\widehat{\beta}_1 x &amp;lt;/math&amp;gt;&lt;br /&gt;
::= &amp;lt;math&amp;gt;17.0016+1.9952 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\approx 17+2 x &amp;lt;/math&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
This line is shown in Figure 4.4.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.4.png|thumb|center|400px|Fitted regression line for the data in Table 4.1. Also shown is the residual for the 21st observation.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
Once the fitted regression line is known, the fitted value of &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; corresponding to any observed data point can be calculated. For example, the fitted value corresponding to the 21st observation in Table 4.1 is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\widehat{y}_{21} = \widehat{\beta}_0 = \widehat{\beta}_1 x_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;(17.0016) + (1.9952) &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;93 &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The observed response at this point is &amp;lt;math&amp;gt;y_{21}=194 &amp;lt;/math&amp;gt;. Therefore, the residual at this point is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;e_{21} &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;y_{21}-\widehat{y}_{21} &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;194-202.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
::=&amp;lt;math&amp;gt;-8.6 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In DOE++, fitted values and residuals are available using the Diagnostic icon in the Control Panel. The values are shown in Figure 4.5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.5.png|thumb|center|400px|Fitted values and residuals for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Tests in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
The following sections discuss hypothesis tests on the regression coefficients in simple linear regression. These tests can be carried out if it can be assumed that the random error term, &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , is normally and independently distributed with a mean of zero and variance of &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; Tests===&lt;br /&gt;
&lt;br /&gt;
The  tests are used to conduct hypothesis tests on the regression coefficients obtained in simple linear regression. A statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution is used to test the two-sided hypothesis that the true slope, &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; , equals some constant value, &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt;. [Note] The statements for the hypothesis test are expressed as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;H_0&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_1 = \beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;H_1&amp;lt;/math&amp;gt; : &amp;lt;math&amp;gt;\beta_{1}\ne\beta_{1,0} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic used for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_1)} &amp;lt;/math&amp;gt;(6) &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; is the least square estimate of &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; is its standard error. The value of &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
(7)&lt;br /&gt;
&lt;br /&gt;
The test statistic, &amp;lt;math&amp;gt;T_0 &amp;lt;/math&amp;gt; , follows a &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with &amp;lt;math&amp;gt;(n-2) &amp;lt;/math&amp;gt; degrees of freedom, where &amp;lt;math&amp;gt;n &amp;lt;/math&amp;gt; is the total number of observations. The null hypothesis, &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, is rejected if the calculated value of the test statistic is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-t_{\alpha/2,n-2}&amp;lt;T_0&amp;lt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; are the critical values for the two-sided hypothesis. &amp;lt;math&amp;gt;t_{\alpha/2,n-2} &amp;lt;/math&amp;gt; is the percentile of the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution corresponding to a cumulative probability of (&amp;lt;math&amp;gt;(1-\alpha/2) &amp;lt;/math&amp;gt;) and &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; is the significance level. &lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
If the value of &amp;lt;math&amp;gt;\beta_{1,0} &amp;lt;/math&amp;gt; used in Eqn. (6) is zero, then the hypothesis tests for the significance of regression. In other words, the test indicates if the fitted regression model is of value in explaining variations in the observations or if you are trying to impose a regression model when no true relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Failure to reject &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; implies that no linear relationship exists between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. This result may be obtained when the scatter plots of  against  are as shown in 4.6 (a) and (b) of the following figure. Figure 4.6 (a) represents the case where no model exits for the observed data. In this case you would be trying to fit a regression model to noise or random variation. Figure 4.6 (b) represents the case where the true relationship between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt; is not linear. Figure 4.6 (c) and (d) represent the case when &amp;lt;math&amp;gt;H_0:\beta_1=0 &amp;lt;/math&amp;gt; is rejected, implying that a model does exist between &amp;lt;math&amp;gt;x &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y &amp;lt;/math&amp;gt;. Figure 4.6 (c) represents the case where the linear model is sufficient. Figure 4.6, (d) represents the case where a higher order model may be needed.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.6.png|thumb|center|400px|Possible scatter plots of  against . Plots (a) and (b) represent cases when  is not rejected. Plots (c) and (d) represent cases when  is rejected.]]&lt;br /&gt;
 &lt;br /&gt;
  &lt;br /&gt;
A similar procedure can be used to test the hypothesis on the intercept . The test statistic used in this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_0=\frac{\widehat{\beta}_0-\beta_{0,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;(8)&lt;br /&gt;
&lt;br /&gt;
where  is the least square estimate of , and  is its standard error which is calculated using:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_0) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\displaystyle\sum_{i=1}^n e_i^2}{n-2} \Bigg[ \frac{1}{n}+\frac{\bar{x}^2}{\displaystyle\sum_{i=1}^n (x_i-\bar{x})^2} \Bigg]} &amp;lt;/math&amp;gt;&lt;br /&gt;
(9)&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Example 4.1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The test for the significance of regression for the data in Table 4.1 is illustrated in this example. The test is carried out using the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test on the coefficient &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;. The hypothesis to be tested is &amp;lt;math&amp;gt;H_0 : \beta_1 = 0 &amp;lt;/math&amp;gt;. To calculate the statistic to test &amp;lt;math&amp;gt;H_0 &amp;lt;/math&amp;gt;, the estimate, &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt;, and the standard error, &amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt;, are needed. The value of &amp;lt;math&amp;gt;\widehat{\beta}_1 &amp;lt;/math&amp;gt; was obtained in Chapter 4, Fitted Regression Line. The standard error can be calculated using Eqn. (7) as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;se(\widehat{\beta}_1) &amp;lt;/math&amp;gt; = &amp;lt;math&amp;gt;\sqrt{\frac{\frac{\displaystyle \sum_{i=1}^n e_i^2}{n-2}}{\displaystyle \sum_{i=1}^n (x_i-\bar{x})^2}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = \sqrt{\frac{(371.627/23)}{5679.36}} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt; = 0.0533 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then, the test statistic can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;t_0=\frac{\widehat{\beta}_1-\beta_{1,0}}{se(\widehat{\beta}_0)} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=\frac{1.9952-0}{0.0533} &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;=37.4058 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value corresponding to this statistic based on the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution with 23(&amp;lt;math&amp;gt;n-2=25-2=23 &amp;lt;/math&amp;gt;) degrees of freedom can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p value = 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-P(T\le t_0) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 2 &amp;lt;/math&amp;gt; x &amp;lt;math&amp;gt;(1-0.999999) &amp;lt;/math&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;= 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance level is 0.1, since &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value &amp;lt; 0.1, &amp;lt;math&amp;gt;H_0 : \beta_1=0 &amp;lt;/math&amp;gt; is rejected indicating that a relation exists between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure 4.2, it can be concluded that the relationship between temperature and yield is linear.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
In DOE++, information related to the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test is displayed in the Regression Information table as shown in Figure 4.7. In this table the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is displayed in the row for the term Temperature because &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; is the coefficient that represents the variable temperature in the regression model. The columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the  test and the &amp;lt;math&amp;gt;p &amp;lt;/math&amp;gt; value for the &amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; test, respectively. These values have been calculated for &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt; in this example. The Coefficient column represents the estimate of regression coefficients. For &amp;lt;math&amp;gt;\beta_1 &amp;lt;/math&amp;gt;, this value was calculated using Eqn. (2). The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Chapter 7, Two Level Factorial Experiments. Columns Low CI and High CI represent the limits of the confidence intervals for the regression coefficients and are explained in Chapter 4, Confidence Interval on Regression Coefficients. The Variance Inflation Factor column displays values that give a measure of multicollinearity. The concept of multicollinearity is only applicable to multiple linear regression models and is explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.7.png|thumb|center|400px|Regression results for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Analysis of Variance Approach to Test the Significance of Regression===&lt;br /&gt;
&lt;br /&gt;
The analysis of variance (ANOVA) is another method to test for the significance of regression. As the name implies, this approach uses the variance of the observed data to determine if a regression model can be applied to the observed data. The observed variance is partitioned into components that are then used in the test for significance of regression.&lt;br /&gt;
&lt;br /&gt;
====Sum of Squares====&lt;br /&gt;
&lt;br /&gt;
The total variance (i.e. the variance of all of the observed data) is estimated using the observed data. As mentioned in Chapter 3, the variance of a population can be estimated using the sample variance, which is calculated using the following relationship:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{y}_{i}}-\bar{y})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The quantity in the numerator of the previous equation is called the sum of squares. It is the sum of the square of deviations of all the observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; , from their mean,  &amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; . In the context of ANOVA this quantity is called the total sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) because it relates to the total variance of the observations. Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The denominator in the relationship of the sample variance is the number of degrees of freedom associated with the sample variance. Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{T}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt; .  The sample variance is also referred to as a mean square because it is obtained by dividing the sum of squares by the respective degrees of freedom. Therefore, the total mean square (abbreviated  &amp;lt;math&amp;gt;M{{S}_{T}}&amp;lt;/math&amp;gt; ) is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}=\frac{S{{S}_{T}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When you attempt to fit a regression model to the observations, you are trying to explain some of the variation of the observations using this model. If the regression model is such that the resulting fitted regression line passes through all of the observations, then you would have a &amp;quot;perfect&amp;quot; model (see Figure PerfectModel (a)). In this case the model would explain all of the variability of the observations. Therefore, the model sum of squares (also referred to as the regression sum of squares and abbreviated  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ) equals the total sum of squares; i.e. the model explains all of the observed variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=S{{S}_{T}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For the perfect model, the regression sum of squares,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , equals the total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; , because all estimated values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , will equal the corresponding observations,  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt;  can be calculated using a relationship similar to the one for obtaining  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt;  by replacing  &amp;lt;math&amp;gt;{{y}_{i}}&amp;lt;/math&amp;gt;  by  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt;  in the relationship of  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; . Therefore:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{R}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{R}})&amp;lt;/math&amp;gt; , is one. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Based on the preceding discussion of ANOVA, a perfect regression model exists when the fitted regression line passes through all observed points. However, this is not usually the case, as seen in Figure PerfectModel (b) or Figure FittedRegressionLine. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.8.png|thumb|center|400px|A perfect regression model will pass through all observed data points as shown in *(a). Most models are imperfect and do not fit perfectly to all data points as shown in (b).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In both of these plots, a number of points do not follow the fitted regression line. This indicates that a part of the total variability of the observed data still remains unexplained. This portion of the total variability or the total sum of squares, that is not explained by the model, is called the residual sum of squares or the error sum of squares (abbreviated  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ). The deviation for this sum of squares is obtained at each observation in the form of the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; . The error sum of squares can be obtained as the sum of squares of these deviations:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,e_{i}^{2}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;dof(S{{S}_{E}})&amp;lt;/math&amp;gt; , is  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The total variability of the observed data (i.e. total sum of squares,  &amp;lt;math&amp;gt;S{{S}_{T}}&amp;lt;/math&amp;gt; ) can be written using the portion of the variability explained by the model,  &amp;lt;math&amp;gt;S{{S}_{R}}&amp;lt;/math&amp;gt; , and the portion unexplained by the model,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{T}}=S{{S}_{R}}+S{{S}_{E}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The above equation is also referred to as the analysis of variance identity and can be expanded as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{\hat{y}}_{i}}-\bar{y})}^{2}}+\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{\hat{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.9.png|thumb|center|400px|Scatter plots showing the deviations for the sum of squares used in ANOVA. (a) shows deviations for , (b) shows deviations for , and (c) shows deviations for .]]&lt;br /&gt;
&lt;br /&gt;
====Mean Squares====&lt;br /&gt;
&lt;br /&gt;
As mentioned previously, mean squares are obtained by dividing the sum of squares by the respective degrees of freedom. For example, the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error mean square is an estimate of the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of the random error term,  &amp;lt;math&amp;gt;\epsilon &amp;lt;/math&amp;gt; , and can be written as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\sigma }}^{2}}=\frac{S{{S}_{E}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the regression mean square,  &amp;lt;math&amp;gt;M{{S}_{R}}&amp;lt;/math&amp;gt; , can be obtained by dividing the regression sum of squares by the respective degrees of freedom as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}=\frac{S{{S}_{R}}}{1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====F Test====&lt;br /&gt;
&lt;br /&gt;
To test the hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt; , the statistic used is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. It can be shown that if the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true, then the statistic:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}=\frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;  degree of freedom in the numerator and  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator.  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected if the calculated statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{f}_{\alpha ,1,n-2}}&amp;lt;/math&amp;gt;  is the percentile of the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution corresponding to a cumulative probability of ( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) and  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is the significance level.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The analysis of variance approach to test the significance of regression can be applied to the yield data in Table 4.1. To calculate the statistic,  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt; , for the test, the sum of squares have to be obtained. The sum of squares can be calculated as shown next.&lt;br /&gt;
The total sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22979.44  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The regression sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-166.32)}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 22607.81  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 371.63  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing the sum of squares, the statistic to test  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{S{{S}_{R}}/1}{S{{S}_{E}}/(n-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81/1}{371.63/(25-2)} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1399.20  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value at a significance level of 0.1 is  &amp;lt;math&amp;gt;{{f}_{0.05,1,23}}=2.937&amp;lt;/math&amp;gt; . Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;gt;{{f}_{\alpha ,1,n-2}},&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected and it is concluded that  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is not zero. Alternatively, the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value can also be used. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value corresponding to the test statistic,  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;/math&amp;gt; , based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with one degree of freedom in the numerator and 23 degrees of freedom in the denominator is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.999999 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 4.17E-22  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assuming that the desired significance is 0.1, since the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value &amp;lt; 0.1, then  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\beta }_{1}}=0&amp;lt;/math&amp;gt;  is rejected, implying that a relation does exist between temperature and yield for the data in Table 4.1. Using this result along with the scatter plot of Figure ScatterPlotSshot, it can be concluded that the relationship that exists between temperature and yield is linear. This result is displayed in the ANOVA table as shown in Figure Ex2ANOVAtableSshot. Note that this is the same result that was obtained from the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test in Section 4.tTest. The ANOVA and Regression Information tables in DOE++ represent two different ways to test for the significance of the regression model. In the case of multiple linear regression models these tables are expanded to allow tests on individual variables used in the model. This is done using extra sum of squares. Multiple linear regression models and the application of extra sum of squares in the analysis of these models are discussed in Chapter 5. The term Partial appearing in Figure Ex2ANOVAtableSshot relates to the extra sum of squares and is also explained in Chapter 5.&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.10.png|thumb|center|400px|ANOVA table for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
==Confidence Intervals in Simple Linear Regression==&lt;br /&gt;
&lt;br /&gt;
A confidence interval represents a closed interval where a certain percentage of the population is likely to lie. For example, a 90% confidence interval with a lower limit of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and an upper limit of  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  implies that 90% of the population lies between the values of  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . Out of the remaining 10% of the population, 5% is less than  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and 5% is greater than  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; . (For details refer to [LDAReference]). This section discusses confidence intervals used in simple linear regression analysis.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Regression Coefficients===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{1}}&amp;lt;/math&amp;gt;  is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{1}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, a 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on  &amp;lt;math&amp;gt;{{\beta }_{0}}&amp;lt;/math&amp;gt;  is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}\pm {{t}_{\alpha /2,n-2}}\cdot se({{\hat{\beta }}_{0}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on Fitted Values===&lt;br /&gt;
&lt;br /&gt;
A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent confidence interval on any fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It can be seen that the width of the confidence interval depends on the value of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and will be a minimum at  &amp;lt;math&amp;gt;{{x}_{i}}=\bar{x}&amp;lt;/math&amp;gt;  and will widen as  &amp;lt;math&amp;gt;\left| {{x}_{i}}-\bar{x} \right|&amp;lt;/math&amp;gt;  increases.&lt;br /&gt;
&lt;br /&gt;
===Confidence Interval on New Observations===&lt;br /&gt;
&lt;br /&gt;
For the data in Table 4.1, assume that a new value of the yield is observed after the regression model is fit to the data. This new observation is independent of the observations used to obtain the regression model. If  &amp;lt;math&amp;gt;{{x}_{p}}&amp;lt;/math&amp;gt;  is the level of the temperature at which the new observation was taken, then the estimate for this new value based on the fitted regression model is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times {{x}_{p}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If a confidence interval needs to be obtained on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt; , then this interval should include both the error from the fitted model and the error associated with future observations. This is because  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  represents the estimate for a value of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  that was not used to obtain the regression model. The confidence interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  is referred to as the prediction interval &amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt;  A 100( &amp;lt;math&amp;gt;1-\alpha &amp;lt;/math&amp;gt; ) percent prediction interval on a new observation is obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To illustrate the calculation of confidence intervals, the 95% confidence intervals on the response at  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  for the data in Table 4.1 is obtained in this example. A 95% prediction interval is also obtained assuming that a new observation for the yield was made at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , corresponding to  &amp;lt;math&amp;gt;x=93&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{21}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{21}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 93 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% confidence interval  &amp;lt;math&amp;gt;(\alpha =0.05)&amp;lt;/math&amp;gt;  on the fitted value,  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}=202.6&amp;lt;/math&amp;gt; , is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{i}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ \frac{1}{n}+\frac{{{({{x}_{i}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.069\sqrt{16.16\left[ \frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 202.6\pm 2.602  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{21}}&amp;lt;/math&amp;gt;  are 199.95 and 205.2, respectively.&lt;br /&gt;
The estimated value based on the fitted regression model for the new observation at  &amp;lt;math&amp;gt;x=91&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{y}}}_{p}}= &amp;amp; {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{p}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 17.0016+1.9952\times 91 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% prediction interval on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}=198.6&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; {{{\hat{y}}}_{p}}\pm {{t}_{\alpha /2,n-2}}\sqrt{{{{\hat{\sigma }}}^{2}}\left[ 1+\frac{1}{n}+\frac{{{({{x}_{p}}-\bar{x})}^{2}}}{\underset{i=1}{\overset{n}{\mathop \sum }}\,{{({{x}_{p}}-\bar{x})}^{2}}} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm {{t}_{0.025,23}}\sqrt{M{{S}_{E}}\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\sqrt{16.16\left[ 1+\frac{1}{25}+\frac{{{(93-74.84)}^{2}}}{5679.36} \right]} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 2.069\times 4.1889 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 198.6\pm 8.67  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The 95% limits on  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  are 189.9 and 207.2, respectively. In DOE++, confidence and prediction intervals are available using the Prediction icon in the Control Panel. The prediction interval values calculated in this example are shown in Figure PredictionInterval as Low PI and High PI respectively. The columns labeled Mean Predicted and Standard Error represent the values of  &amp;lt;math&amp;gt;{{\hat{y}}_{p}}&amp;lt;/math&amp;gt;  and the standard error used in the calculations. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.11.png|thumb|center|400px|Calculation of prediction intervals in DOE++.]]&lt;br /&gt;
&lt;br /&gt;
==Measures of Model Adequacy==&lt;br /&gt;
&lt;br /&gt;
It is important to analyze the regression model before inferences based on the model are undertaken. The following sections present some techniques that can be used to check the appropriateness of the model for the given data. These techniques help to determine if any of the model assumptions have been violated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Coefficient of Determination (&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt;)===&lt;br /&gt;
The coefficient of determination is a measure of the amount of variability in the data accounted for by the regression model. As mentioned previously, the total variability of the data is measured by the total sum of squares, . The amount of this variability explained by the regression model is the regression sum of squares, . The coefficient of determination is the ratio of the regression sum of squares to the total sum of squares.&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt;(22)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can take on values between 0 and 1 since &amp;lt;math&amp;gt;R^2 = \frac{SS_R}{SS_T} &amp;lt;/math&amp;gt; . For the yield data example, &amp;lt;math&amp;gt;R^2 &amp;lt;/math&amp;gt; can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{R}^{2}}= &amp;amp; \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{22607.81}{22979.44} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.98  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  indicate a better fitting regression model. However,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  should be used cautiously as this is not always the case. The value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , for the new model is larger than the  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt;  of the older model, even though the new model will show an increased value of  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt; . In the results obtained from DOE++,  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  is displayed as R-sq under the ANOVA table (as shown in Figure FullAnalysisEx1 which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
The other values displayed with  &amp;lt;math&amp;gt;{{R}^{2}}&amp;lt;/math&amp;gt;  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square,  &amp;lt;math&amp;gt;M{{S}_{E}}&amp;lt;/math&amp;gt; , and represents Therefore, 98% of the variability in the yield data is explained by the regression model, indicating a very good fit of the model. It may appear that larger values of  indicate a better fitting regression model. However,  should be used cautiously as this is not always the case. The value of  increases as more terms are added to the model, even if the new term does not contribute significantly to the model. Therefore, an increase in the value of  cannot be taken as a sign to conclude that the new model is superior to the older model. Adding a new term may make the regression model worse if the error mean square, , for the new model is larger than the  of the older model, even though the new model will show an increased value of . In the results obtained from DOE++,  is displayed as R-sq under the ANOVA table (as shown in Figure 4.12, which displays the complete analysis sheet for the data in Table 4.1).&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The other values displayed with  are S, R-sq(adj), PRESS and R-sq(pred). These values measure different aspects of the adequacy of the regression model. For example, the value of S is the square root of the error mean square, , and represents the &amp;quot;standard error of the model.&amp;quot; A lower value of S indicates a better fitting model. The values of S, R-sq and R-sq(adj) indicate how well the model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. R-sq(adj), PRESS and R-sq(pred) are explained in Chapter 5, Multiple Linear Regression Analysis.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.12.png|thumb|center|400px|Complete analysis for the data in Table 4.1.]]&lt;br /&gt;
&lt;br /&gt;
===Residual Analysis===&lt;br /&gt;
In the simple linear regression model the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}}&amp;lt;/math&amp;gt; , are never known. The residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; , may be thought of as the observed error terms that are similar to the true error terms. Since the true error terms,  &amp;lt;math&amp;gt;{{\epsilon }_{i}},&amp;lt;/math&amp;gt;  are assumed to be normally distributed with a mean of zero and a variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , in a good model the observed error terms, (i.e. the residuals,  &amp;lt;math&amp;gt;{{e}_{i}}&amp;lt;/math&amp;gt; ,) should also follow these assumptions.  Thus the residuals in the simple linear regression should be normally distributed with a mean of zero and a constant variance of  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; . Residuals are usually plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , against the predictor variable values,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , and against time or run-order sequence, in addition to the normal probability plot. Plots of residuals are used to check for the following:&lt;br /&gt;
 &lt;br /&gt;
:1. Residuals follow the normal distribution. &lt;br /&gt;
:2. Residuals have a constant variance. &lt;br /&gt;
:3. Regression function is linear. &lt;br /&gt;
:4. A pattern does not exist when residuals are plotted in a time or run-order sequence. &lt;br /&gt;
:5. There are no outliers.  &lt;br /&gt;
&lt;br /&gt;
Examples of residual plots are shown in Figure DiffrResidualPlots. The plot of  Figure DiffrResidualPlots (a) is a satisfactory plot with the residuals falling in a horizontal band with no systematic pattern. Such a plot indicates an appropriate regression model. The plot of Figure DiffrResidualPlots (b) shows residuals falling in a funnel shape. Such a plot indicates increase in variance of residuals and the assumption of constant variance is violated here. Transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be helpful in this case (see Section 4.Transformations). If the residuals follow the pattern of Figure DiffrResidualPlots (c) or (d) then this is an indication that the linear regression model is not adequate. Addition of higher order terms to the regression model or transformation on  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may be required in such cases. A plot of residuals may also show a pattern as seen in Figure DiffrResidualPlots (e) indicating that the residuals increase (or decrease) as the run order sequence or time progresses. This may be due to factors such as operator-learning or instrument-creep and should be investigated further. &lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.13.png|thumb|center|300px|Possible residual plots (against fitted values, time or run-order) that can be obtained from simple linear regression analysis.]] &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&#039;&#039;&#039;Example 4.4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Residual plots for the data of Table 4.1 are shown in Figures ResidualNPP to ResidualVsRun. Figure ResidualNPP is the normal probability plot. It can be observed that the residuals follow the normal distribution and the assumption of normality is valid here. In Figure ResidualVsFitted the residuals are plotted against the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , and in Figure ResidualVsRun the residuals are plotted against the run order. Both of these plots show that the 21st observation seems to be an outlier. Further investigations are needed to study the cause of this oulier. &lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe4.14.png|thumb|center|300px|Normal probability plot of residuals for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.15.png|thumb|center|300px|Plot of residuals against fitted values for the data in Table 4.1.]]&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.16.png|thumb|center|300px|Plot of residuals against run order for the data in Table 4.1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Lack-of-Fit Test===&lt;br /&gt;
&lt;br /&gt;
As mentioned in Section 4.ANOVA, a perfect regression model results in a fitted line that passes exactly through all observed data points. This perfect model will give us a zero error sum of squares ( &amp;lt;math&amp;gt;S{{S}_{E}}=0&amp;lt;/math&amp;gt; ). Thus, no error exists for the perfect model. However, if you record the response values for the same values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  for a second time, in conditions maintained as strictly identical as possible to the first time, observations from the second time will not all fall along the perfect model. The deviations in observations recorded for the second time constitute the &amp;quot;purely&amp;quot; random variation or noise. The sum of squares due to pure error (abbreviated  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) quantifies these variations.  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is calculated by taking repeated observations at some or all values of  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  and adding up the square of deviations at each level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  using the respective repeated observations at that  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  value. &lt;br /&gt;
Assume that there are  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations are taken at each  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level. The data is collected as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp;  &amp;amp; {{y}_{11}},{{y}_{12}},....,{{y}_{1{{m}_{1}}}}\text{     repeated observations at }{{x}_{1}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{21}},{{y}_{22}},....,{{y}_{2{{m}_{2}}}}\text{     repeated observations at }{{x}_{2}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{i1}},{{y}_{i2}},....,{{y}_{i{{m}_{i}}}}\text{       repeated observations at }{{x}_{i}} \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; ... \\ &lt;br /&gt;
 &amp;amp;  &amp;amp; {{y}_{n1}},{{y}_{n2}},....,{{y}_{n{{m}_{n}}}}\text{    repeated observations at }{{x}_{n}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sum of squares of the deviations from the mean of the observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt;  is the mean of the  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  repeated observations corresponding to  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;{{\bar{y}}_{i}}=(1/{{m}_{i}})\mathop{}_{j=1}^{{{m}_{i}}}{{y}_{ij}}&amp;lt;/math&amp;gt; ). The number of degrees of freedom for these deviations is ( &amp;lt;math&amp;gt;{{m}_{i}}-1&amp;lt;/math&amp;gt; ) as there are  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  observations at  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; th level of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  but one degree of freedom is lost in calculating the mean,  &amp;lt;math&amp;gt;{{\bar{y}}_{i}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
The total sum of square deviations (or  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; ) for all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  can be obtained by summing the deviations for all  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt;  as shown next:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The total number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,({{m}_{i}}-1) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,{{m}_{i}}-n  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If all  &amp;lt;math&amp;gt;{{m}_{i}}=m&amp;lt;/math&amp;gt; , (i.e.  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; ), then  &amp;lt;math&amp;gt;\mathop{}_{i=1}^{n}{{m}_{i}}=nm&amp;lt;/math&amp;gt;  and the degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  are: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square in this case will be:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{PE}}=\frac{S{{S}_{PE}}}{nm-n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
When repeated observations are used for a perfect regression model, the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , is also considered as the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . For the case when repeated observations are used with imperfect regression models, there are two components of the error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; . One portion is the pure error due to the repeated observations. The other portion is the error that represents variation not captured because of the imperfect model. The second portion is termed as the sum of squares due to lack-of-fit (abbreviated  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; ) to point to the deficiency in fit due to departure from the perfect-fit model. Thus, for an imperfect regression model:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{E}}=S{{S}_{PE}}+S{{S}_{LOF}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , the previous equation can be used to obtain  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained in a similar manner using subtraction. For the case when  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  repeated observations are taken at all levels of  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{PE}})=nm-n&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since there are  &amp;lt;math&amp;gt;nm&amp;lt;/math&amp;gt;  total observations, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;dof(S{{S}_{E}})=nm-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; = &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; n-2  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;M{{S}_{LOF}}=\frac{S{{S}_{LOF}}}{n-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The magnitude of  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;  will provide an indication of how far the regression model is from the perfect model. An  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  test exists to examine the lack-of-fit at a particular significance level.  The quantity  &amp;lt;math&amp;gt;M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt;  follows an  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;(n-2)&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;(nm-n)&amp;lt;/math&amp;gt;  degrees of freedom in the denominator when all  &amp;lt;math&amp;gt;{{m}_{i}}&amp;lt;/math&amp;gt;  equal  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; . The test statistic for the lack-of-fit test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{M{{S}_{LOF}}}{M{{S}_{PE}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the critical value  &amp;lt;math&amp;gt;{{f}_{\alpha ,n-2,mn-n}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}&amp;gt;{{f}_{\alpha ,n-2,nm-n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
it will lead to the rejection of the hypothesis that the model adequately fits the data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that a second set of observations are taken for the yield data of Table 4.1. The resulting observations are recorded in Table 4.2. To conduct a lack-of-fit test on this data, the statistic  &amp;lt;math&amp;gt;{{F}_{0}}=M{{S}_{LOF}}/M{{S}_{PE}}&amp;lt;/math&amp;gt; , can be calculated as shown next.&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.2.png|thumb|center|400px|Yield data from the first and second observation sets for the chemical process example in Section 4.1.]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of Least Square Estimates&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The parameters of the fitted regression model can be obtained using Eqns. (beta0) and (beta1) as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{{\hat{\beta }}}_{1}}= &amp;amp; \frac{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}}{{x}_{i}}-\frac{\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{y}_{i}} \right)\left( \underset{i=1}{\overset{50}{\mathop \sum }}\,{{x}_{i}} \right)}{50}}{\underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{x}_{i}}-\bar{x})}^{2}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{648532-\frac{8356\times 3742}{50}}{11358.72} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 2.04 \\ &lt;br /&gt;
 &amp;amp;  &amp;amp;  \\ &lt;br /&gt;
 &amp;amp; {{{\hat{\beta }}}_{0}}= &amp;amp; \bar{y}-{{{\hat{\beta }}}_{1}}\bar{x} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 167.12-2.04\times 74.84 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14.47  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Knowing  &amp;lt;math&amp;gt;{{\hat{\beta }}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\hat{\beta }}_{0}}&amp;lt;/math&amp;gt; , the fitted values,  &amp;lt;math&amp;gt;{{\hat{y}}_{i}}&amp;lt;/math&amp;gt; , can be calculated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Sum of Squares&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Using the fitted values, the sum of squares can be obtained as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{T}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47907.28 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{R}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{{\hat{y}}}_{i}}-\bar{y})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 47258.91 \\ &lt;br /&gt;
 &amp;amp; S{{S}_{E}}= &amp;amp; \underset{i=1}{\overset{50}{\mathop \sum }}\,{{({{y}_{i}}-{{{\hat{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of  &amp;lt;math&amp;gt;M{{S}_{LOF}}&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The error sum of squares,  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt; , can now be split into the sum of squares due to pure error,  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt; , and the sum of squares due to lack-of-fit,  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt; .  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  can be calculated as follows considering that in this example  &amp;lt;math&amp;gt;n=25&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;m=2&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{PE}}= &amp;amp; \underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{{{m}_{i}}}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \underset{i=1}{\overset{25}{\mathop \sum }}\,\underset{j=1}{\overset{2}{\mathop \sum }}\,{{({{y}_{ij}}-{{{\bar{y}}}_{i}})}^{2}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 350  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{PE}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{PE}})= &amp;amp; nm-n \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25\times 2-25 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 25  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The corresponding mean square,  &amp;lt;math&amp;gt;M{{S}_{PE}}&amp;lt;/math&amp;gt; , can now be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{PE}}= &amp;amp; \frac{S{{S}_{PE}}}{dof(S{{S}_{PE}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{350}{25} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 14  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  can be obtained by subtraction from  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  as:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; S{{S}_{LOF}}= &amp;amp; S{{S}_{E}}-S{{S}_{PE}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 648.37-350 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 298.37  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Similarly, the number of degrees of freedom associated with  &amp;lt;math&amp;gt;S{{S}_{LOF}}&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; dof(S{{S}_{LOF}})= &amp;amp; dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; (nm-2)-(nm-n) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 23  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The lack-of-fit mean square is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; M{{S}_{LOF}}= &amp;amp; \frac{M{{S}_{LOF}}}{dof(M{{S}_{LOF}})} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{298.37}{23} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 12.97  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Calculation of the Test Statistic&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic for the lack-of-fit test can now be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{f}_{0}}= &amp;amp; \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; \frac{12.97}{14} \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.93  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The critical value for this test is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{0.05,23,25}}=1.97&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{f}_{0}}&amp;lt;{{f}_{0.05,23,25}}&amp;lt;/math&amp;gt; , we fail to reject the hypothesis that the model adequately fits the data. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value for this case is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; p\text{ }value= &amp;amp; 1-P(F\le {{f}_{0}}) \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 1-0.43 \\ &lt;br /&gt;
 &amp;amp; = &amp;amp; 0.57  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, at a significance level of 0.05 we conclude that the simple linear regression model,  &amp;lt;math&amp;gt;y=14.47+2.04x&amp;lt;/math&amp;gt; , is adequate for the observed data. Table 4.3 presents a summary of the ANOVA calculations for the lack-of-fit test.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doet4.3.png|thumb|center|500px|ANOVA table for the lack-of-fit test of the yield data example.]]&lt;br /&gt;
&lt;br /&gt;
==Transformations==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The linear regression model may not be directly applicable to certain data. Non-linearity may be detected from scatter plots or may be known through the underlying theory of the product or process or from past experience. Transformations on either the predictor variable,  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; , or the response variable,  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; , may often be sufficient to make the linear regression model appropriate for the transformed data.&lt;br /&gt;
If it is known that the data follows the logarithmic distribution, then a logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) might be useful. For data following the Poisson distribution, a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is generally applicable.&lt;br /&gt;
&lt;br /&gt;
Transformations on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  may also be applied based on the type of scatter plot obtained from the data. Figure TransformationScatterPlots shows a few such examples. For the scatter plot of Figure (a), a square root transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=\sqrt{Y}&amp;lt;/math&amp;gt; ) is applicable. While for Figure (b), a logarithmic transformation (i.e.  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (Y)&amp;lt;/math&amp;gt; ) may be applied. For Figure (c), the reciprocal transformation ( &amp;lt;math&amp;gt;{{Y}^{*}}=1/Y&amp;lt;/math&amp;gt; ) is applicable. At times it may be helpful to introduce a constant into the transformation of  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; . For example, if  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  is negative and the logarithmic transformation on  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  seems applicable, a suitable constant,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , may be chosen to make all observed  &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;  positive. Thus the transformation in this case would be  &amp;lt;math&amp;gt;{{Y}^{*}}=\log (k+Y)&amp;lt;/math&amp;gt; . &lt;br /&gt;
The Box-Cox method may also be used to automatically identify a suitable power transformation for the data based on the relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Y}^{*}}={{Y}^{\lambda }}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here the parameter  &amp;lt;math&amp;gt;\lambda &amp;lt;/math&amp;gt;  is determined using the given data such that  &amp;lt;math&amp;gt;S{{S}_{E}}&amp;lt;/math&amp;gt;  is minimized (details on this method are presented in Chapter 6).&lt;br /&gt;
&lt;br /&gt;
[[Image:doe4.17.png|thumb|center|400px|Transformations on  for a few possible scatter plots. Plot (a) may require , (b) may require  and (c) may require.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32831</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32831"/>
		<updated>2012-08-21T04:30:31Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Inference on Variance of a Normal Population */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;. The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt;, of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;.&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.8.png|thumb|center|300px|&amp;lt;math&amp;gt;F &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot;&amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot;&amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply&amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;&amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{z}_{0}}&amp;amp;=&amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp;=&amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp;=&amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
       \chi _{0}^{2} &amp;amp;=&amp;amp;\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32829</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32829"/>
		<updated>2012-08-21T04:26:55Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Inference on the Mean of a Population When the Variance Is Known */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;. The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt;, of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;.&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.8.png|thumb|center|300px|&amp;lt;math&amp;gt;F &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot;&amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot;&amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply&amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;&amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{z}_{0}}&amp;amp;=&amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp;=&amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp;=&amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32828</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32828"/>
		<updated>2012-08-21T04:22:15Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Hypothesis Testing */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;. The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt;, of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;.&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.8.png|thumb|center|300px|&amp;lt;math&amp;gt;F &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot;&amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot;&amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply&amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;&amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{z}_{0}}= &amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32818</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32818"/>
		<updated>2012-08-21T01:29:48Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* F  Distribution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;. The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt;, of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;.&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.8.png|thumb|center|300px|&amp;lt;math&amp;gt;F &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{z}_{0}}= &amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32815</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32815"/>
		<updated>2012-08-21T01:27:22Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* F  Distribution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;. The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt;, of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;.&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{z}_{0}}= &amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32796</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32796"/>
		<updated>2012-08-21T00:12:53Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Central Limit Theorem */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;. The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt;, of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;.&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{z}_{0}}= &amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32767</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32767"/>
		<updated>2012-08-20T22:53:53Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt; , of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; .&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom (dof)==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{z}_{0}}= &amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32763</id>
		<title>Statistical Background on DOE</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Statistical_Background_on_DOE&amp;diff=32763"/>
		<updated>2012-08-20T22:49:57Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Degrees of Freedom ( dof ) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:Doebook|2}}&lt;br /&gt;
&lt;br /&gt;
==Introduction==&lt;br /&gt;
Variations occur in nature, be it the tensile strength of a particular grade of steel, caffeine content in your energy drink or the distance traveled by your vehicle in a day. Variations are also seen in the observations recorded during multiple executions of a process, even when all factors are strictly maintained at their respective levels and all the executions are run as identically as possible. The natural variations that occur in a process, even when all conditions are maintained at the same level, are often termed as noise. When the effect of a particular factor on a process is studied it becomes extremely important to distinguish the changes in the process caused by the factor from noise. A number of statistical methods are available to achieve this. This chapter covers basic statistical concepts that are useful in understanding the statistical analysis of data obtained from designed experiments. The initial sections of this chapter discuss the normal distribution and related concepts. The assumption of the normal distribution is widely used in the analysis of designed experiments. The subsequent sections introduce the standard normal, Chi-Squared,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distributions that are widely used in calculations related to hypothesis testing and confidence bounds. The final sections of this chapter cover hypothesis testing. It is important to gain a clear understanding of hypothesis testing because this concept finds direct application in the analysis of designed experiments to determine whether a particular factor is significant or not [[EDAR Appendix F|[Montgomery and Runger, 1991]]].&lt;br /&gt;
 &lt;br /&gt;
==Random Variables and the Normal Distribution==&lt;br /&gt;
If you record the distance traveled by your car everyday then these values would show some variation because it is unlikely that your car travels the same distance each day. If a variable  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is used to denote these values then  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is termed as a random variable (because of the diverse and unpredicted values  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  can have). Random variables are denoted by uppercase letters while a measured value of the random variable is denoted by the corresponding lowercase letter. For example, if the distance traveled by your car on January 1 was 10.7 miles then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;x=10.7\text{ miles} &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A commonly used distribution to describe the behavior of random variables is the normal distribution. When you calculate the mean and standard deviation for a given data set, you are assuming that the data follows a normal distribution. A normal distribution (also referred to as the Gaussian distribution) is a bell shaped curved (see Figure Ch3NormalDist). The mean and standard deviation are the two parameters of this distribution. The mean determines the location of the distribution on the  &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;  axis and is also called the location parameter of the normal distribution. The standard deviation determines the spread of the distribution (how narrow or wide) and is thus called the scale parameter of the normal distribution. The standard deviation, or its square called variance, gives an indication of the variability or spread of data. A large value of the standard deviation (or variance) implies that a large amount of variability exists in the data.&lt;br /&gt;
 &lt;br /&gt;
Any curve in Figure Ch3NormalDist is also referred to as the probability density function or pdf of the normal distribution as the area under the curve gives the probability of occurrence of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  for a particular interval. For instance, if you obtained the mean and standard deviation for the distance data of your car as 15 miles and 2.5 miles respectively, then the probability that your car travels a distance between 7 miles and 14 miles is given by the area under the curve covered between these two values which is calculated as 34.4% (see Figure Ch3MilesDistEx). This means that on 34.4 days out of every 100 days your car travels, you car can be expected to cover a distance in the range of 7 to 14 miles.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.1.png|thumb|center|300px|Normal probability density functions for different values of mean and standard deviation.]]&lt;br /&gt;
&lt;br /&gt;
On a normal probability density function, the area under the curve between the values of  &amp;lt;math&amp;gt;Mean-(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;Mean+(3\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  is approximately 99.7% of the total area under the curve. This implies that almost all the time (or 99.7% of the time) the distance traveled will fall in the range of 7.5 miles  &amp;lt;math&amp;gt;(15-3\times 2.5)&amp;lt;/math&amp;gt;  and 22.5 miles  &amp;lt;math&amp;gt;(15+3\times 2.5)&amp;lt;/math&amp;gt; . Similarly,  &amp;lt;math&amp;gt;Mean\pm (2\times Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 95% of the area under the curve and  &amp;lt;math&amp;gt;Mean\pm (Standard&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;Deviation)&amp;lt;/math&amp;gt;  covers approximately 68% of the area under the curve.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.2.png|thumb|center|300px|Normal probability density function with the shaded area representing the probability of occurrence of data between 7 and 14 miles.]]&lt;br /&gt;
&lt;br /&gt;
==Population Mean, Sample Mean and Variance==&lt;br /&gt;
&lt;br /&gt;
If data for all of the population under investigation is known, then the mean and variance for this population can be calculated as follows:&lt;br /&gt;
&lt;br /&gt;
Population Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\mu =\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{x}_{i}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Population Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\sigma }^{2}}=\frac{\underset{i=1}{\overset{N}{\mathop{\sum }}}\,{{({{x}_{i}}-\mu )}^{2}}}{N}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;  is the size of the population.&lt;br /&gt;
&lt;br /&gt;
The population standard deviation is the positive square root of the population variance.&lt;br /&gt;
&lt;br /&gt;
Most of the time it is not possible to obtain data for the entire population. For example, it is impossible to measure the height of every male in a country to determine the average height and variance for males of a particular country. In such cases, results for the population have to be estimated using samples. This process is known as statistical inference. Mean and variance for a sample are calculated using the following relations:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Sample Mean:&lt;br /&gt;
::&amp;lt;math&amp;gt;\bar{x}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{x}_{i}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
Sample Variance:&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
The sample standard deviation is the positive square root of the sample variance.&lt;br /&gt;
The sample mean and variance of a random sample can be used as estimators of the population mean and variance respectively. The sample mean and variance may be referred to as statistics. A statistic is any function of observations in a random sample.&lt;br /&gt;
You may have noticed that the denominator in the calculation of sample variance, unlike the denominator in the calculation of population variance, is  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  and not  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; . The reason for this difference is explained in Section 3.BiasedEstimators.&lt;br /&gt;
&lt;br /&gt;
==Central Limit Theorem==&lt;br /&gt;
&lt;br /&gt;
The Central Limit Theorem states that for large sample size  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The sample means from a population are normally distributed with a mean value equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , even if the population is not normally distributed.&lt;br /&gt;
What this means is that if random samples are drawn from any population and the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , calculated for each of these samples, then these sample means would follow the normal distribution with a mean (or location parameter) equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the distribution of the statistic,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , would be a normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . The distribution of a statistic is called the sampling distribution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:•	The variance,  &amp;lt;math&amp;gt;{{s}^{2}}\,\!&amp;lt;/math&amp;gt; , of the sample means would be  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  times smaller than the variance of the population,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
This implies that the sampling distribution of the sample means would have a variance equal to  &amp;lt;math&amp;gt;{{\sigma }^{2}}/n\,\!&amp;lt;/math&amp;gt;  (or a scale parameter equal to  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; ), where  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation. The standard deviation of the sampling distribution of an estimator is called the standard error of the estimator. Thus the standard error of sample mean  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  is  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; .&lt;br /&gt;
In short, the Central Limit Theorem states that the sampling distribution of the sample mean is a normal distribution with parameters  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt;  as shown in the figure below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.3.png|thumb|center|300px|Sampling distribution of the sample emna. The distribution is normal with the mean equal to the population mean and the variance equal to the &#039;&#039;n&#039;&#039;th fraction of the population variance.]]&lt;br /&gt;
&lt;br /&gt;
==Unbiased and Biased Estimators==&lt;br /&gt;
&lt;br /&gt;
If the mean value of an estimator equals the true value of the quantity it estimates, then the estimator is called an unbiased estimator (see Figure Ch3BiasedEstimator). For example, assume that the sample mean is being used to estimate the mean of a population. Using the Central Limit Theorem, the mean value of the sample means equals the population mean. Therefore, the sample mean is an unbiased estimator of the population mean.&lt;br /&gt;
If the mean value of an estimator is either less than or greater than the true value of the quantity it estimates, then the estimator is called a biased. For example, suppose you decide to choose the smallest observation in a sample to be the estimator of the population mean. Such an estimator would be biased because the average of the values of this estimator would always be less than the true population mean. In other words, the mean of the sampling distribution of this estimator would be less than the true value of the population mean it is trying to estimate. Consequently, the estimator is a biased estimator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.4.png|thumb|center|300px|Exmaple showing the distribution of a biased estimator which underestimated the parameter in question, along with the distribution of an unbiased estimator.]]&lt;br /&gt;
&lt;br /&gt;
A case of biased estimation is seen to occur when sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , is used to estimate the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , if the following relation is used to calculate the sample variance:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The sample variance calculated using this relation is always less than the true population variance. This is because to calculate the sample variance, deviations with respect to the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , are used.  Sample observations,  &amp;lt;math&amp;gt;{{x}_{i}}&amp;lt;/math&amp;gt; , tend to be closer to  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt;  than to  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; . Thus, the calculated deviations  &amp;lt;math&amp;gt;({{x}_{i}}-\bar{x})&amp;lt;/math&amp;gt;  are smaller. As a result, the sample variance obtained is smaller than the population variance. To compensate for this,  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  is used as the denominator in place of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  in the calculation of sample variance. Thus, the correct formula to obtain the sample variance is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{s}^{2}}=\frac{\underset{i=1}{\overset{n}{\mathop{\sum }}}\,{{({{x}_{i}}-\bar{x})}^{2}}}{n-1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is important to note that although using  &amp;lt;math&amp;gt;(n-1)&amp;lt;/math&amp;gt;  as the denominator makes the sample variance,  &amp;lt;math&amp;gt;{{s}^{2}}&amp;lt;/math&amp;gt; , an unbiased estimator of the population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , still remains a biased estimator of the population standard deviation,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; . For large sample sizes this bias is negligible.&lt;br /&gt;
&lt;br /&gt;
==Degrees of Freedom ( &amp;lt;math&amp;gt;dof&amp;lt;/math&amp;gt; )==&lt;br /&gt;
&lt;br /&gt;
Degrees of freedom refer to the number of independent observations made in excess of the unknowns. If there are 3 unknowns and 7 independent observations are taken then the number of degrees of freedom is  &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;7-3=4&amp;lt;/math&amp;gt; ). As another example, two parameters are needed to specify a line, therefore, there are 2 unknowns. If 10 points are available to fit the line, the number of degrees of freedom is  &amp;lt;math&amp;gt;8&amp;lt;/math&amp;gt;  ( &amp;lt;math&amp;gt;10-2=8&amp;lt;/math&amp;gt; ).&lt;br /&gt;
&lt;br /&gt;
==Standard Normal Distribution==&lt;br /&gt;
&lt;br /&gt;
A normal distribution with mean  &amp;lt;math&amp;gt;\mu =0&amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is called the standard normal distribution (see Figure Ch3StdNormDist). Standard normal random variables are denoted by  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  represents a normal random variable that follows the normal distribution with mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  and variance  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , then the corresponding standard normal random variable is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z=(X-\mu )/\sigma &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  represents the distance of  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  from the mean  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt;  in terms of the standard deviation  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.5.png|thumb|center|300px|Standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
==Chi-Squared Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, then the distribution of  &amp;lt;math&amp;gt;{{Z}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared distribution (see Figure Ch3ChiSqDist). A Chi-Squared random variable is represented by  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt; . Thus:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\chi }^{2}}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.6.png|thumb|center|300px|Chi-Squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The distribution of the variable  &amp;lt;math&amp;gt;{{\chi }^{2}}&amp;lt;/math&amp;gt;  mentioned in the previous equation is also referred to as centrally distributed Chi-Squared with one degree of freedom. The degree of freedom is one here because here the Chi-Squared random variable is obtained from a single standard normal random variable  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; . The previous equation may also be represented by including the degree of freedom into the equation as: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{Z}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{Z}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{Z}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{Z}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=Z_{1}^{2}+Z_{2}^{2}+Z_{3}^{2}...+Z_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
is also a Chi-Squared random variable. The distribution of  &amp;lt;math&amp;gt;\chi _{m}^{2}&amp;lt;/math&amp;gt;  is said to be centrally Chi-Squared with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom, as the Chi-Squared random variable is obtained from  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent standard normal random variables.&lt;br /&gt;
If  &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;  is a normal random variable then the distribution of  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is said to be non-centrally distributed Chi-Squared with one degree of freedom. Therefore,  &amp;lt;math&amp;gt;{{X}^{2}}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable and can be represented as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{1}^{2}={{X}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{X}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{X}_{3}}&amp;lt;/math&amp;gt; ... &amp;lt;math&amp;gt;{{X}_{m}}&amp;lt;/math&amp;gt;  are  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  independent normal random variables then: &lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{m}^{2}=X_{1}^{2}+X_{2}^{2}+X_{3}^{2}...+X_{m}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is a non-centrally distributed Chi-Squared random variable with  &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
==Student&#039;s  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution ( &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  Distribution)==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt;  is a standard normal random variable, and  &amp;lt;math&amp;gt;\chi _{k}^{2}&amp;lt;/math&amp;gt;  is a Chi-Squared random variable with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom, and both of these random variables are independent, then the distribution of the random variable  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T=\frac{Z}{\sqrt{\chi _{k}^{2}/k}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;  degrees of freedom.  &lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution is similar in appearance to the standard normal distribution (see Figure Ch3tDist). Both of these distributions are symmetric, reaching a maximum at the mean value of zero. However, the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution has heavier tails than the standard normal distribution implying that it has more probability in the tails. As the degrees of freedom,  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; , of the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution approach infinity, the distribution approaches the standard normal distribution.&lt;br /&gt;
 &lt;br /&gt;
[[Image:doe3.7.png|thumb|center|300px|&amp;lt;math&amp;gt;t &amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  Distribution==&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;\chi _{u}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\chi _{v}^{2}&amp;lt;/math&amp;gt;  are two independent Chi-Squared random variables with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom, respectively, then the distribution of the random variable  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;F=\frac{\chi _{u}^{2}/u}{\chi _{v}^{2}/v}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
is said to follow the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;u&amp;lt;/math&amp;gt;  degrees of freedom in the numerator and  &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;  degrees of freedom in the denominator. The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution resembles the Chi-Squared distribution (see Figure Ch3FDist). This is because the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable, like the Chi-Squared random variable, is non-negative and the distribution is skewed to the right (a right skew means that the distribution is unsymmetrical and has a right tail). The  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  random variable is usually abbreviated by including the degrees of freedom as  &amp;lt;math&amp;gt;{{F}_{u,v}}&amp;lt;/math&amp;gt; .&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
==Hypothesis Testing==&lt;br /&gt;
A statistical hypothesis is a statement about the population under study or about the distribution of a quantity under consideration. The null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , is the hypothesis to be tested. It is a statement about a theory that is believed to be true but has not been proven. For instance, if a new product design is thought to perform consistently, regardless of the region of operation, then the null hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : New product design performance is not affected by region.&amp;quot; Statements in  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  always include exact values of parameters under consideration, e.g. &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; : The population mean is 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , leads to the possibility that the alternative hypothesis,  &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; , may be true. Given the previous null hypothesis, the alternate hypothesis may be &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : New product design performance is affected by region.&amp;quot; In the case of the example regarding inference on the population mean, the alternative hypothesis may be stated as &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}&amp;lt;/math&amp;gt; : The population mean is not 100&amp;quot; or simply &amp;quot; &amp;lt;math&amp;gt;{{H}_{1}}\ \ :\ \ \mu \ne 100&amp;lt;/math&amp;gt; .&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing involves the calculation of a test statistic based on a random sample drawn from the population. The test statistic is then compared to the critical value(s) and used to make a decision about the null hypothesis. The critical values are set by the analyst.&lt;br /&gt;
The outcome of a hypothesis test is that we either &amp;quot;reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; or we &amp;quot;fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; .&amp;quot; Failing to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  implies that we did not find sufficient evidence to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . It does not necessarily mean that there is a high probability that  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is true. As such, the terminology &amp;quot;accept  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; &amp;quot; is not preferred.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a certain population is 100 or not. The statements for this hypothesis can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The analyst decides to use the sample mean as the test statistic for this test. The analyst further decides that if the sample mean lies between 98 and 102 it can be concluded that the population mean is 100. Thus, the critical values set for this test by the analyst are 98 and 102. It is also decided to draw out a random sample of size 25 from the population.&lt;br /&gt;
&lt;br /&gt;
Now assume that the true population mean is 100 (i.e.  &amp;lt;math&amp;gt;\mu =100&amp;lt;/math&amp;gt; ) and the true population standard deviation is 5 (i.e.  &amp;lt;math&amp;gt;\sigma =5&amp;lt;/math&amp;gt; ). This information is not known to the analyst. Using the Central Limit Theorem, the test statistic (sample mean) will follow a normal distribution with a mean equal to the population mean,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , and a standard deviation of  &amp;lt;math&amp;gt;\sigma /\sqrt{n}&amp;lt;/math&amp;gt; , where  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size. Therefore, the distribution of the test statistic has a mean of 100 and a standard deviation of  &amp;lt;math&amp;gt;5/\sqrt{25}=1&amp;lt;/math&amp;gt; . This distribution is shown in Figure Ch3HypoTestEx1.&lt;br /&gt;
 &lt;br /&gt;
The unshaded area in the figure bound by the critical values of 98 and 102 is called the acceptance region. The acceptance region gives the probability that a random sample drawn from the population would have a sample mean that lies between 98 and 102. Therefore, this is the region that will lead to the &amp;quot;acceptance&amp;quot; of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; . On the other hand, the shaded area gives the probability that the sample mean obtained from the random sample lies outside of the critical values. In other words, it gives the probability of rejection of the null hypothesis when the true mean is 100. The shaded area is referred to as the critical region or the rejection region. Rejection of the null hypothesis  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  when it is true is referred to as type I error. Thus, there is a 4.56% chance of making a type I error in this hypothesis test. This percentage is called the significance level of the test and is denoted by  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt; . Here  &amp;lt;math&amp;gt;\alpha =0.0456&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;4.56%&amp;lt;/math&amp;gt;  (area of the shaded region in the figure). The value of  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  is set by the analyst when he/she chooses the critical values.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.9.png|thumb|center|400px|Acceptance region and critical regions for the hypothesis test in Example 1.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
A type II error is also defined in hypothesis testing. This error occurs when the analyst fails to reject the null hypothesis when it is actually false. Such an error would occur if the value of the sample mean obtained is in the acceptance region bounded by 98 and 102 even though the true population mean is not 100. The probability of occurrence of type II error is denoted by  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; .&lt;br /&gt;
&lt;br /&gt;
===Two-sided and One-sided Hypotheses===&lt;br /&gt;
&lt;br /&gt;
As seen in the previous section, the critical region for the hypothesis test is split into two parts, with equal areas in each tail of the distribution of the test statistic. Such a hypothesis, in which the values for which we can reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  are in both tails of the probability distribution, is called a two-sided hypothesis.&lt;br /&gt;
The hypothesis for which the critical region lies only in one tail of the probability distribution is called a one-sided hypothesis. For instance, consider the following hypothesis test:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;gt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is an example of a one-sided hypothesis. Here the critical region lies entirely in the right tail of the distribution as shown in Figure Ch3OneSidedHypo.&lt;br /&gt;
The hypothesis test may also be set up as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;100  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
This is also a one-sided hypothesis. Here the critical region lies entirely in the left tail of the distribution as shown in Figure Ch3OneSidedHypoL.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for a Single Sample==&lt;br /&gt;
&lt;br /&gt;
Hypothesis testing forms an important part of statistical inference. As stated previously, statistical inference refers to the process of estimating results for the population based on measurements from a sample. In the next sections, statistical inference for a single sample is discussed briefly.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Mean of a Population When the Variance Is Known===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the standard normal distribution. If  &amp;lt;math&amp;gt;\bar{X}&amp;lt;/math&amp;gt;  is the calculated sample mean, then the standard normal test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{\sigma /\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;{{\mu }_{0}}&amp;lt;/math&amp;gt;  is the hypothesized population mean,  &amp;lt;math&amp;gt;\sigma &amp;lt;/math&amp;gt;  is the population standard deviation and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  is the sample size.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.10.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the right tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.11.png|thumb|center|300px|One-sided hypothesis where the critical region lies in the left tail.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is 100. The population variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , is known to be 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \mu =100 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \mu \ne 100  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It is a clear that this is a two-sided hypothesis. Thus the critical region will lie in both of the tails of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Assume that the analyst chooses a significance level of 0.05. Thus  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . The significance level determines the critical values of the test statistic. Here the test statistic is based on the standard normal distribution. For the two-sided hypothesis these values are obtained as: &lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{z}_{\alpha /2}}={{z}_{0.025}}=1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}=-{{z}_{0.025}}=-1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx2. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic,  &amp;lt;math&amp;gt;{{Z}_{0}}&amp;lt;/math&amp;gt; , is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{z}_{\alpha /2}}\le {{Z}_{0}}\le {{z}_{\alpha /2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
 &lt;br /&gt;
3) Next the analyst draws a random sample from the population. Assume that the sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25 and the sample mean is obtained as  &amp;lt;math&amp;gt;\bar{x}=103&amp;lt;/math&amp;gt; .&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.12.png|thumb|center|300px|Critical values and rejection region for Example 2 marked on the standard normal distribution.]]&lt;br /&gt;
&lt;br /&gt;
	&lt;br /&gt;
4) The value of the test statistic corresponding to the sample mean value of 103 is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{z}_{0}}= &amp;amp; \frac{\bar{x}-{{\mu }_{0}}}{\sigma /\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{103-100}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since this value does not lie in the acceptance region  &amp;lt;math&amp;gt;-1.96\le {{Z}_{0}}\le 1.96&amp;lt;/math&amp;gt; , we reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =100&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
===&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  Value===&lt;br /&gt;
&lt;br /&gt;
In the previous example the null hypothesis was rejected at a significance level of 0.05. This statement does not provide information as to how far out the test statistic was into the critical region. At times it is necessary to know if the test statistic was just into the critical region or was far out into the region. This information can be provided by using the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value.&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability of occurrence of the values of the test statistic that are either equal to the one obtained from the sample or more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  than the one obtained from the sample. It is the lowest significance level that would lead to the rejection of the null hypothesis,  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; , at the given value of the test statistic. The value of the test statistic is referred to as significant when  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected. The  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  at which the statistic is significant and  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is rejected.&lt;br /&gt;
&lt;br /&gt;
For instance, in the previous example the test statistic was obtained as  &amp;lt;math&amp;gt;{{z}_{0}}=3&amp;lt;/math&amp;gt; . Values that are more unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  in this case are values greater than 3. Then the required probability is the probability of getting a test statistic value either equal to or greater than 3 (this is abbreviated as  &amp;lt;math&amp;gt;P(Z\ge 3)&amp;lt;/math&amp;gt; ). This probability is shown in Figure Ch3Pvalue as the dark shaded area on the right tail of the distribution and is equal to 0.0013 or 0.13% (i.e.  &amp;lt;math&amp;gt;P(Z\ge 3)=0.0013&amp;lt;/math&amp;gt; ). Since this is a two-sided test the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;p\text{ }value=2\times 0.0013=0.0026&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Therefore, the smallest  &amp;lt;math&amp;gt;\alpha &amp;lt;/math&amp;gt;  (corresponding to the test static value of 3) that would lead to the rejection of  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  is 0.0026.   &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.13.png|thumb|center|400px|&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; value for Example 2.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on Mean of a Population When Variance Is Unknown===&lt;br /&gt;
When the variance,  &amp;lt;math&amp;gt;{{\sigma }^{2}}&amp;lt;/math&amp;gt; , of a population (that can be assumed to be normally distributed) is unknown the sample variance,  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt; , is used in its place in the calculation of the test statistic. The test statistic used in this case is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution and is obtained using the following relation:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the mean of a population,  &amp;lt;math&amp;gt;\mu &amp;lt;/math&amp;gt; , is less than 50 at a significance level of 0.05. A random sample drawn from the population gives the sample mean,  &amp;lt;math&amp;gt;\bar{x}&amp;lt;/math&amp;gt; , as 47.7 and the sample standard deviation,  &amp;lt;math&amp;gt;s&amp;lt;/math&amp;gt; , as 5. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 25. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
:1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; \mu =50 \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; \mu &amp;lt;50  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a one-sided hypothesis. Here the critical region will lie in the left tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
:2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution. Thus, for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;-{{t}_{\alpha ,dof}}=-{{t}_{0.05,n-1}}=-{{t}_{0.05,24}}=-1.7109&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx3tDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}&amp;gt;-{{t}_{0.05,24}}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
		&lt;br /&gt;
:3) The value of the test statistic,  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt; , corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{t}_{0}}= &amp;amp; \frac{\bar{X}-{{\mu }_{0}}}{S/\sqrt{n}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{47.7-50}{5/\sqrt{25}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; -2.3  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  is less than the critical value of -1.7109,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \mu =50&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population mean is less than 50.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is either less than or equal to  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  (since values less than  &amp;lt;math&amp;gt;-2.3&amp;lt;/math&amp;gt;  are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is equal to 0.0152. &lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.14.png|thumb|center|300px|Critical value and rejection region for Example 3 marked on the &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Inference on Variance of a Normal Population===&lt;br /&gt;
&lt;br /&gt;
The test statistic used in this case is based on the Chi-Squared distribution. If  &amp;lt;math&amp;gt;{{S}^{2}}&amp;lt;/math&amp;gt;  is the calculated sample variance and  &amp;lt;math&amp;gt;\sigma _{0}^{2}&amp;lt;/math&amp;gt;  the hypothesized population variance then the Chi-Squared test statistic is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}=\frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the Chi-Squared distribution with  &amp;lt;math&amp;gt;n-1&amp;lt;/math&amp;gt;  degrees of freedom.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 4&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variance of a population exceeds 1 at a significance level of 0.05. A random sample drawn from the population gives the sample variance as 2. The sample size,  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , is 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; {{\sigma }^{2}}=1 \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; {{\sigma }^{2}}&amp;gt;1  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This is a one-sided hypothesis. Here the critical region will lie in the right tail of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level,  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here, the test statistic is based on the Chi-Squared distribution. Thus for the one-sided hypothesis the critical value is obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{\alpha ,n-1}^{2}=\chi _{0.05,19}^{2}=30.1435&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
This value and the critical regions are shown in Figure Ch3HypoTestEx4ChiDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;\chi _{0.05,19}^{2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  corresponding to the given sample data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; \chi _{0}^{2}= &amp;amp; \frac{(n-1){{S}^{2}}}{\sigma _{0}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{(20-1)2}{1}=38  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;\chi _{0}^{2}&amp;lt;/math&amp;gt;  is greater than the critical value of 30.1435,  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ {{\sigma }^{2}}=1&amp;lt;/math&amp;gt;  is rejected and it is concluded that at a significance level of 0.05 the population variance exceeds 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.15.png|thumb|center|300px|Critical value and rejection region for Example 4 marked on the chi-squared distribution.]]&lt;br /&gt;
&lt;br /&gt;
4) &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;  value&lt;br /&gt;
&lt;br /&gt;
In this case the  &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;  value is the probability that the test statistic is greater than or equal to 38 (since values greater than 38 are unfavorable to  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt; ). This probability is determined to be 0.0059.&lt;br /&gt;
&lt;br /&gt;
==Statistical Inference for Two Samples==&lt;br /&gt;
&lt;br /&gt;
This section briefly covers statistical inference for two samples.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Known===&lt;br /&gt;
The test statistic used here is based on the standard normal distribution. Let  &amp;lt;math&amp;gt;{{\mu }_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\mu }_{2}}&amp;lt;/math&amp;gt;  represent the means of two populations, and  &amp;lt;math&amp;gt;\sigma _{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\sigma _{2}^{2}&amp;lt;/math&amp;gt;  their variances, respectively. Let  &amp;lt;math&amp;gt;{{\Delta }_{0}}&amp;lt;/math&amp;gt;  be the hypothesized difference in the population means and  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt;  be the sample means obtained from two samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  drawn randomly from the two populations, respectively. The test statistic can be obtained as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{Z}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{\sigma _{1}^{2}}{{{n}_{1}}}+\frac{\sigma _{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The statements for the hypothesis test are:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
  &amp;amp; {{H}_{0}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}={{\Delta }_{0}} \\ &lt;br /&gt;
 &amp;amp; {{H}_{1}}: &amp;amp; {{\mu }_{1}}-{{\mu }_{2}}\ne {{\Delta }_{0}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If  &amp;lt;math&amp;gt;{{\Delta }_{0}}=0&amp;lt;/math&amp;gt; , then the hypothesis will test for the equality of the two population means.&lt;br /&gt;
&lt;br /&gt;
===Inference on the Difference in Population Means When Variances Are Unknown===&lt;br /&gt;
&lt;br /&gt;
If the population variances can be assumed to be equal then the following test statistic based on the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution can be used. Let  &amp;lt;math&amp;gt;{{\bar{X}}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{\bar{X}}_{2}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  be the sample means and variances obtained from randomly drawn samples of sizes  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  from the two populations, respectively. The weighted average,  &amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt; , of the two sample variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}=\frac{({{n}_{1}}-1)S_{1}^{2}+({{n}_{2}}-1)S_{2}^{2}}{{{n}_{1}}+{{n}_{2}}-2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;S_{p}^{2}&amp;lt;/math&amp;gt;  has ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. The test statistic can be calculated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{T}_{0}}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{{{S}_{p}}\sqrt{\frac{1}{{{n}_{1}}}+\frac{1}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;{{T}_{0}}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  +  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 2) degrees of freedom. This test is also referred to as the two-sample pooled  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  test.&lt;br /&gt;
If the population variances cannot be assumed to be equal then the following test statistic is used:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;T_{0}^{*}=\frac{{{{\bar{X}}}_{1}}-{{{\bar{X}}}_{2}}-{{\Delta }_{0}}}{\sqrt{\frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;T_{0}^{*}&amp;lt;/math&amp;gt;  follows the  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  distribution with  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  degrees of freedom.  &amp;lt;math&amp;gt;\upsilon &amp;lt;/math&amp;gt;  is defined as follows:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\upsilon =\frac{{{\left( \frac{S_{1}^{2}}{{{n}_{1}}}+\frac{S_{2}^{2}}{{{n}_{2}}} \right)}^{2}}}{\frac{{{\left( S_{1}^{2}/{{n}_{1}} \right)}^{2}}}{{{n}_{1}}+1}+\frac{{{\left( S_{2}^{2}/{{n}_{2}} \right)}^{2}}}{{{n}_{2}}+1}}-2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Inference on the Variances of Two Normal Populations===&lt;br /&gt;
&lt;br /&gt;
The test statistic used here is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. If  &amp;lt;math&amp;gt;S_{1}^{2}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;S_{2}^{2}&amp;lt;/math&amp;gt;  are the sample variances drawn randomly from the two populations and  &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  are the two sample sizes, respectively, then the test statistic that can be used to test the equality of the population variances is:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{F}_{0}}=\frac{S_{1}^{2}}{S_{2}^{2}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The test statistic follows the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution with ( &amp;lt;math&amp;gt;{{n}_{1}}&amp;lt;/math&amp;gt;  -- &lt;br /&gt;
1) degrees of freedom in the numerator and ( &amp;lt;math&amp;gt;{{n}_{2}}&amp;lt;/math&amp;gt;  -- 1) degrees of freedom in the denominator.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assume that an analyst wants to know if the variances of two normal populations are equal at a significance level of 0.05. Random samples drawn from the two populations give the sample standard deviations as 1.84 and 2, respectively. Both the sample sizes are 20. The hypothesis test may be conducted as follows:&lt;br /&gt;
&lt;br /&gt;
1) The statements for this hypothesis test may be formulated as:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{H}_{0}}: &amp;amp; \sigma _{1}^{2}=\sigma _{2}^{2} \\ &lt;br /&gt;
	 &amp;amp; {{H}_{1}}: &amp;amp; \sigma _{1}^{2}\ne \sigma _{2}^{2}  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
It is clear that this is a two-sided hypothesis and the critical region will be located on both sides of the probability distribution.&lt;br /&gt;
&lt;br /&gt;
2) Significance level  &amp;lt;math&amp;gt;\alpha =0.05&amp;lt;/math&amp;gt; . Here the test statistic is based on the  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  distribution. For the two-sided hypothesis the critical values are obtained as:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.025,19,19}}=2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
and&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}={{f}_{0.975,19,19}}=0.40&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
These values and the critical regions are shown in Figure Ch3HypoTestEx5FDist. The analyst would fail to reject  &amp;lt;math&amp;gt;{{H}_{0}}&amp;lt;/math&amp;gt;  if the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  is such that:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1-\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}\le {{F}_{0}}\le {{f}_{\alpha /2,{{n}_{1}}-1,{{n}_{2}}-1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;0.40\le {{F}_{0}}\le 2.53&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
3) The value of the test statistic  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  corresponding to the given data is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  &amp;amp; {{f}_{0}}= &amp;amp; \frac{S_{1}^{2}}{S_{2}^{2}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; \frac{{{1.84}^{2}}}{{{2}^{2}}} \\ &lt;br /&gt;
	 &amp;amp; = &amp;amp; 0.8464  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
	&lt;br /&gt;
Since  &amp;lt;math&amp;gt;{{F}_{0}}&amp;lt;/math&amp;gt;  lies in the acceptance region, the analyst fails to reject  &amp;lt;math&amp;gt;{{H}_{0}}\ \ :\ \ \sigma _{1}^{2}=\sigma _{2}^{2}&amp;lt;/math&amp;gt;  at a significance level of 0.05.&lt;br /&gt;
&lt;br /&gt;
[[Image:doe3.16.png|thumb|center|300px|Critical values and rejection region for Example 5 marked on the &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; distribution.]]&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15668</id>
		<title>Reliability Importance and Optimized Reliability Allocation (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15668"/>
		<updated>2012-02-13T23:56:38Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Reliability Allocation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|6}}&lt;br /&gt;
&lt;br /&gt;
=Component Reliability Importance=&lt;br /&gt;
===Static Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Once the reliability of a system has been determined, engineers are often faced with the task of identifying the least reliable component(s) in the system in order to improve the design.  For example, it was observed in Chapter 4 that the least reliable component in a series system has the biggest effect on the system reliability.  In this case, if the reliability of the system is to be improved, then the efforts can best be concentrated on improving the reliability of that component first.   In simple systems such as a series system, it is easy to identify the weak components.  However, in more complex systems this becomes quite a difficult task.  For complex systems, the analyst needs a mathematical approach that will provide the means of identifying and quantifying the importance of each component in the system.&lt;br /&gt;
&lt;br /&gt;
Using reliability importance measures is one method of identifying the relative importance of each component in a system with respect to the overall reliability of the system.  The reliability importance,  &amp;lt;math&amp;gt;{{I}_{R}}&amp;lt;/math&amp;gt; , of component  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;  in a system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  components is given by [[Appendix D: Weibull References | Leemis [17]]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{i}}}   \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{s}}&amp;lt;/math&amp;gt;  is the system reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  is the component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The value of the reliability importance given by Eqn.6.1 depends both on the reliability of a component and its corresponding position in the system.  In Chapter 4 we observed that for a simple series system (three components in series with reliabilities of 0.7, 0.8 and 0.9) the rate of increase of the system reliability was greatest when the least reliable component was improved.  In other words, it was observed that Component 1 had the largest reliability importance in the system relative to the other two components (see Figure 6.1).  The same conclusion can be drawn by using Eqn.6.1 and obtaining the reliability importance in terms of a value for each component.&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Using BlockSim, the reliability importance values for these components can be calculated with Eqn.6.1.  Using the plot option and selecting the Static Reliability Importance plot type, Figure 6.2 can be obtained.  Note that the time input required to create this plot is irrelevant for this example because the components are static.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The values shown in Figure fig1a for each component were obtained using Eqn.6.1.  The reliability equation for this series system was given by: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}   \ (eqn 2)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Taking the partial derivative of Eqn.6.2 with respect to  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{I}_{{{R}_{1}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{1}}}= &amp;amp; {{R}_{2}}{{R}_{3}} \\ &lt;br /&gt;
= &amp;amp; 0.8\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.72  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus the reliability importance of Component 1 is  &amp;lt;math&amp;gt;{{I}_{{{R}_{1}}}}=&amp;lt;/math&amp;gt;  0.72.  The reliability importance values for Components 2 and 3 are obtained in a similar manner.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.1.png|thumb|center|300px|Rate of change of system reliability when increasing the reliability of each component.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:6.2.gif|thumb|center|300px|Static reliability importance plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Time-Dependent Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
\The same concept applies if the components have a time-varying reliability.  That is, if  &amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt; , then one could compute  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  at any time  &amp;lt;math&amp;gt;x,&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}{{(t)}_{_{t=x}}}.&amp;lt;/math&amp;gt;   This is quantified in Eqn. (importance time). &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)=\frac{\partial {{R}_{s}}(t)}{\partial {{R}_{i}}(t)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In turn, this can be viewed as either a static plot (at a given time) or as time-varying plot, as illustrated in the next figures.  Specifically, Figures Ch6fig3, Ch6fig4 and Ch6fig5 present the analysis for three components configured reliability-wise in series following a Weibull distribution with  &amp;lt;math&amp;gt;\beta =3&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=1,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\eta }_{2}}=2,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=3,000&amp;lt;/math&amp;gt; .  Figure Ch6fig3 shows a bar chart of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  while Figure Ch6fig4 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  in BlockSim&#039;s tableau chart format.  In this chart, the area of the square is  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt; .  Lastly, Figure Ch6fig5 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  vs. time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that a system has failure modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; .  Furthermore, assume that failure of the entire system will occur if:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  occurs.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  occur.&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, assume the following failure probabilities for each mode.&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  have a mean time to occurrence of 1,000 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=1,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  has a mean time to occurrence of 100 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=100).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a mean time to occurrence of 700,000, 1,000,000 and 2,000,000 hours respectively (i.e. exponential with  &amp;lt;math&amp;gt;MTT{{F}_{B}}=700,000&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;MTT{{F}_{C}}=1,000,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;MTT{{F}_{F}}=2,000,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examine the mode importance for operating times of 100 and 500 hours.&lt;br /&gt;
&lt;br /&gt;
[[Image:6.3.gif|thumb|center|400px|Static Reliability Importance plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.4.png|thumb|center|400px|Static Reliability Importance tableau plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.5.png|thumb|center|400px|Reliability Importance vs. time plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
The RBD for this example is (from Chapter 4, Example 18):&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6ex1.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig6 illustrates  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt; .  It can be seen that even though  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a much rarer rate of occurrence, they are much more significant at 100 hours.  By 500 hours,  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt; , the effects of the lower reliability components become greatly pronounced and thus they become more important, as can be seen in Figure Ch6fig7.  Finally, the behavior of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  can be observed in Figure Ch6fig8.  Note that not all lines are plainly visible in Figure Ch6fig8 due to overlap.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Reliability Allocation=&lt;br /&gt;
&lt;br /&gt;
In the process of  developing a new product, the engineer is often faced with the task of designing a system that conforms to a set of reliability specifications.  The engineer is given the goal for the system and must then develop a design that will achieve the desired reliability of the system, while performing all of the system&#039;s intended functions at a minimum cost. This involves a balancing act of determining how to allocate reliability to the components in the system so the system will meet its reliability goal while at the same time ensuring that the system meets all of the other associated performance specifications.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.6.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.7.gif|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.8.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
Blocksim provide 3 allocation methods: equal allocation, weighted reliability allocation and cost optimzation allocation. In these 3 methods, the simplest method is equal reliability allocatio, which distributes the reliabilities uniformly among all components. For example, suppose a system with five components in series has a reliability objective of 90% for a given operating time. The uniform allocation of the objective to all components would require each component to have a reliability of 98% for the specified operating time, since  &amp;lt;math&amp;gt;{{0.98}^{5}}\tilde{=}0.90&amp;lt;/math&amp;gt;. While this manner of allocation is easy to calculate, it is generally not the best way to allocate reliability for a system. The optimum method of allocating reliability would take into account the cost or relative difficulty of improving the reliability of different subsystems or components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability optimization process begins with the development of a model that represents the entire system.  This is accomplished with the construction of a system reliability block diagram that represents the reliability relationships of the components in the system.  From this model, the system reliability impact of different component modifications can be estimated and considered alongside the costs that would be incurred in the process of making those modifications.  It is then possible to perform an optimization analysis for this problem, finding the best combination of component reliability improvements that meet or exceed the performance goals at the lowest cost.&lt;br /&gt;
&lt;br /&gt;
===Importance Measures and FMEA/FMECA===&lt;br /&gt;
&lt;br /&gt;
Traditional Failure Mode and Effects analysis (FMEA/FMECA) relies on Risk Priority Numbers (RPNs) or criticality calculations to identify and prioritize the significance/importance of different failure modes.  The RPN methodology (and to some extent, the criticality methodology) tend to be subjective.  When conducting these types of analyses, one may wish to incorporate more quantitative metrics, such as the importance measure presented here and/or the RS FCI and RS DECI for repairable systems (which are discussed in later chapters).  ReliaSoft&#039;s Xfmea software can be used to export an FMEA/FMECA analysis to BlockSim.  The documentation that accompanies Xfmea provides more information on FMEA/FMECA, including both methods of risk assessment.&lt;br /&gt;
&lt;br /&gt;
=Improving Reliability=&lt;br /&gt;
Reliability engineers are very often called upon to make decisions as to whether to improve a certain component or components in order to achieve a minimum required system reliability.  There are two approaches to improving the reliability of a system: fault avoidance and fault tolerance.  Fault avoidance is achieved by using high-quality and high-reliability components and is usually less expensive than fault tolerance.  Fault tolerance, on the other hand, is achieved by redundancy.  Redundancy can result in increased design complexity and increased costs through additional weight, space, etc.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before deciding whether to improve the reliability of a system by fault tolerance or fault avoidance, a reliability assessment for each component in the system should be made.  Once the reliability values for the components have been quantified, an analysis can be performed in order to determine if that system&#039;s reliability goal will be met.  If it becomes apparent that the system&#039;s reliability will not be adequate to meet the desired goal at the specified mission duration, steps can be taken to determine the best way to improve the system&#039;s reliability so that it will reach the desired target.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a system with three components connected reliability-wise in series.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 70%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 80% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 90%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 85%, is required for this system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}=50.4%&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, this is far short of the system&#039;s required reliability performance.  It is apparent that the reliability of the system&#039;s constituent components will need to be increased in order for the system to meet its goal.  First, we will try increasing the reliability of one component at a time to see whether the reliability goal can be achieved.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig9 shows that even by raising the individual component reliability to a hypothetical value of 1 (100% reliability, which implies that the component will never fail), the overall system reliability goal will not be met by improving the reliability of just one component.  The next logical step would be to try to increase the reliability of two components.  The question now becomes: which two?  One might also suggest increasing the reliability of all three components.  A basis for making such decisions needs to be found in order to avoid the ``trial and error&#039;&#039; aspect of altering the system&#039;s components randomly in an attempt to achieve the system reliability goal.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.9.gif|thumb|center|400px|Change in system reliability of a three-unit series system due to increasing the reliability of just one component.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As we have seen, the reliability goal for the preceding example could not be achieved by increasing the reliability of just one component.  There are cases, however, where increasing the reliability of one component results in achieving the system reliability goal.  Consider, for example, a system with three components connected reliability-wise in parallel.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 60%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 70% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 80%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 99%, is required for this system.  The initial system reliability is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-(1-0.6)\cdot (1-0.7)\cdot (1-0.8)=0.976&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current system reliability is inadequate to meet the goal.  Once again, we can try to meet the system reliability goal by raising the reliability of just one of the three components in the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From Figure fig10, it can be seen that the reliability goal can be reached by improving Component 1, Component 2 or Component 3.  The reliability engineer is now faced with another dilemma:  which component&#039;s reliability should be improved? This presents a new aspect to the problem of allocating the reliability of the system.  Since we know that the system reliability goal can be achieved by increasing at least one unit, the question becomes one of how to do this most efficiently and cost effectively.  We will need more information to make an informed decision as to how to go about improving the system&#039;s reliability.  How much does each component need to be improved for the system to meet its goal?  How feasible is it to improve the reliability of each component?  Would it actually be more efficient to slightly raise the reliability of two or three components rather than radically improving only one?&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to answer these questions, we must introduce another variable into the problem &amp;lt;math&amp;gt;:\ \ \ &amp;lt;/math&amp;gt; cost.  Cost does not necessarily have to be in dollars.  It could be described in terms of non-monetary resources, such as time.  By associating cost values to the reliabilities of the system&#039;s components, we can find an optimum design that will provide the required reliability at a minimum cost.&lt;br /&gt;
&lt;br /&gt;
===Cost/Penalty Function===&lt;br /&gt;
&lt;br /&gt;
There is always a cost associated with changing a design due to change of vendors, use of higher-quality materials, retooling costs, administrative fees, etc.  The cost as a function of the reliability for each component must be quantified before attempting to improve the reliability.  Otherwise, the design changes may result in a system that is needlessly expensive or overdesigned.  Developing the ``cost of reliability&#039;&#039; relationship will give the engineer an understanding of which components to improve and how to best concentrate the effort and allocate resources in doing so.  The first step will be to obtain a relationship between the cost of improvement and reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.10.png|thumb|center|400px|Meeting a reliability goal requirement by increasing a component&#039;s reliability]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The preferred approach would be to formulate the cost function from actual cost data.  This can be done from past experience.  If a reliability growth program is in place, the costs associated with each stage of improvement can also be quantified.  Defining the different costs associated with different vendors or different component models is also useful in formulating a model of component cost as a function of reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, there are many cases where no such information is available.  For this reason, a general (default) behavior model of the cost versus the component&#039;s reliability was developed for performing reliability optimization in BlockSim.  The objective of this function is to model an overall cost behavior for all types of components.  Of course, it is impossible to formulate a model that will be precisely applicable to every situation; but the proposed relationship is general enough to cover most applications.  In addition to the default model formulation, BlockSim does allow the definition of user-defined cost models.&lt;br /&gt;
&lt;br /&gt;
====Quantifying the Cost/Penalty Function====&lt;br /&gt;
&lt;br /&gt;
One needs to quantify a cost function for each component,  &amp;lt;math&amp;gt;{{C}_{i}}&amp;lt;/math&amp;gt; , in terms of the reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , of each component, or:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}=f({{R}_{i}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This function should:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the current reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Current}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the maximum possible reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Max}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Allow for different levels of difficulty (or cost) in increasing the reliability of each component.  It can take into account:&amp;lt;br&amp;gt;&lt;br /&gt;
::o	design issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	supplier issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	state of technology.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	time-to-market issues, etc.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus, for the cost function to comply with these needs, the following conditions should be adhered to:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be constrained by the minimum and maximum reliabilities of each component (i.e. reliability must be less than one and greater than the current reliability of the component or at least greater than zero).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should not be linear, but rather quantify the fact that it is incrementally harder to improve reliability.  For example, it is considerably easier to increase the reliability from 90% to 91% than to increase it from 99.99% to 99.999%, even though the increase is larger in the first case.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be asymptotic to the maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The following default cost function (also used in BlockSim) adheres to all of these conditions and acts like a penalty function for increasing a component&#039;s reliability.  Furthermore, an exponential behavior for the cost is assumed since it should get exponentially more difficult to increase the reliability. See Mettas [21]. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})&amp;lt;/math&amp;gt;  is the penalty (or cost) function as a function of component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  is the feasibility (or cost index) of improving a component&#039;s reliability relative to the other components in the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{min,i}}&amp;lt;/math&amp;gt;  is the current reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{max,i}}&amp;lt;/math&amp;gt;  is the maximum achievable reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
Note that this penalty function is dimensionless.  It essentially acts as a weighting factor that describes the difficulty in increasing the component reliability from its current value, relative to the other components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examining the cost function given by Eqn. (Default Cost), the following observations can be made:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability departs from the minimum or current value of reliability.  It is assumed that the reliabilities for the components will not take values any lower than they already have.  Depending on the optimization, a component&#039;s reliability may not need to be increased from its current value but it will not drop any lower.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability approaches the maximum achievable reliability.  This is a reliability value that is approached asymptotically as the cost increases but is never actually reached.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost is a function of the range of improvement, which is the difference between the component&#039;s initial reliability and the corresponding maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The exponent in Eqn. (Default Cost) approaches infinity as the component&#039;s reliability approaches its maximum achievable value.  This means that it is easier to increase the reliability of a component from a lower initial value.  For example, it is easier to increase a component&#039;s reliability from 70% to 75% than increasing its reliability from 90% to 95%.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====The Feasibility Term,  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
The feasibility term in Eqn. (Default Cost) is a constant (or an equation parameter) that represents the difficulty in increasing a component&#039;s reliability relative to the rest of the components in the system.  Depending on the design complexity, technological limitations, etc., certain components can be very hard to improve.  Clearly, the more difficult it is to improve the reliability of the component, the greater the cost.  Figure feasplot illustrates the behavior of the function defined in Eqn. (Default Cost) for different values of  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; .  It can be seen that the lower the feasibility value, the more rapidly the cost function approaches infinity.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Several methods can be used to obtain a feasibility value.  Weighting factors for allocating reliability have been proposed by many authors and can be used to quantify feasibility.  These weights depend on certain factors of influence, such as the complexity of the component, the state of the art, the operational profile, the criticality, etc.  Engineering judgment based on past experience, supplier quality, supplier availability and other factors can also be used in determining a feasibility value.  Overall, the assignment of a feasibility value is going to be a subjective process.  Of course, this problem is negated if the relationship between the cost and the reliability for each component is known because one can use regression methods to estimate the parameter value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.11.gif|thumb|center|400px|Behavior of the cost function for different feasibility values.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Maximum Achievable Reliability====&lt;br /&gt;
&lt;br /&gt;
For the purposes of reliability optimization, we also need to define a limiting reliability that a component will approach, but not reach.  The costs near the maximum achievable reliability are very high and the actual value for the maximum reliability is usually dictated by technological or financial constraints.  In deciding on a value to use for the maximum achievable reliability, the current state of the art of the component in question and other similar factors will have to be considered.  In the end, a realistic estimation based on engineering judgment and experience will be necessary to assign a value to this input.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that the time associated with this maximum achievable reliability is the same as that of the overall system reliability goal.  Almost any component can achieve a very high reliability value, provided the mission time is short enough.  For example, a component with an exponential distribution and a failure rate of one failure per hour has a reliability that drops below 1% for missions greater than five hours.  However, it can achieve a reliability of 99.9% as long as the mission is no longer than four seconds.  For the purposes of optimization in BlockSim, the reliability values of the components are associated with the time for which the system reliability goal is specified.  For example, if the problem is to achieve a system goal of 99% reliability at 1,000 hours, the maximum achievable reliability values entered for the individual components would be the maximum reliability that each component could attain for a mission of 1,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As the component reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , approaches the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function approaches infinity.  The maximum achievable reliability acts as a scale parameter for the cost function.  By decreasing  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function is compressed between  &amp;lt;math&amp;gt;{{R}_{i,min}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , as shown in Figure oldfig5.&lt;br /&gt;
 &lt;br /&gt;
====Cost Function====&lt;br /&gt;
Once the cost functions for the individual components have been determined, it becomes necessary to develop an expression for the overall system cost.  This takes the form of:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{s}}({{R}_{G}})={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+...+{{C}_{n}}({{R}_{n}}),i=1,2,...,n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, the cost of the system is simply the sum of the costs of its components.  This is regardless of the form of the individual component cost functions.  They can be of the general behavior model in BlockSim or they can be user-defined.   Once the overall cost function for the system has been defined, the problem becomes one of minimizing the cost function while remaining within the constraints defined by the target system reliability and the reliability ranges for the components.  The latter constraints in this case are defined by the minimum and maximum reliability values for the individual components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.12.png|thumb|center|400px|Effect5 of the maximum achievable reliability on the cost function.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim employs a nonlinear programming technique to minimize the system cost function.  The system has a minimum (current) and theoretical maximum reliability value that is defined by the minimum and maximum reliabilities of the components and by the way the system is configured.  That is, the structural properties of the system are accounted for in the determination of the optimum solution.  For example, the optimization for a system of three units in series will be different from the optimization for a system consisting of the same three units in parallel.  The optimization occurs by varying the reliability values of the components within their respective constraints of maximum and minimum reliability in a way that the overall system goal is achieved.  Obviously, there can be any number of different combinations of component reliability values that might achieve the system goal.  The optimization routine essentially finds the combination that results in the lowest overall system cost. &lt;br /&gt;
&lt;br /&gt;
====Determining the Optimum Allocation Scheme====&lt;br /&gt;
&lt;br /&gt;
To determine the optimum reliability allocation, the analyst first determines the system reliability equation (the objective function).  As an example, and again for a trivial system with three components in series, this would be:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a target reliability of 90% is sought, then Eqn. (optAlloc) is recast as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The objective now is to solve for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  so that the equality in Eqn. (optAlloc90) is satisfied.  To obtain an optimum solution, we also need to use our cost functions (i.e. define the total allocation costs) as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
With the cost equation defined, then the optimum values for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  are the values that satisfy the reliability requirement, Eqn. (optAlloc90), at the minimum cost, Eqn. (optcost).  BlockSim uses this methodology during the optimization task.&lt;br /&gt;
&lt;br /&gt;
====Defining a Feasibility Policy in BlockSim====&lt;br /&gt;
&lt;br /&gt;
In BlockSim you can choose to use the default feasibility function, as defined by Eqn. (Default Cost), or use your own function.  Figure BSfvalues illustrates the use of the default values using the slider control. Figure BSFcustom shows the use of an associated feasibility policy to create a user-defined cost function.  When defining your own cost function, you should be aware of/adhere to the following guidelines:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Because the cost functions are evaluated relative to each other, they should be correlated.  In other words, if one function evaluates to 10,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=10&amp;lt;/math&amp;gt;  for one block and 20 for another,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=20&amp;lt;/math&amp;gt; , then the implication is that there is a 1 to 2 cost relation.  &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Do not mix your own function with the software&#039;s default functions unless you have verified that your cost functions are defined and correlated to the default cost functions, as defined by Eqn. (Default Cost).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Your function should adhere to the guidelines presented earlier.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Lastly, and since the evaluation is relative, it is preferable to use the pre-defined functions unless you have a compelling reason (or data) to do otherwise.  The last section in this chapter describes cases where user-defined functions are preferred.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.13.png|thumb|center|300px|Setting the default feasibility function in BlockSim with the feasibility slider. Note that the feasibility slider displays values, &#039;&#039;SV&#039;&#039;, from 1 to 9 when moved by the user, with SV=9 being the hardest. The relationship between &#039;&#039;f&#039;&#039; and &#039;&#039;SV&#039;&#039; is ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.14.png|thumb|center|400px|Setting a user-defined feasibility function in BlockSim utilizing an assiciated feasibility policy. Any user-defined equation can be entered as a function of &#039;&#039;R.&#039;&#039;]]&lt;br /&gt;
&lt;br /&gt;
=Implementing the Optimization=&lt;br /&gt;
&lt;br /&gt;
As was mentioned earlier, there are two different methods of implementing the changes suggested by the reliability optimization routine: fault tolerance and fault avoidance.  When the optimized component reliabilities have been determined, it does not matter which of the two methods is employed to realize the optimum reliability for the component in question.  For example, suppose we have determined that a component must have its reliability for a certain mission time raised from 50% to 75%.  The engineer must now decide how to go about implementing the increase in reliability.  If the engineer decides to do this via fault avoidance, another component must be found (or the existing component must be redesigned) so that it will perform the same function with a higher reliability.  On the other hand, if the engineer decides to go the fault tolerance route, the optimized reliability can be achieved merely by placing a second identical component in parallel with the first one.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, the method of implementing the reliability optimization is going to be related to the cost function and this is something the reliability engineer must take into account when deciding on what type of cost function is used for the optimization.  In fact, if we take a closer look at the fault tolerance scheme, we can see some parallels with the general behavior cost model included in BlockSim.  For example, consider a system that consists of a single unit.  The cost of that unit, including all associated mounting and hardware costs, is one dollar.  The reliability of this unit for a given mission time is 30%.  It has been determined that this is inadequate and that a second component is to be added in parallel to increase the reliability.  Thus, the reliability for the two-unit parallel system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-{{(1-0.3)}^{2}}=0.51\text{ or }51%&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
So, the reliability has increased by a value of 21% and the cost has increased by one dollar.  In a similar fashion, we can continue to add more units in parallel, thus increasing the reliability and the cost.  We now have an array of reliability values and the associated costs that we can use to develop a cost function for this fault tolerance scheme.  Figure costredundant shows the relationship between cost and reliability for this example.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As can be seen, this looks quite similar to the general behavior cost model presented earlier.  In fact, a standard regression analysis available in Weibull++ indicates that an exponential model fits this cost model quite well.   The function is given by the following equation, where  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is the cost in dollars and  &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt;  is the fractional reliability value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C(R)=0.3756\cdot {{e}^{3.1972\cdot R}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.15.gif|thumb|center|400px|Cost function for redundant parallel units.]]&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
&lt;br /&gt;
Consider a system consisting of three components connected reliability-wise in series.  Assume the objective reliability for the system is 90% for a mission time of 100 hours.  Five cases will be considered for the allocation problem. See Mettas [21].&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1 - All three components are identical with times-to-failure that are described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.318 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 312 hours. All three components have the same feasibility value of Moderate (5).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2 - Same as in Case 1, but Component 1 has a feasibility of Easy, Component 2 has a feasibility of Moderate and Component 3 has a feasibility of Hard.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3 - Component 1 has 70% reliability, Component 2 has 80% reliability and Component 3 has 90% reliability, all for a mission duration of 100 hours.  All three components have the same feasibility of Easy.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 4 - Component 1 has 70% reliability and Easy feasibility, Component 2 has 80% reliability and Moderate feasibility, and Component 3 has 90% reliability and Hard feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 5 - Component 1 has 70% reliability and Hard feasibility, Component 2 has 80% reliability and Easy feasibility and Component 3 has 90% reliability and Moderate feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In all cases, the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , for each component is 99.9% for a mission duration of 100 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.16.gif|thumb|center|300px|Optimization inputs in BlockSim&#039;s Analytical QCP for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Solution====&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Case 1&#039;&#039;&#039; - The reliability equation for Case 1 is: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the equality desired is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}(t=100)\cdot {{R}_{2}}(t=100)\cdot {{R}_{3}}(t=100)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{1,2,3}}={{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cost or feasibility function is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{1,2,3}}({{R}_{1,2,3}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
And where  &amp;lt;math&amp;gt;{{R}_{\max _{1,2,3}^{}}}=0.999&amp;lt;/math&amp;gt;  (arbitrarily set),  &amp;lt;math&amp;gt;{{R}_{\min _{1,2,3}^{}}}&amp;lt;/math&amp;gt;  computed from the reliability function of each component at the time of interest,  &amp;lt;math&amp;gt;t=100&amp;lt;/math&amp;gt; , or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{\min _{1,2,3}^{}}}= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( \tfrac{100}{312} \right)}^{1.318}}}} \\ &lt;br /&gt;
= &amp;amp; 0.79995  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  obtained from: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
f= &amp;amp; \left( 1-\frac{5}{10} \right) \\ &lt;br /&gt;
= &amp;amp; 0.5  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The solution,  &amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}&amp;lt;/math&amp;gt; , is the one that satisfies Eqn. (exbjective2) while minimizing Eqn. (exonstraint).  In this case (and since all the components are identical), the target reliability is found to be: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}(t=100)=0.9655&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Figures QCPOpt and OptResults show related BlockSim screens.  Based on this, each component&#039;s reliability should be at least 96.55% at 100 hours in order for the system&#039;s reliability to be 90% at 100 hours.  Note the column labeled N.E.P.U. in the Results Panel shown in Figure OptResults.  This stands for &amp;quot;Number of Equivalent Parallel Units&amp;quot; and represents the number of redundant units that would be required to bring that particular component up to the recommended reliability.  In the case where the fault tolerance approach is to be implemented, the N.E.P.U value should be rounded up to an integer.  Therefore, some manipulation by the engineer is required in order to ensure that the chosen integer values will yield the required system reliability goal (or exceed it).  In addition, further cost analysis should be performed in order to account for the costs of adding redundancy to the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Additionally, and when the results have been obtained, the engineer may wish to re-scale the components based on their distribution parameters instead of the fixed reliability value.  In the case of these components, one may wish to re-scale the scale parameter of the distribution ,  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt; , for the components, or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{100}{\eta } \right)}^{1.318}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.17.png|thumb|center|300px|Optimization results for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Which yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\eta }_{{{O}_{i}}}}=1269.48&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The Parameter Experimenter in BlockSim can also be used for this (Figure paramexper).  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The results from the other cases can be obtained in a similar fashion.  The results for Cases 1 through 5 are summarized next.&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{matrix}&lt;br /&gt;
   {} &amp;amp; Case 1 &amp;amp; Case 2 &amp;amp; Case 3 &amp;amp; Case 4 &amp;amp; Case 5  \\&lt;br /&gt;
   Component 1 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9874} &amp;amp; \text{0}\text{.9552} &amp;amp; \text{0}\text{.9790} &amp;amp; \text{0}\text{.9295}  \\&lt;br /&gt;
   Component 2 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9633} &amp;amp; \text{0}\text{.9649} &amp;amp; \text{0}\text{.9553} &amp;amp; \text{0}\text{.9884}  \\&lt;br /&gt;
   Component 3 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9463} &amp;amp; \text{0}\text{.9765} &amp;amp; \text{0}\text{.9624} &amp;amp; \text{0}\text{.9797}  \\&lt;br /&gt;
\end{matrix}&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 2&#039;&#039;&#039; - It can be seen that the highest reliability was allocated to Component 1 with the Easy feasibility.  The lowest reliability was assigned to Component 3 with the Hard feasibility.  This makes sense in that an optimized reliability scheme will call for the greatest reliability changes in those components that are the easiest to change.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 3&#039;&#039;&#039; - The components were different but had the same feasibility values.  &lt;br /&gt;
&lt;br /&gt;
[[Image:BS6.18.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, all three components have the same opportunity for improvement.  This case differs from Cases 1 and 2 since there are two factors, not present previously, that will affect the outcome of the allocation in this case.  First, each component in this case has a different reliability importance (impact of a component on the system&#039;s reliability); whereas in Cases 1 and 2, all three components were identical and had the same reliability importance.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure relimp shows the reliability importance for each component, where it can be seen that Component 1 has the greatest reliability importance and Component 3 has the smallest (this reliability importance also applies in Cases 4 and 5).  This indicates that the reliability of Component 1 should be significantly increased because it has the biggest impact on the overall system reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In addition, each component&#039;s cost function in Case 3 also depends on the difference between each component&#039;s initial reliability and its corresponding maximum achievable reliability.  (In Cases 1 and 2 this was not an issue because the components were identical.)  The greater this difference, the greater the cost of improving the reliability of a particular component relative to the other two components.  This difference between the initial reliability of a component and its maximum achievable reliability is called the range of improvement for that component.  Since all three components have the same maximum achievable reliability, Component 1, with the largest range for improvement, is the most cost efficient component to improve.  The improvement ranges for all three components are illustrated in Figure Rangeofimprovement. At the same time, however, there is a reliability value between the initial and the maximum achievable reliability beyond which it becomes cost prohibitive to improve any further.  This reliability value is dictated by the feasibility value.  From the table of results, it can be seen that in Case 3 there was a 25.52% improvement for Component 1, 16.49% for Component 2 and 7.65% for Component 3.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.19.png|thumb|center|300px|Reliability importance for Example 2, Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.20.png|thumb|center|300px|Range of improvement for each component for Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 4&#039;&#039;&#039; - As opposed to Case 3, Component 1 was assigned an even greater increase of 27.9%, with Components 2 and 3 receiving lesser increases than in Case 3, of 15.53% and 6.24% respectively.  This is due to the fact that Component 1 has an Easy feasibility and Component 3 has a Hard feasibility, which means that it is more difficult to increase the reliability of Component 3 than to increase the reliability of Component 1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 5&#039;&#039;&#039; - The feasibility values here are reversed with Component 1 having a Hard feasibility and Component 3 an Easy feasibility.  The recommended increase in Component 1&#039;s reliability is less compared to its increase for Cases 3 and 4.  Note, however, that Components 2 and 3 still received a smaller increase in reliability than Component 1 because their ranges of improvement are smaller.  In other words, Component 3 was assigned the smallest increase in reliability in Cases 3, 4 and 5 because its initial reliability is very close to its maximum achievable reliability. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Setting Specifications=&lt;br /&gt;
&lt;br /&gt;
This methodology could also be used to arrive at initial specifications for a set of components.  In the prior examples, we assumed a current reliability for the components.  One could repeat these steps by choosing an arbitrary (lower) initial reliability for each component, thus allowing the algorithm to travel up to the target.  When doing this, it is important to keep in mind the fact that both the distance from the target (the distance from the initial arbitrary value and the target value) for each component is also a significant contributor to the final results, as presented in the prior example.  If one wishes to arrive at the results using only the cost functions then it may be advantageous to set equal initial reliabilities for all components.&lt;br /&gt;
&lt;br /&gt;
=Other Notes on User-Defined Cost Functions=&lt;br /&gt;
&lt;br /&gt;
The optimization method in BlockSim is a very powerful tool for allocating reliability to the components of a system while minimizing an overall cost of improvement.  The default cost function in BlockSim was derived in order to model a general relationship between the cost and the component reliability.  However, if actual cost information is available, then one can use the cost data instead of using the default function.  Additionally, one can also view the feasibility in the default function as a measure of the difficulty in increasing the reliability of the component relative to the rest of the components to be optimized, assuming that they also follow the same cost function with the corresponding feasibility values.  If fault tolerance is a viable option, a reliability cost function for adding parallel units can be developed as demonstrated previously.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Another method for developing a reliability cost function would be to obtain different samples of components from different suppliers and test the samples to determine the reliability of each sample type.  From this data, a curve could be fitted through standard regression techniques and an equation defining the cost as a function of reliability could be developed.  Figure RGplot shows such a curve.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, and in cases where a reliability growth program is in place, the simplest way of obtaining a relationship between cost and reliability is by associating a cost to each development stage of the growth process.  Reliability growth models such as the Crow (AMSAA), Duane, Gompertz and Logistic models can be used to describe the cost as a function of reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.21.png|thumb|center|300px|Typical reliability growth curve generated using ReliaSoft&#039;s Reliability Growth software.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a reliability growth model has been successfully implemented, the development costs over the respective development time stages can be applied to the growth model, resulting in equations that describe reliability/cost relationships.  These equations can then be entered into BlockSim as user-defined cost functions (feasibility policies).  The only potential drawback to using growth model data is the lack of flexibility in applying the optimum results.  Making the cost projection for future stages of the project would require the assumption that development costs will be accrued at a similar rate in the future, which may not always be a valid assumption.  Also, if the optimization result suggests using a high reliability value for a component, it may take more time than is allotted for that project to attain the required reliability given the current reliability growth of the project.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15643</id>
		<title>Reliability Importance and Optimized Reliability Allocation (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15643"/>
		<updated>2012-02-13T23:39:01Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Static Reliability Importance */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|6}}&lt;br /&gt;
&lt;br /&gt;
=Component Reliability Importance=&lt;br /&gt;
===Static Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Once the reliability of a system has been determined, engineers are often faced with the task of identifying the least reliable component(s) in the system in order to improve the design.  For example, it was observed in Chapter 4 that the least reliable component in a series system has the biggest effect on the system reliability.  In this case, if the reliability of the system is to be improved, then the efforts can best be concentrated on improving the reliability of that component first.   In simple systems such as a series system, it is easy to identify the weak components.  However, in more complex systems this becomes quite a difficult task.  For complex systems, the analyst needs a mathematical approach that will provide the means of identifying and quantifying the importance of each component in the system.&lt;br /&gt;
&lt;br /&gt;
Using reliability importance measures is one method of identifying the relative importance of each component in a system with respect to the overall reliability of the system.  The reliability importance,  &amp;lt;math&amp;gt;{{I}_{R}}&amp;lt;/math&amp;gt; , of component  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;  in a system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  components is given by [[Appendix D: Weibull References | Leemis [17]]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{i}}}   \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{s}}&amp;lt;/math&amp;gt;  is the system reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  is the component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The value of the reliability importance given by Eqn.6.1 depends both on the reliability of a component and its corresponding position in the system.  In Chapter 4 we observed that for a simple series system (three components in series with reliabilities of 0.7, 0.8 and 0.9) the rate of increase of the system reliability was greatest when the least reliable component was improved.  In other words, it was observed that Component 1 had the largest reliability importance in the system relative to the other two components (see Figure 6.1).  The same conclusion can be drawn by using Eqn.6.1 and obtaining the reliability importance in terms of a value for each component.&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Using BlockSim, the reliability importance values for these components can be calculated with Eqn.6.1.  Using the plot option and selecting the Static Reliability Importance plot type, Figure 6.2 can be obtained.  Note that the time input required to create this plot is irrelevant for this example because the components are static.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The values shown in Figure fig1a for each component were obtained using Eqn.6.1.  The reliability equation for this series system was given by: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}   \ (eqn 2)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Taking the partial derivative of Eqn.6.2 with respect to  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{I}_{{{R}_{1}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{1}}}= &amp;amp; {{R}_{2}}{{R}_{3}} \\ &lt;br /&gt;
= &amp;amp; 0.8\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.72  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus the reliability importance of Component 1 is  &amp;lt;math&amp;gt;{{I}_{{{R}_{1}}}}=&amp;lt;/math&amp;gt;  0.72.  The reliability importance values for Components 2 and 3 are obtained in a similar manner.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.1.png|thumb|center|300px|Rate of change of system reliability when increasing the reliability of each component.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:6.2.gif|thumb|center|300px|Static reliability importance plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Time-Dependent Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
\The same concept applies if the components have a time-varying reliability.  That is, if  &amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt; , then one could compute  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  at any time  &amp;lt;math&amp;gt;x,&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}{{(t)}_{_{t=x}}}.&amp;lt;/math&amp;gt;   This is quantified in Eqn. (importance time). &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)=\frac{\partial {{R}_{s}}(t)}{\partial {{R}_{i}}(t)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In turn, this can be viewed as either a static plot (at a given time) or as time-varying plot, as illustrated in the next figures.  Specifically, Figures Ch6fig3, Ch6fig4 and Ch6fig5 present the analysis for three components configured reliability-wise in series following a Weibull distribution with  &amp;lt;math&amp;gt;\beta =3&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=1,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\eta }_{2}}=2,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=3,000&amp;lt;/math&amp;gt; .  Figure Ch6fig3 shows a bar chart of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  while Figure Ch6fig4 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  in BlockSim&#039;s tableau chart format.  In this chart, the area of the square is  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt; .  Lastly, Figure Ch6fig5 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  vs. time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that a system has failure modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; .  Furthermore, assume that failure of the entire system will occur if:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  occurs.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  occur.&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, assume the following failure probabilities for each mode.&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  have a mean time to occurrence of 1,000 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=1,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  has a mean time to occurrence of 100 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=100).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a mean time to occurrence of 700,000, 1,000,000 and 2,000,000 hours respectively (i.e. exponential with  &amp;lt;math&amp;gt;MTT{{F}_{B}}=700,000&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;MTT{{F}_{C}}=1,000,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;MTT{{F}_{F}}=2,000,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examine the mode importance for operating times of 100 and 500 hours.&lt;br /&gt;
&lt;br /&gt;
[[Image:6.3.gif|thumb|center|400px|Static Reliability Importance plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.4.png|thumb|center|400px|Static Reliability Importance tableau plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.5.png|thumb|center|400px|Reliability Importance vs. time plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
The RBD for this example is (from Chapter 4, Example 18):&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6ex1.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig6 illustrates  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt; .  It can be seen that even though  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a much rarer rate of occurrence, they are much more significant at 100 hours.  By 500 hours,  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt; , the effects of the lower reliability components become greatly pronounced and thus they become more important, as can be seen in Figure Ch6fig7.  Finally, the behavior of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  can be observed in Figure Ch6fig8.  Note that not all lines are plainly visible in Figure Ch6fig8 due to overlap.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Reliability Allocation=&lt;br /&gt;
&lt;br /&gt;
In the process of  developing a new product, the engineer is often faced with the task of designing a system that conforms to a set of reliability specifications.  The engineer is given the goal for the system and must then develop a design that will achieve the desired reliability of the system, while performing all of the system&#039;s intended functions at a minimum cost. This involves a balancing act of determining how to allocate reliability to the components in the system so the system will meet its reliability goal while at the same time ensuring that the system meets all of the other associated performance specifications.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.6.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.7.gif|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.8.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
The simplest method for allocating reliability is to distribute the reliabilities uniformly among all components. For example, suppose a system with five components in series has a reliability objective of 90% for a given operating time. The uniform allocation of the objective to all components would require each component to have a reliability of 98% for the specified operating time, since  &amp;lt;math&amp;gt;{{0.98}^{5}}\tilde{=}0.90&amp;lt;/math&amp;gt;. While this manner of allocation is easy to calculate, it is generally not the best way to allocate reliability for a system. The optimum method of allocating reliability would take into account the cost or relative difficulty of improving the reliability of different subsystems or components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability optimization process begins with the development of a model that represents the entire system.  This is accomplished with the construction of a system reliability block diagram that represents the reliability relationships of the components in the system.  From this model, the system reliability impact of different component modifications can be estimated and considered alongside the costs that would be incurred in the process of making those modifications.  It is then possible to perform an optimization analysis for this problem, finding the best combination of component reliability improvements that meet or exceed the performance goals at the lowest cost.&lt;br /&gt;
&lt;br /&gt;
===Importance Measures and FMEA/FMECA===&lt;br /&gt;
&lt;br /&gt;
Traditional Failure Mode and Effects analysis (FMEA/FMECA) relies on Risk Priority Numbers (RPNs) or criticality calculations to identify and prioritize the significance/importance of different failure modes.  The RPN methodology (and to some extent, the criticality methodology) tend to be subjective.  When conducting these types of analyses, one may wish to incorporate more quantitative metrics, such as the importance measure presented here and/or the RS FCI and RS DECI for repairable systems (which are discussed in later chapters).  ReliaSoft&#039;s Xfmea software can be used to export an FMEA/FMECA analysis to BlockSim.  The documentation that accompanies Xfmea provides more information on FMEA/FMECA, including both methods of risk assessment.&lt;br /&gt;
&lt;br /&gt;
=Improving Reliability=&lt;br /&gt;
Reliability engineers are very often called upon to make decisions as to whether to improve a certain component or components in order to achieve a minimum required system reliability.  There are two approaches to improving the reliability of a system: fault avoidance and fault tolerance.  Fault avoidance is achieved by using high-quality and high-reliability components and is usually less expensive than fault tolerance.  Fault tolerance, on the other hand, is achieved by redundancy.  Redundancy can result in increased design complexity and increased costs through additional weight, space, etc.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before deciding whether to improve the reliability of a system by fault tolerance or fault avoidance, a reliability assessment for each component in the system should be made.  Once the reliability values for the components have been quantified, an analysis can be performed in order to determine if that system&#039;s reliability goal will be met.  If it becomes apparent that the system&#039;s reliability will not be adequate to meet the desired goal at the specified mission duration, steps can be taken to determine the best way to improve the system&#039;s reliability so that it will reach the desired target.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a system with three components connected reliability-wise in series.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 70%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 80% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 90%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 85%, is required for this system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}=50.4%&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, this is far short of the system&#039;s required reliability performance.  It is apparent that the reliability of the system&#039;s constituent components will need to be increased in order for the system to meet its goal.  First, we will try increasing the reliability of one component at a time to see whether the reliability goal can be achieved.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig9 shows that even by raising the individual component reliability to a hypothetical value of 1 (100% reliability, which implies that the component will never fail), the overall system reliability goal will not be met by improving the reliability of just one component.  The next logical step would be to try to increase the reliability of two components.  The question now becomes: which two?  One might also suggest increasing the reliability of all three components.  A basis for making such decisions needs to be found in order to avoid the ``trial and error&#039;&#039; aspect of altering the system&#039;s components randomly in an attempt to achieve the system reliability goal.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.9.gif|thumb|center|400px|Change in system reliability of a three-unit series system due to increasing the reliability of just one component.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As we have seen, the reliability goal for the preceding example could not be achieved by increasing the reliability of just one component.  There are cases, however, where increasing the reliability of one component results in achieving the system reliability goal.  Consider, for example, a system with three components connected reliability-wise in parallel.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 60%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 70% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 80%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 99%, is required for this system.  The initial system reliability is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-(1-0.6)\cdot (1-0.7)\cdot (1-0.8)=0.976&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current system reliability is inadequate to meet the goal.  Once again, we can try to meet the system reliability goal by raising the reliability of just one of the three components in the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From Figure fig10, it can be seen that the reliability goal can be reached by improving Component 1, Component 2 or Component 3.  The reliability engineer is now faced with another dilemma:  which component&#039;s reliability should be improved? This presents a new aspect to the problem of allocating the reliability of the system.  Since we know that the system reliability goal can be achieved by increasing at least one unit, the question becomes one of how to do this most efficiently and cost effectively.  We will need more information to make an informed decision as to how to go about improving the system&#039;s reliability.  How much does each component need to be improved for the system to meet its goal?  How feasible is it to improve the reliability of each component?  Would it actually be more efficient to slightly raise the reliability of two or three components rather than radically improving only one?&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to answer these questions, we must introduce another variable into the problem &amp;lt;math&amp;gt;:\ \ \ &amp;lt;/math&amp;gt; cost.  Cost does not necessarily have to be in dollars.  It could be described in terms of non-monetary resources, such as time.  By associating cost values to the reliabilities of the system&#039;s components, we can find an optimum design that will provide the required reliability at a minimum cost.&lt;br /&gt;
&lt;br /&gt;
===Cost/Penalty Function===&lt;br /&gt;
&lt;br /&gt;
There is always a cost associated with changing a design due to change of vendors, use of higher-quality materials, retooling costs, administrative fees, etc.  The cost as a function of the reliability for each component must be quantified before attempting to improve the reliability.  Otherwise, the design changes may result in a system that is needlessly expensive or overdesigned.  Developing the ``cost of reliability&#039;&#039; relationship will give the engineer an understanding of which components to improve and how to best concentrate the effort and allocate resources in doing so.  The first step will be to obtain a relationship between the cost of improvement and reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.10.png|thumb|center|400px|Meeting a reliability goal requirement by increasing a component&#039;s reliability]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The preferred approach would be to formulate the cost function from actual cost data.  This can be done from past experience.  If a reliability growth program is in place, the costs associated with each stage of improvement can also be quantified.  Defining the different costs associated with different vendors or different component models is also useful in formulating a model of component cost as a function of reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, there are many cases where no such information is available.  For this reason, a general (default) behavior model of the cost versus the component&#039;s reliability was developed for performing reliability optimization in BlockSim.  The objective of this function is to model an overall cost behavior for all types of components.  Of course, it is impossible to formulate a model that will be precisely applicable to every situation; but the proposed relationship is general enough to cover most applications.  In addition to the default model formulation, BlockSim does allow the definition of user-defined cost models.&lt;br /&gt;
&lt;br /&gt;
====Quantifying the Cost/Penalty Function====&lt;br /&gt;
&lt;br /&gt;
One needs to quantify a cost function for each component,  &amp;lt;math&amp;gt;{{C}_{i}}&amp;lt;/math&amp;gt; , in terms of the reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , of each component, or:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}=f({{R}_{i}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This function should:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the current reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Current}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the maximum possible reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Max}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Allow for different levels of difficulty (or cost) in increasing the reliability of each component.  It can take into account:&amp;lt;br&amp;gt;&lt;br /&gt;
::o	design issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	supplier issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	state of technology.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	time-to-market issues, etc.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus, for the cost function to comply with these needs, the following conditions should be adhered to:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be constrained by the minimum and maximum reliabilities of each component (i.e. reliability must be less than one and greater than the current reliability of the component or at least greater than zero).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should not be linear, but rather quantify the fact that it is incrementally harder to improve reliability.  For example, it is considerably easier to increase the reliability from 90% to 91% than to increase it from 99.99% to 99.999%, even though the increase is larger in the first case.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be asymptotic to the maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The following default cost function (also used in BlockSim) adheres to all of these conditions and acts like a penalty function for increasing a component&#039;s reliability.  Furthermore, an exponential behavior for the cost is assumed since it should get exponentially more difficult to increase the reliability. See Mettas [21]. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})&amp;lt;/math&amp;gt;  is the penalty (or cost) function as a function of component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  is the feasibility (or cost index) of improving a component&#039;s reliability relative to the other components in the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{min,i}}&amp;lt;/math&amp;gt;  is the current reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{max,i}}&amp;lt;/math&amp;gt;  is the maximum achievable reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
Note that this penalty function is dimensionless.  It essentially acts as a weighting factor that describes the difficulty in increasing the component reliability from its current value, relative to the other components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examining the cost function given by Eqn. (Default Cost), the following observations can be made:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability departs from the minimum or current value of reliability.  It is assumed that the reliabilities for the components will not take values any lower than they already have.  Depending on the optimization, a component&#039;s reliability may not need to be increased from its current value but it will not drop any lower.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability approaches the maximum achievable reliability.  This is a reliability value that is approached asymptotically as the cost increases but is never actually reached.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost is a function of the range of improvement, which is the difference between the component&#039;s initial reliability and the corresponding maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The exponent in Eqn. (Default Cost) approaches infinity as the component&#039;s reliability approaches its maximum achievable value.  This means that it is easier to increase the reliability of a component from a lower initial value.  For example, it is easier to increase a component&#039;s reliability from 70% to 75% than increasing its reliability from 90% to 95%.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====The Feasibility Term,  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
The feasibility term in Eqn. (Default Cost) is a constant (or an equation parameter) that represents the difficulty in increasing a component&#039;s reliability relative to the rest of the components in the system.  Depending on the design complexity, technological limitations, etc., certain components can be very hard to improve.  Clearly, the more difficult it is to improve the reliability of the component, the greater the cost.  Figure feasplot illustrates the behavior of the function defined in Eqn. (Default Cost) for different values of  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; .  It can be seen that the lower the feasibility value, the more rapidly the cost function approaches infinity.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Several methods can be used to obtain a feasibility value.  Weighting factors for allocating reliability have been proposed by many authors and can be used to quantify feasibility.  These weights depend on certain factors of influence, such as the complexity of the component, the state of the art, the operational profile, the criticality, etc.  Engineering judgment based on past experience, supplier quality, supplier availability and other factors can also be used in determining a feasibility value.  Overall, the assignment of a feasibility value is going to be a subjective process.  Of course, this problem is negated if the relationship between the cost and the reliability for each component is known because one can use regression methods to estimate the parameter value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.11.gif|thumb|center|400px|Behavior of the cost function for different feasibility values.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Maximum Achievable Reliability====&lt;br /&gt;
&lt;br /&gt;
For the purposes of reliability optimization, we also need to define a limiting reliability that a component will approach, but not reach.  The costs near the maximum achievable reliability are very high and the actual value for the maximum reliability is usually dictated by technological or financial constraints.  In deciding on a value to use for the maximum achievable reliability, the current state of the art of the component in question and other similar factors will have to be considered.  In the end, a realistic estimation based on engineering judgment and experience will be necessary to assign a value to this input.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that the time associated with this maximum achievable reliability is the same as that of the overall system reliability goal.  Almost any component can achieve a very high reliability value, provided the mission time is short enough.  For example, a component with an exponential distribution and a failure rate of one failure per hour has a reliability that drops below 1% for missions greater than five hours.  However, it can achieve a reliability of 99.9% as long as the mission is no longer than four seconds.  For the purposes of optimization in BlockSim, the reliability values of the components are associated with the time for which the system reliability goal is specified.  For example, if the problem is to achieve a system goal of 99% reliability at 1,000 hours, the maximum achievable reliability values entered for the individual components would be the maximum reliability that each component could attain for a mission of 1,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As the component reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , approaches the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function approaches infinity.  The maximum achievable reliability acts as a scale parameter for the cost function.  By decreasing  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function is compressed between  &amp;lt;math&amp;gt;{{R}_{i,min}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , as shown in Figure oldfig5.&lt;br /&gt;
 &lt;br /&gt;
====Cost Function====&lt;br /&gt;
Once the cost functions for the individual components have been determined, it becomes necessary to develop an expression for the overall system cost.  This takes the form of:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{s}}({{R}_{G}})={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+...+{{C}_{n}}({{R}_{n}}),i=1,2,...,n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, the cost of the system is simply the sum of the costs of its components.  This is regardless of the form of the individual component cost functions.  They can be of the general behavior model in BlockSim or they can be user-defined.   Once the overall cost function for the system has been defined, the problem becomes one of minimizing the cost function while remaining within the constraints defined by the target system reliability and the reliability ranges for the components.  The latter constraints in this case are defined by the minimum and maximum reliability values for the individual components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.12.png|thumb|center|400px|Effect5 of the maximum achievable reliability on the cost function.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim employs a nonlinear programming technique to minimize the system cost function.  The system has a minimum (current) and theoretical maximum reliability value that is defined by the minimum and maximum reliabilities of the components and by the way the system is configured.  That is, the structural properties of the system are accounted for in the determination of the optimum solution.  For example, the optimization for a system of three units in series will be different from the optimization for a system consisting of the same three units in parallel.  The optimization occurs by varying the reliability values of the components within their respective constraints of maximum and minimum reliability in a way that the overall system goal is achieved.  Obviously, there can be any number of different combinations of component reliability values that might achieve the system goal.  The optimization routine essentially finds the combination that results in the lowest overall system cost. &lt;br /&gt;
&lt;br /&gt;
====Determining the Optimum Allocation Scheme====&lt;br /&gt;
&lt;br /&gt;
To determine the optimum reliability allocation, the analyst first determines the system reliability equation (the objective function).  As an example, and again for a trivial system with three components in series, this would be:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a target reliability of 90% is sought, then Eqn. (optAlloc) is recast as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The objective now is to solve for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  so that the equality in Eqn. (optAlloc90) is satisfied.  To obtain an optimum solution, we also need to use our cost functions (i.e. define the total allocation costs) as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
With the cost equation defined, then the optimum values for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  are the values that satisfy the reliability requirement, Eqn. (optAlloc90), at the minimum cost, Eqn. (optcost).  BlockSim uses this methodology during the optimization task.&lt;br /&gt;
&lt;br /&gt;
====Defining a Feasibility Policy in BlockSim====&lt;br /&gt;
&lt;br /&gt;
In BlockSim you can choose to use the default feasibility function, as defined by Eqn. (Default Cost), or use your own function.  Figure BSfvalues illustrates the use of the default values using the slider control. Figure BSFcustom shows the use of an associated feasibility policy to create a user-defined cost function.  When defining your own cost function, you should be aware of/adhere to the following guidelines:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Because the cost functions are evaluated relative to each other, they should be correlated.  In other words, if one function evaluates to 10,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=10&amp;lt;/math&amp;gt;  for one block and 20 for another,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=20&amp;lt;/math&amp;gt; , then the implication is that there is a 1 to 2 cost relation.  &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Do not mix your own function with the software&#039;s default functions unless you have verified that your cost functions are defined and correlated to the default cost functions, as defined by Eqn. (Default Cost).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Your function should adhere to the guidelines presented earlier.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Lastly, and since the evaluation is relative, it is preferable to use the pre-defined functions unless you have a compelling reason (or data) to do otherwise.  The last section in this chapter describes cases where user-defined functions are preferred.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.13.png|thumb|center|300px|Setting the default feasibility function in BlockSim with the feasibility slider. Note that the feasibility slider displays values, &#039;&#039;SV&#039;&#039;, from 1 to 9 when moved by the user, with SV=9 being the hardest. The relationship between &#039;&#039;f&#039;&#039; and &#039;&#039;SV&#039;&#039; is ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.14.png|thumb|center|400px|Setting a user-defined feasibility function in BlockSim utilizing an assiciated feasibility policy. Any user-defined equation can be entered as a function of &#039;&#039;R.&#039;&#039;]]&lt;br /&gt;
&lt;br /&gt;
=Implementing the Optimization=&lt;br /&gt;
&lt;br /&gt;
As was mentioned earlier, there are two different methods of implementing the changes suggested by the reliability optimization routine: fault tolerance and fault avoidance.  When the optimized component reliabilities have been determined, it does not matter which of the two methods is employed to realize the optimum reliability for the component in question.  For example, suppose we have determined that a component must have its reliability for a certain mission time raised from 50% to 75%.  The engineer must now decide how to go about implementing the increase in reliability.  If the engineer decides to do this via fault avoidance, another component must be found (or the existing component must be redesigned) so that it will perform the same function with a higher reliability.  On the other hand, if the engineer decides to go the fault tolerance route, the optimized reliability can be achieved merely by placing a second identical component in parallel with the first one.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, the method of implementing the reliability optimization is going to be related to the cost function and this is something the reliability engineer must take into account when deciding on what type of cost function is used for the optimization.  In fact, if we take a closer look at the fault tolerance scheme, we can see some parallels with the general behavior cost model included in BlockSim.  For example, consider a system that consists of a single unit.  The cost of that unit, including all associated mounting and hardware costs, is one dollar.  The reliability of this unit for a given mission time is 30%.  It has been determined that this is inadequate and that a second component is to be added in parallel to increase the reliability.  Thus, the reliability for the two-unit parallel system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-{{(1-0.3)}^{2}}=0.51\text{ or }51%&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
So, the reliability has increased by a value of 21% and the cost has increased by one dollar.  In a similar fashion, we can continue to add more units in parallel, thus increasing the reliability and the cost.  We now have an array of reliability values and the associated costs that we can use to develop a cost function for this fault tolerance scheme.  Figure costredundant shows the relationship between cost and reliability for this example.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As can be seen, this looks quite similar to the general behavior cost model presented earlier.  In fact, a standard regression analysis available in Weibull++ indicates that an exponential model fits this cost model quite well.   The function is given by the following equation, where  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is the cost in dollars and  &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt;  is the fractional reliability value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C(R)=0.3756\cdot {{e}^{3.1972\cdot R}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.15.gif|thumb|center|400px|Cost function for redundant parallel units.]]&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
&lt;br /&gt;
Consider a system consisting of three components connected reliability-wise in series.  Assume the objective reliability for the system is 90% for a mission time of 100 hours.  Five cases will be considered for the allocation problem. See Mettas [21].&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1 - All three components are identical with times-to-failure that are described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.318 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 312 hours. All three components have the same feasibility value of Moderate (5).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2 - Same as in Case 1, but Component 1 has a feasibility of Easy, Component 2 has a feasibility of Moderate and Component 3 has a feasibility of Hard.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3 - Component 1 has 70% reliability, Component 2 has 80% reliability and Component 3 has 90% reliability, all for a mission duration of 100 hours.  All three components have the same feasibility of Easy.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 4 - Component 1 has 70% reliability and Easy feasibility, Component 2 has 80% reliability and Moderate feasibility, and Component 3 has 90% reliability and Hard feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 5 - Component 1 has 70% reliability and Hard feasibility, Component 2 has 80% reliability and Easy feasibility and Component 3 has 90% reliability and Moderate feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In all cases, the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , for each component is 99.9% for a mission duration of 100 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.16.gif|thumb|center|300px|Optimization inputs in BlockSim&#039;s Analytical QCP for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Solution====&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Case 1&#039;&#039;&#039; - The reliability equation for Case 1 is: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the equality desired is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}(t=100)\cdot {{R}_{2}}(t=100)\cdot {{R}_{3}}(t=100)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{1,2,3}}={{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cost or feasibility function is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{1,2,3}}({{R}_{1,2,3}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
And where  &amp;lt;math&amp;gt;{{R}_{\max _{1,2,3}^{}}}=0.999&amp;lt;/math&amp;gt;  (arbitrarily set),  &amp;lt;math&amp;gt;{{R}_{\min _{1,2,3}^{}}}&amp;lt;/math&amp;gt;  computed from the reliability function of each component at the time of interest,  &amp;lt;math&amp;gt;t=100&amp;lt;/math&amp;gt; , or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{\min _{1,2,3}^{}}}= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( \tfrac{100}{312} \right)}^{1.318}}}} \\ &lt;br /&gt;
= &amp;amp; 0.79995  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  obtained from: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
f= &amp;amp; \left( 1-\frac{5}{10} \right) \\ &lt;br /&gt;
= &amp;amp; 0.5  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The solution,  &amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}&amp;lt;/math&amp;gt; , is the one that satisfies Eqn. (exbjective2) while minimizing Eqn. (exonstraint).  In this case (and since all the components are identical), the target reliability is found to be: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}(t=100)=0.9655&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Figures QCPOpt and OptResults show related BlockSim screens.  Based on this, each component&#039;s reliability should be at least 96.55% at 100 hours in order for the system&#039;s reliability to be 90% at 100 hours.  Note the column labeled N.E.P.U. in the Results Panel shown in Figure OptResults.  This stands for &amp;quot;Number of Equivalent Parallel Units&amp;quot; and represents the number of redundant units that would be required to bring that particular component up to the recommended reliability.  In the case where the fault tolerance approach is to be implemented, the N.E.P.U value should be rounded up to an integer.  Therefore, some manipulation by the engineer is required in order to ensure that the chosen integer values will yield the required system reliability goal (or exceed it).  In addition, further cost analysis should be performed in order to account for the costs of adding redundancy to the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Additionally, and when the results have been obtained, the engineer may wish to re-scale the components based on their distribution parameters instead of the fixed reliability value.  In the case of these components, one may wish to re-scale the scale parameter of the distribution ,  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt; , for the components, or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{100}{\eta } \right)}^{1.318}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.17.png|thumb|center|300px|Optimization results for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Which yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\eta }_{{{O}_{i}}}}=1269.48&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The Parameter Experimenter in BlockSim can also be used for this (Figure paramexper).  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The results from the other cases can be obtained in a similar fashion.  The results for Cases 1 through 5 are summarized next.&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{matrix}&lt;br /&gt;
   {} &amp;amp; Case 1 &amp;amp; Case 2 &amp;amp; Case 3 &amp;amp; Case 4 &amp;amp; Case 5  \\&lt;br /&gt;
   Component 1 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9874} &amp;amp; \text{0}\text{.9552} &amp;amp; \text{0}\text{.9790} &amp;amp; \text{0}\text{.9295}  \\&lt;br /&gt;
   Component 2 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9633} &amp;amp; \text{0}\text{.9649} &amp;amp; \text{0}\text{.9553} &amp;amp; \text{0}\text{.9884}  \\&lt;br /&gt;
   Component 3 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9463} &amp;amp; \text{0}\text{.9765} &amp;amp; \text{0}\text{.9624} &amp;amp; \text{0}\text{.9797}  \\&lt;br /&gt;
\end{matrix}&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 2&#039;&#039;&#039; - It can be seen that the highest reliability was allocated to Component 1 with the Easy feasibility.  The lowest reliability was assigned to Component 3 with the Hard feasibility.  This makes sense in that an optimized reliability scheme will call for the greatest reliability changes in those components that are the easiest to change.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 3&#039;&#039;&#039; - The components were different but had the same feasibility values.  &lt;br /&gt;
&lt;br /&gt;
[[Image:BS6.18.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, all three components have the same opportunity for improvement.  This case differs from Cases 1 and 2 since there are two factors, not present previously, that will affect the outcome of the allocation in this case.  First, each component in this case has a different reliability importance (impact of a component on the system&#039;s reliability); whereas in Cases 1 and 2, all three components were identical and had the same reliability importance.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure relimp shows the reliability importance for each component, where it can be seen that Component 1 has the greatest reliability importance and Component 3 has the smallest (this reliability importance also applies in Cases 4 and 5).  This indicates that the reliability of Component 1 should be significantly increased because it has the biggest impact on the overall system reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In addition, each component&#039;s cost function in Case 3 also depends on the difference between each component&#039;s initial reliability and its corresponding maximum achievable reliability.  (In Cases 1 and 2 this was not an issue because the components were identical.)  The greater this difference, the greater the cost of improving the reliability of a particular component relative to the other two components.  This difference between the initial reliability of a component and its maximum achievable reliability is called the range of improvement for that component.  Since all three components have the same maximum achievable reliability, Component 1, with the largest range for improvement, is the most cost efficient component to improve.  The improvement ranges for all three components are illustrated in Figure Rangeofimprovement. At the same time, however, there is a reliability value between the initial and the maximum achievable reliability beyond which it becomes cost prohibitive to improve any further.  This reliability value is dictated by the feasibility value.  From the table of results, it can be seen that in Case 3 there was a 25.52% improvement for Component 1, 16.49% for Component 2 and 7.65% for Component 3.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.19.png|thumb|center|300px|Reliability importance for Example 2, Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.20.png|thumb|center|300px|Range of improvement for each component for Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 4&#039;&#039;&#039; - As opposed to Case 3, Component 1 was assigned an even greater increase of 27.9%, with Components 2 and 3 receiving lesser increases than in Case 3, of 15.53% and 6.24% respectively.  This is due to the fact that Component 1 has an Easy feasibility and Component 3 has a Hard feasibility, which means that it is more difficult to increase the reliability of Component 3 than to increase the reliability of Component 1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 5&#039;&#039;&#039; - The feasibility values here are reversed with Component 1 having a Hard feasibility and Component 3 an Easy feasibility.  The recommended increase in Component 1&#039;s reliability is less compared to its increase for Cases 3 and 4.  Note, however, that Components 2 and 3 still received a smaller increase in reliability than Component 1 because their ranges of improvement are smaller.  In other words, Component 3 was assigned the smallest increase in reliability in Cases 3, 4 and 5 because its initial reliability is very close to its maximum achievable reliability. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Setting Specifications=&lt;br /&gt;
&lt;br /&gt;
This methodology could also be used to arrive at initial specifications for a set of components.  In the prior examples, we assumed a current reliability for the components.  One could repeat these steps by choosing an arbitrary (lower) initial reliability for each component, thus allowing the algorithm to travel up to the target.  When doing this, it is important to keep in mind the fact that both the distance from the target (the distance from the initial arbitrary value and the target value) for each component is also a significant contributor to the final results, as presented in the prior example.  If one wishes to arrive at the results using only the cost functions then it may be advantageous to set equal initial reliabilities for all components.&lt;br /&gt;
&lt;br /&gt;
=Other Notes on User-Defined Cost Functions=&lt;br /&gt;
&lt;br /&gt;
The optimization method in BlockSim is a very powerful tool for allocating reliability to the components of a system while minimizing an overall cost of improvement.  The default cost function in BlockSim was derived in order to model a general relationship between the cost and the component reliability.  However, if actual cost information is available, then one can use the cost data instead of using the default function.  Additionally, one can also view the feasibility in the default function as a measure of the difficulty in increasing the reliability of the component relative to the rest of the components to be optimized, assuming that they also follow the same cost function with the corresponding feasibility values.  If fault tolerance is a viable option, a reliability cost function for adding parallel units can be developed as demonstrated previously.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Another method for developing a reliability cost function would be to obtain different samples of components from different suppliers and test the samples to determine the reliability of each sample type.  From this data, a curve could be fitted through standard regression techniques and an equation defining the cost as a function of reliability could be developed.  Figure RGplot shows such a curve.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, and in cases where a reliability growth program is in place, the simplest way of obtaining a relationship between cost and reliability is by associating a cost to each development stage of the growth process.  Reliability growth models such as the Crow (AMSAA), Duane, Gompertz and Logistic models can be used to describe the cost as a function of reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.21.png|thumb|center|300px|Typical reliability growth curve generated using ReliaSoft&#039;s Reliability Growth software.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a reliability growth model has been successfully implemented, the development costs over the respective development time stages can be applied to the growth model, resulting in equations that describe reliability/cost relationships.  These equations can then be entered into BlockSim as user-defined cost functions (feasibility policies).  The only potential drawback to using growth model data is the lack of flexibility in applying the optimum results.  Making the cost projection for future stages of the project would require the assumption that development costs will be accrued at a similar rate in the future, which may not always be a valid assumption.  Also, if the optimization result suggests using a high reliability value for a component, it may take more time than is allotted for that project to attain the required reliability given the current reliability growth of the project.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15638</id>
		<title>Reliability Importance and Optimized Reliability Allocation (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15638"/>
		<updated>2012-02-13T23:35:26Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Static Reliability Importance */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|6}}&lt;br /&gt;
&lt;br /&gt;
=Component Reliability Importance=&lt;br /&gt;
===Static Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Once the reliability of a system has been determined, engineers are often faced with the task of identifying the least reliable component(s) in the system in order to improve the design.  For example, it was observed in Chapter 4 that the least reliable component in a series system has the biggest effect on the system reliability.  In this case, if the reliability of the system is to be improved, then the efforts can best be concentrated on improving the reliability of that component first.   In simple systems such as a series system, it is easy to identify the weak components.  However, in more complex systems this becomes quite a difficult task.  For complex systems, the analyst needs a mathematical approach that will provide the means of identifying and quantifying the importance of each component in the system.&lt;br /&gt;
&lt;br /&gt;
Using reliability importance measures is one method of identifying the relative importance of each component in a system with respect to the overall reliability of the system.  The reliability importance,  &amp;lt;math&amp;gt;{{I}_{R}}&amp;lt;/math&amp;gt; , of component  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;  in a system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  components is given by [[Appendix D: Weibull References | Leemis [17]]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{i}}}   \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{s}}&amp;lt;/math&amp;gt;  is the system reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  is the component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The value of the reliability importance given by Eqn.6.1 depends both on the reliability of a component and its corresponding position in the system.  In Chapter 4 we observed that for a simple series system (three components in series with reliabilities of 0.7, 0.8 and 0.9) the rate of increase of the system reliability was greatest when the least reliable component was improved.  In other words, it was observed that Component 1 had the largest reliability importance in the system relative to the other two components (see Figure Ch6fig1).  The same conclusion can be drawn by using Eqn.6.1 and obtaining the reliability importance in terms of a value for each component.&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Using BlockSim, the reliability importance values for these components can be calculated with Eqn.6.1.  Using the plot option and selecting the Static Reliability Importance plot type, Figure 6.2 can be obtained.  Note that the time input required to create this plot is irrelevant for this example because the components are static.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The values shown in Figure fig1a for each component were obtained using Eqn.6.1.  The reliability equation for this series system was given by: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Taking the partial derivative of Eqn.6.2 with respect to  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{I}_{{{R}_{1}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{1}}}= &amp;amp; {{R}_{2}}{{R}_{3}} \\ &lt;br /&gt;
= &amp;amp; 0.8\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.72  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus the reliability importance of Component 1 is  &amp;lt;math&amp;gt;{{I}_{{{R}_{1}}}}=&amp;lt;/math&amp;gt;  0.72.  The reliability importance values for Components 2 and 3 are obtained in a similar manner.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.1.png|thumb|center|300px|Rate of change of system reliability when increasing the reliability of each component.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:6.2.gif|thumb|center|300px|Static reliability importance plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Time-Dependent Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
\The same concept applies if the components have a time-varying reliability.  That is, if  &amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt; , then one could compute  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  at any time  &amp;lt;math&amp;gt;x,&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}{{(t)}_{_{t=x}}}.&amp;lt;/math&amp;gt;   This is quantified in Eqn. (importance time). &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)=\frac{\partial {{R}_{s}}(t)}{\partial {{R}_{i}}(t)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In turn, this can be viewed as either a static plot (at a given time) or as time-varying plot, as illustrated in the next figures.  Specifically, Figures Ch6fig3, Ch6fig4 and Ch6fig5 present the analysis for three components configured reliability-wise in series following a Weibull distribution with  &amp;lt;math&amp;gt;\beta =3&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=1,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\eta }_{2}}=2,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=3,000&amp;lt;/math&amp;gt; .  Figure Ch6fig3 shows a bar chart of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  while Figure Ch6fig4 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  in BlockSim&#039;s tableau chart format.  In this chart, the area of the square is  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt; .  Lastly, Figure Ch6fig5 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  vs. time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that a system has failure modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; .  Furthermore, assume that failure of the entire system will occur if:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  occurs.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  occur.&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, assume the following failure probabilities for each mode.&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  have a mean time to occurrence of 1,000 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=1,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  has a mean time to occurrence of 100 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=100).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a mean time to occurrence of 700,000, 1,000,000 and 2,000,000 hours respectively (i.e. exponential with  &amp;lt;math&amp;gt;MTT{{F}_{B}}=700,000&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;MTT{{F}_{C}}=1,000,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;MTT{{F}_{F}}=2,000,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examine the mode importance for operating times of 100 and 500 hours.&lt;br /&gt;
&lt;br /&gt;
[[Image:6.3.gif|thumb|center|400px|Static Reliability Importance plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.4.png|thumb|center|400px|Static Reliability Importance tableau plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.5.png|thumb|center|400px|Reliability Importance vs. time plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
The RBD for this example is (from Chapter 4, Example 18):&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6ex1.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig6 illustrates  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt; .  It can be seen that even though  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a much rarer rate of occurrence, they are much more significant at 100 hours.  By 500 hours,  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt; , the effects of the lower reliability components become greatly pronounced and thus they become more important, as can be seen in Figure Ch6fig7.  Finally, the behavior of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  can be observed in Figure Ch6fig8.  Note that not all lines are plainly visible in Figure Ch6fig8 due to overlap.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Reliability Allocation=&lt;br /&gt;
&lt;br /&gt;
In the process of  developing a new product, the engineer is often faced with the task of designing a system that conforms to a set of reliability specifications.  The engineer is given the goal for the system and must then develop a design that will achieve the desired reliability of the system, while performing all of the system&#039;s intended functions at a minimum cost. This involves a balancing act of determining how to allocate reliability to the components in the system so the system will meet its reliability goal while at the same time ensuring that the system meets all of the other associated performance specifications.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.6.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.7.gif|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.8.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
The simplest method for allocating reliability is to distribute the reliabilities uniformly among all components. For example, suppose a system with five components in series has a reliability objective of 90% for a given operating time. The uniform allocation of the objective to all components would require each component to have a reliability of 98% for the specified operating time, since  &amp;lt;math&amp;gt;{{0.98}^{5}}\tilde{=}0.90&amp;lt;/math&amp;gt;. While this manner of allocation is easy to calculate, it is generally not the best way to allocate reliability for a system. The optimum method of allocating reliability would take into account the cost or relative difficulty of improving the reliability of different subsystems or components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability optimization process begins with the development of a model that represents the entire system.  This is accomplished with the construction of a system reliability block diagram that represents the reliability relationships of the components in the system.  From this model, the system reliability impact of different component modifications can be estimated and considered alongside the costs that would be incurred in the process of making those modifications.  It is then possible to perform an optimization analysis for this problem, finding the best combination of component reliability improvements that meet or exceed the performance goals at the lowest cost.&lt;br /&gt;
&lt;br /&gt;
===Importance Measures and FMEA/FMECA===&lt;br /&gt;
&lt;br /&gt;
Traditional Failure Mode and Effects analysis (FMEA/FMECA) relies on Risk Priority Numbers (RPNs) or criticality calculations to identify and prioritize the significance/importance of different failure modes.  The RPN methodology (and to some extent, the criticality methodology) tend to be subjective.  When conducting these types of analyses, one may wish to incorporate more quantitative metrics, such as the importance measure presented here and/or the RS FCI and RS DECI for repairable systems (which are discussed in later chapters).  ReliaSoft&#039;s Xfmea software can be used to export an FMEA/FMECA analysis to BlockSim.  The documentation that accompanies Xfmea provides more information on FMEA/FMECA, including both methods of risk assessment.&lt;br /&gt;
&lt;br /&gt;
=Improving Reliability=&lt;br /&gt;
Reliability engineers are very often called upon to make decisions as to whether to improve a certain component or components in order to achieve a minimum required system reliability.  There are two approaches to improving the reliability of a system: fault avoidance and fault tolerance.  Fault avoidance is achieved by using high-quality and high-reliability components and is usually less expensive than fault tolerance.  Fault tolerance, on the other hand, is achieved by redundancy.  Redundancy can result in increased design complexity and increased costs through additional weight, space, etc.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before deciding whether to improve the reliability of a system by fault tolerance or fault avoidance, a reliability assessment for each component in the system should be made.  Once the reliability values for the components have been quantified, an analysis can be performed in order to determine if that system&#039;s reliability goal will be met.  If it becomes apparent that the system&#039;s reliability will not be adequate to meet the desired goal at the specified mission duration, steps can be taken to determine the best way to improve the system&#039;s reliability so that it will reach the desired target.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a system with three components connected reliability-wise in series.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 70%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 80% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 90%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 85%, is required for this system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}=50.4%&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, this is far short of the system&#039;s required reliability performance.  It is apparent that the reliability of the system&#039;s constituent components will need to be increased in order for the system to meet its goal.  First, we will try increasing the reliability of one component at a time to see whether the reliability goal can be achieved.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig9 shows that even by raising the individual component reliability to a hypothetical value of 1 (100% reliability, which implies that the component will never fail), the overall system reliability goal will not be met by improving the reliability of just one component.  The next logical step would be to try to increase the reliability of two components.  The question now becomes: which two?  One might also suggest increasing the reliability of all three components.  A basis for making such decisions needs to be found in order to avoid the ``trial and error&#039;&#039; aspect of altering the system&#039;s components randomly in an attempt to achieve the system reliability goal.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.9.gif|thumb|center|400px|Change in system reliability of a three-unit series system due to increasing the reliability of just one component.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As we have seen, the reliability goal for the preceding example could not be achieved by increasing the reliability of just one component.  There are cases, however, where increasing the reliability of one component results in achieving the system reliability goal.  Consider, for example, a system with three components connected reliability-wise in parallel.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 60%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 70% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 80%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 99%, is required for this system.  The initial system reliability is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-(1-0.6)\cdot (1-0.7)\cdot (1-0.8)=0.976&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current system reliability is inadequate to meet the goal.  Once again, we can try to meet the system reliability goal by raising the reliability of just one of the three components in the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From Figure fig10, it can be seen that the reliability goal can be reached by improving Component 1, Component 2 or Component 3.  The reliability engineer is now faced with another dilemma:  which component&#039;s reliability should be improved? This presents a new aspect to the problem of allocating the reliability of the system.  Since we know that the system reliability goal can be achieved by increasing at least one unit, the question becomes one of how to do this most efficiently and cost effectively.  We will need more information to make an informed decision as to how to go about improving the system&#039;s reliability.  How much does each component need to be improved for the system to meet its goal?  How feasible is it to improve the reliability of each component?  Would it actually be more efficient to slightly raise the reliability of two or three components rather than radically improving only one?&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to answer these questions, we must introduce another variable into the problem &amp;lt;math&amp;gt;:\ \ \ &amp;lt;/math&amp;gt; cost.  Cost does not necessarily have to be in dollars.  It could be described in terms of non-monetary resources, such as time.  By associating cost values to the reliabilities of the system&#039;s components, we can find an optimum design that will provide the required reliability at a minimum cost.&lt;br /&gt;
&lt;br /&gt;
===Cost/Penalty Function===&lt;br /&gt;
&lt;br /&gt;
There is always a cost associated with changing a design due to change of vendors, use of higher-quality materials, retooling costs, administrative fees, etc.  The cost as a function of the reliability for each component must be quantified before attempting to improve the reliability.  Otherwise, the design changes may result in a system that is needlessly expensive or overdesigned.  Developing the ``cost of reliability&#039;&#039; relationship will give the engineer an understanding of which components to improve and how to best concentrate the effort and allocate resources in doing so.  The first step will be to obtain a relationship between the cost of improvement and reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.10.png|thumb|center|400px|Meeting a reliability goal requirement by increasing a component&#039;s reliability]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The preferred approach would be to formulate the cost function from actual cost data.  This can be done from past experience.  If a reliability growth program is in place, the costs associated with each stage of improvement can also be quantified.  Defining the different costs associated with different vendors or different component models is also useful in formulating a model of component cost as a function of reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, there are many cases where no such information is available.  For this reason, a general (default) behavior model of the cost versus the component&#039;s reliability was developed for performing reliability optimization in BlockSim.  The objective of this function is to model an overall cost behavior for all types of components.  Of course, it is impossible to formulate a model that will be precisely applicable to every situation; but the proposed relationship is general enough to cover most applications.  In addition to the default model formulation, BlockSim does allow the definition of user-defined cost models.&lt;br /&gt;
&lt;br /&gt;
====Quantifying the Cost/Penalty Function====&lt;br /&gt;
&lt;br /&gt;
One needs to quantify a cost function for each component,  &amp;lt;math&amp;gt;{{C}_{i}}&amp;lt;/math&amp;gt; , in terms of the reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , of each component, or:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}=f({{R}_{i}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This function should:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the current reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Current}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the maximum possible reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Max}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Allow for different levels of difficulty (or cost) in increasing the reliability of each component.  It can take into account:&amp;lt;br&amp;gt;&lt;br /&gt;
::o	design issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	supplier issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	state of technology.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	time-to-market issues, etc.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus, for the cost function to comply with these needs, the following conditions should be adhered to:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be constrained by the minimum and maximum reliabilities of each component (i.e. reliability must be less than one and greater than the current reliability of the component or at least greater than zero).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should not be linear, but rather quantify the fact that it is incrementally harder to improve reliability.  For example, it is considerably easier to increase the reliability from 90% to 91% than to increase it from 99.99% to 99.999%, even though the increase is larger in the first case.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be asymptotic to the maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The following default cost function (also used in BlockSim) adheres to all of these conditions and acts like a penalty function for increasing a component&#039;s reliability.  Furthermore, an exponential behavior for the cost is assumed since it should get exponentially more difficult to increase the reliability. See Mettas [21]. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})&amp;lt;/math&amp;gt;  is the penalty (or cost) function as a function of component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  is the feasibility (or cost index) of improving a component&#039;s reliability relative to the other components in the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{min,i}}&amp;lt;/math&amp;gt;  is the current reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{max,i}}&amp;lt;/math&amp;gt;  is the maximum achievable reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
Note that this penalty function is dimensionless.  It essentially acts as a weighting factor that describes the difficulty in increasing the component reliability from its current value, relative to the other components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examining the cost function given by Eqn. (Default Cost), the following observations can be made:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability departs from the minimum or current value of reliability.  It is assumed that the reliabilities for the components will not take values any lower than they already have.  Depending on the optimization, a component&#039;s reliability may not need to be increased from its current value but it will not drop any lower.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability approaches the maximum achievable reliability.  This is a reliability value that is approached asymptotically as the cost increases but is never actually reached.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost is a function of the range of improvement, which is the difference between the component&#039;s initial reliability and the corresponding maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The exponent in Eqn. (Default Cost) approaches infinity as the component&#039;s reliability approaches its maximum achievable value.  This means that it is easier to increase the reliability of a component from a lower initial value.  For example, it is easier to increase a component&#039;s reliability from 70% to 75% than increasing its reliability from 90% to 95%.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====The Feasibility Term,  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
The feasibility term in Eqn. (Default Cost) is a constant (or an equation parameter) that represents the difficulty in increasing a component&#039;s reliability relative to the rest of the components in the system.  Depending on the design complexity, technological limitations, etc., certain components can be very hard to improve.  Clearly, the more difficult it is to improve the reliability of the component, the greater the cost.  Figure feasplot illustrates the behavior of the function defined in Eqn. (Default Cost) for different values of  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; .  It can be seen that the lower the feasibility value, the more rapidly the cost function approaches infinity.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Several methods can be used to obtain a feasibility value.  Weighting factors for allocating reliability have been proposed by many authors and can be used to quantify feasibility.  These weights depend on certain factors of influence, such as the complexity of the component, the state of the art, the operational profile, the criticality, etc.  Engineering judgment based on past experience, supplier quality, supplier availability and other factors can also be used in determining a feasibility value.  Overall, the assignment of a feasibility value is going to be a subjective process.  Of course, this problem is negated if the relationship between the cost and the reliability for each component is known because one can use regression methods to estimate the parameter value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.11.gif|thumb|center|400px|Behavior of the cost function for different feasibility values.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Maximum Achievable Reliability====&lt;br /&gt;
&lt;br /&gt;
For the purposes of reliability optimization, we also need to define a limiting reliability that a component will approach, but not reach.  The costs near the maximum achievable reliability are very high and the actual value for the maximum reliability is usually dictated by technological or financial constraints.  In deciding on a value to use for the maximum achievable reliability, the current state of the art of the component in question and other similar factors will have to be considered.  In the end, a realistic estimation based on engineering judgment and experience will be necessary to assign a value to this input.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that the time associated with this maximum achievable reliability is the same as that of the overall system reliability goal.  Almost any component can achieve a very high reliability value, provided the mission time is short enough.  For example, a component with an exponential distribution and a failure rate of one failure per hour has a reliability that drops below 1% for missions greater than five hours.  However, it can achieve a reliability of 99.9% as long as the mission is no longer than four seconds.  For the purposes of optimization in BlockSim, the reliability values of the components are associated with the time for which the system reliability goal is specified.  For example, if the problem is to achieve a system goal of 99% reliability at 1,000 hours, the maximum achievable reliability values entered for the individual components would be the maximum reliability that each component could attain for a mission of 1,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As the component reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , approaches the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function approaches infinity.  The maximum achievable reliability acts as a scale parameter for the cost function.  By decreasing  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function is compressed between  &amp;lt;math&amp;gt;{{R}_{i,min}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , as shown in Figure oldfig5.&lt;br /&gt;
 &lt;br /&gt;
====Cost Function====&lt;br /&gt;
Once the cost functions for the individual components have been determined, it becomes necessary to develop an expression for the overall system cost.  This takes the form of:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{s}}({{R}_{G}})={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+...+{{C}_{n}}({{R}_{n}}),i=1,2,...,n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, the cost of the system is simply the sum of the costs of its components.  This is regardless of the form of the individual component cost functions.  They can be of the general behavior model in BlockSim or they can be user-defined.   Once the overall cost function for the system has been defined, the problem becomes one of minimizing the cost function while remaining within the constraints defined by the target system reliability and the reliability ranges for the components.  The latter constraints in this case are defined by the minimum and maximum reliability values for the individual components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.12.png|thumb|center|400px|Effect5 of the maximum achievable reliability on the cost function.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim employs a nonlinear programming technique to minimize the system cost function.  The system has a minimum (current) and theoretical maximum reliability value that is defined by the minimum and maximum reliabilities of the components and by the way the system is configured.  That is, the structural properties of the system are accounted for in the determination of the optimum solution.  For example, the optimization for a system of three units in series will be different from the optimization for a system consisting of the same three units in parallel.  The optimization occurs by varying the reliability values of the components within their respective constraints of maximum and minimum reliability in a way that the overall system goal is achieved.  Obviously, there can be any number of different combinations of component reliability values that might achieve the system goal.  The optimization routine essentially finds the combination that results in the lowest overall system cost. &lt;br /&gt;
&lt;br /&gt;
====Determining the Optimum Allocation Scheme====&lt;br /&gt;
&lt;br /&gt;
To determine the optimum reliability allocation, the analyst first determines the system reliability equation (the objective function).  As an example, and again for a trivial system with three components in series, this would be:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a target reliability of 90% is sought, then Eqn. (optAlloc) is recast as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The objective now is to solve for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  so that the equality in Eqn. (optAlloc90) is satisfied.  To obtain an optimum solution, we also need to use our cost functions (i.e. define the total allocation costs) as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
With the cost equation defined, then the optimum values for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  are the values that satisfy the reliability requirement, Eqn. (optAlloc90), at the minimum cost, Eqn. (optcost).  BlockSim uses this methodology during the optimization task.&lt;br /&gt;
&lt;br /&gt;
====Defining a Feasibility Policy in BlockSim====&lt;br /&gt;
&lt;br /&gt;
In BlockSim you can choose to use the default feasibility function, as defined by Eqn. (Default Cost), or use your own function.  Figure BSfvalues illustrates the use of the default values using the slider control. Figure BSFcustom shows the use of an associated feasibility policy to create a user-defined cost function.  When defining your own cost function, you should be aware of/adhere to the following guidelines:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Because the cost functions are evaluated relative to each other, they should be correlated.  In other words, if one function evaluates to 10,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=10&amp;lt;/math&amp;gt;  for one block and 20 for another,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=20&amp;lt;/math&amp;gt; , then the implication is that there is a 1 to 2 cost relation.  &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Do not mix your own function with the software&#039;s default functions unless you have verified that your cost functions are defined and correlated to the default cost functions, as defined by Eqn. (Default Cost).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Your function should adhere to the guidelines presented earlier.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Lastly, and since the evaluation is relative, it is preferable to use the pre-defined functions unless you have a compelling reason (or data) to do otherwise.  The last section in this chapter describes cases where user-defined functions are preferred.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.13.png|thumb|center|300px|Setting the default feasibility function in BlockSim with the feasibility slider. Note that the feasibility slider displays values, &#039;&#039;SV&#039;&#039;, from 1 to 9 when moved by the user, with SV=9 being the hardest. The relationship between &#039;&#039;f&#039;&#039; and &#039;&#039;SV&#039;&#039; is ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.14.png|thumb|center|400px|Setting a user-defined feasibility function in BlockSim utilizing an assiciated feasibility policy. Any user-defined equation can be entered as a function of &#039;&#039;R.&#039;&#039;]]&lt;br /&gt;
&lt;br /&gt;
=Implementing the Optimization=&lt;br /&gt;
&lt;br /&gt;
As was mentioned earlier, there are two different methods of implementing the changes suggested by the reliability optimization routine: fault tolerance and fault avoidance.  When the optimized component reliabilities have been determined, it does not matter which of the two methods is employed to realize the optimum reliability for the component in question.  For example, suppose we have determined that a component must have its reliability for a certain mission time raised from 50% to 75%.  The engineer must now decide how to go about implementing the increase in reliability.  If the engineer decides to do this via fault avoidance, another component must be found (or the existing component must be redesigned) so that it will perform the same function with a higher reliability.  On the other hand, if the engineer decides to go the fault tolerance route, the optimized reliability can be achieved merely by placing a second identical component in parallel with the first one.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, the method of implementing the reliability optimization is going to be related to the cost function and this is something the reliability engineer must take into account when deciding on what type of cost function is used for the optimization.  In fact, if we take a closer look at the fault tolerance scheme, we can see some parallels with the general behavior cost model included in BlockSim.  For example, consider a system that consists of a single unit.  The cost of that unit, including all associated mounting and hardware costs, is one dollar.  The reliability of this unit for a given mission time is 30%.  It has been determined that this is inadequate and that a second component is to be added in parallel to increase the reliability.  Thus, the reliability for the two-unit parallel system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-{{(1-0.3)}^{2}}=0.51\text{ or }51%&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
So, the reliability has increased by a value of 21% and the cost has increased by one dollar.  In a similar fashion, we can continue to add more units in parallel, thus increasing the reliability and the cost.  We now have an array of reliability values and the associated costs that we can use to develop a cost function for this fault tolerance scheme.  Figure costredundant shows the relationship between cost and reliability for this example.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As can be seen, this looks quite similar to the general behavior cost model presented earlier.  In fact, a standard regression analysis available in Weibull++ indicates that an exponential model fits this cost model quite well.   The function is given by the following equation, where  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is the cost in dollars and  &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt;  is the fractional reliability value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C(R)=0.3756\cdot {{e}^{3.1972\cdot R}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.15.gif|thumb|center|400px|Cost function for redundant parallel units.]]&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
&lt;br /&gt;
Consider a system consisting of three components connected reliability-wise in series.  Assume the objective reliability for the system is 90% for a mission time of 100 hours.  Five cases will be considered for the allocation problem. See Mettas [21].&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1 - All three components are identical with times-to-failure that are described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.318 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 312 hours. All three components have the same feasibility value of Moderate (5).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2 - Same as in Case 1, but Component 1 has a feasibility of Easy, Component 2 has a feasibility of Moderate and Component 3 has a feasibility of Hard.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3 - Component 1 has 70% reliability, Component 2 has 80% reliability and Component 3 has 90% reliability, all for a mission duration of 100 hours.  All three components have the same feasibility of Easy.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 4 - Component 1 has 70% reliability and Easy feasibility, Component 2 has 80% reliability and Moderate feasibility, and Component 3 has 90% reliability and Hard feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 5 - Component 1 has 70% reliability and Hard feasibility, Component 2 has 80% reliability and Easy feasibility and Component 3 has 90% reliability and Moderate feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In all cases, the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , for each component is 99.9% for a mission duration of 100 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.16.gif|thumb|center|300px|Optimization inputs in BlockSim&#039;s Analytical QCP for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Solution====&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Case 1&#039;&#039;&#039; - The reliability equation for Case 1 is: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the equality desired is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}(t=100)\cdot {{R}_{2}}(t=100)\cdot {{R}_{3}}(t=100)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{1,2,3}}={{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cost or feasibility function is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{1,2,3}}({{R}_{1,2,3}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
And where  &amp;lt;math&amp;gt;{{R}_{\max _{1,2,3}^{}}}=0.999&amp;lt;/math&amp;gt;  (arbitrarily set),  &amp;lt;math&amp;gt;{{R}_{\min _{1,2,3}^{}}}&amp;lt;/math&amp;gt;  computed from the reliability function of each component at the time of interest,  &amp;lt;math&amp;gt;t=100&amp;lt;/math&amp;gt; , or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{\min _{1,2,3}^{}}}= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( \tfrac{100}{312} \right)}^{1.318}}}} \\ &lt;br /&gt;
= &amp;amp; 0.79995  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  obtained from: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
f= &amp;amp; \left( 1-\frac{5}{10} \right) \\ &lt;br /&gt;
= &amp;amp; 0.5  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The solution,  &amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}&amp;lt;/math&amp;gt; , is the one that satisfies Eqn. (exbjective2) while minimizing Eqn. (exonstraint).  In this case (and since all the components are identical), the target reliability is found to be: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}(t=100)=0.9655&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Figures QCPOpt and OptResults show related BlockSim screens.  Based on this, each component&#039;s reliability should be at least 96.55% at 100 hours in order for the system&#039;s reliability to be 90% at 100 hours.  Note the column labeled N.E.P.U. in the Results Panel shown in Figure OptResults.  This stands for &amp;quot;Number of Equivalent Parallel Units&amp;quot; and represents the number of redundant units that would be required to bring that particular component up to the recommended reliability.  In the case where the fault tolerance approach is to be implemented, the N.E.P.U value should be rounded up to an integer.  Therefore, some manipulation by the engineer is required in order to ensure that the chosen integer values will yield the required system reliability goal (or exceed it).  In addition, further cost analysis should be performed in order to account for the costs of adding redundancy to the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Additionally, and when the results have been obtained, the engineer may wish to re-scale the components based on their distribution parameters instead of the fixed reliability value.  In the case of these components, one may wish to re-scale the scale parameter of the distribution ,  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt; , for the components, or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{100}{\eta } \right)}^{1.318}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.17.png|thumb|center|300px|Optimization results for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Which yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\eta }_{{{O}_{i}}}}=1269.48&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The Parameter Experimenter in BlockSim can also be used for this (Figure paramexper).  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The results from the other cases can be obtained in a similar fashion.  The results for Cases 1 through 5 are summarized next.&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{matrix}&lt;br /&gt;
   {} &amp;amp; Case 1 &amp;amp; Case 2 &amp;amp; Case 3 &amp;amp; Case 4 &amp;amp; Case 5  \\&lt;br /&gt;
   Component 1 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9874} &amp;amp; \text{0}\text{.9552} &amp;amp; \text{0}\text{.9790} &amp;amp; \text{0}\text{.9295}  \\&lt;br /&gt;
   Component 2 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9633} &amp;amp; \text{0}\text{.9649} &amp;amp; \text{0}\text{.9553} &amp;amp; \text{0}\text{.9884}  \\&lt;br /&gt;
   Component 3 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9463} &amp;amp; \text{0}\text{.9765} &amp;amp; \text{0}\text{.9624} &amp;amp; \text{0}\text{.9797}  \\&lt;br /&gt;
\end{matrix}&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 2&#039;&#039;&#039; - It can be seen that the highest reliability was allocated to Component 1 with the Easy feasibility.  The lowest reliability was assigned to Component 3 with the Hard feasibility.  This makes sense in that an optimized reliability scheme will call for the greatest reliability changes in those components that are the easiest to change.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 3&#039;&#039;&#039; - The components were different but had the same feasibility values.  &lt;br /&gt;
&lt;br /&gt;
[[Image:BS6.18.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, all three components have the same opportunity for improvement.  This case differs from Cases 1 and 2 since there are two factors, not present previously, that will affect the outcome of the allocation in this case.  First, each component in this case has a different reliability importance (impact of a component on the system&#039;s reliability); whereas in Cases 1 and 2, all three components were identical and had the same reliability importance.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure relimp shows the reliability importance for each component, where it can be seen that Component 1 has the greatest reliability importance and Component 3 has the smallest (this reliability importance also applies in Cases 4 and 5).  This indicates that the reliability of Component 1 should be significantly increased because it has the biggest impact on the overall system reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In addition, each component&#039;s cost function in Case 3 also depends on the difference between each component&#039;s initial reliability and its corresponding maximum achievable reliability.  (In Cases 1 and 2 this was not an issue because the components were identical.)  The greater this difference, the greater the cost of improving the reliability of a particular component relative to the other two components.  This difference between the initial reliability of a component and its maximum achievable reliability is called the range of improvement for that component.  Since all three components have the same maximum achievable reliability, Component 1, with the largest range for improvement, is the most cost efficient component to improve.  The improvement ranges for all three components are illustrated in Figure Rangeofimprovement. At the same time, however, there is a reliability value between the initial and the maximum achievable reliability beyond which it becomes cost prohibitive to improve any further.  This reliability value is dictated by the feasibility value.  From the table of results, it can be seen that in Case 3 there was a 25.52% improvement for Component 1, 16.49% for Component 2 and 7.65% for Component 3.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.19.png|thumb|center|300px|Reliability importance for Example 2, Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.20.png|thumb|center|300px|Range of improvement for each component for Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 4&#039;&#039;&#039; - As opposed to Case 3, Component 1 was assigned an even greater increase of 27.9%, with Components 2 and 3 receiving lesser increases than in Case 3, of 15.53% and 6.24% respectively.  This is due to the fact that Component 1 has an Easy feasibility and Component 3 has a Hard feasibility, which means that it is more difficult to increase the reliability of Component 3 than to increase the reliability of Component 1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 5&#039;&#039;&#039; - The feasibility values here are reversed with Component 1 having a Hard feasibility and Component 3 an Easy feasibility.  The recommended increase in Component 1&#039;s reliability is less compared to its increase for Cases 3 and 4.  Note, however, that Components 2 and 3 still received a smaller increase in reliability than Component 1 because their ranges of improvement are smaller.  In other words, Component 3 was assigned the smallest increase in reliability in Cases 3, 4 and 5 because its initial reliability is very close to its maximum achievable reliability. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Setting Specifications=&lt;br /&gt;
&lt;br /&gt;
This methodology could also be used to arrive at initial specifications for a set of components.  In the prior examples, we assumed a current reliability for the components.  One could repeat these steps by choosing an arbitrary (lower) initial reliability for each component, thus allowing the algorithm to travel up to the target.  When doing this, it is important to keep in mind the fact that both the distance from the target (the distance from the initial arbitrary value and the target value) for each component is also a significant contributor to the final results, as presented in the prior example.  If one wishes to arrive at the results using only the cost functions then it may be advantageous to set equal initial reliabilities for all components.&lt;br /&gt;
&lt;br /&gt;
=Other Notes on User-Defined Cost Functions=&lt;br /&gt;
&lt;br /&gt;
The optimization method in BlockSim is a very powerful tool for allocating reliability to the components of a system while minimizing an overall cost of improvement.  The default cost function in BlockSim was derived in order to model a general relationship between the cost and the component reliability.  However, if actual cost information is available, then one can use the cost data instead of using the default function.  Additionally, one can also view the feasibility in the default function as a measure of the difficulty in increasing the reliability of the component relative to the rest of the components to be optimized, assuming that they also follow the same cost function with the corresponding feasibility values.  If fault tolerance is a viable option, a reliability cost function for adding parallel units can be developed as demonstrated previously.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Another method for developing a reliability cost function would be to obtain different samples of components from different suppliers and test the samples to determine the reliability of each sample type.  From this data, a curve could be fitted through standard regression techniques and an equation defining the cost as a function of reliability could be developed.  Figure RGplot shows such a curve.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, and in cases where a reliability growth program is in place, the simplest way of obtaining a relationship between cost and reliability is by associating a cost to each development stage of the growth process.  Reliability growth models such as the Crow (AMSAA), Duane, Gompertz and Logistic models can be used to describe the cost as a function of reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.21.png|thumb|center|300px|Typical reliability growth curve generated using ReliaSoft&#039;s Reliability Growth software.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a reliability growth model has been successfully implemented, the development costs over the respective development time stages can be applied to the growth model, resulting in equations that describe reliability/cost relationships.  These equations can then be entered into BlockSim as user-defined cost functions (feasibility policies).  The only potential drawback to using growth model data is the lack of flexibility in applying the optimum results.  Making the cost projection for future stages of the project would require the assumption that development costs will be accrued at a similar rate in the future, which may not always be a valid assumption.  Also, if the optimization result suggests using a high reliability value for a component, it may take more time than is allotted for that project to attain the required reliability given the current reliability growth of the project.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15628</id>
		<title>Reliability Importance and Optimized Reliability Allocation (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15628"/>
		<updated>2012-02-13T23:24:16Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Static Reliability Importance */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|6}}&lt;br /&gt;
&lt;br /&gt;
=Component Reliability Importance=&lt;br /&gt;
===Static Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Once the reliability of a system has been determined, engineers are often faced with the task of identifying the least reliable component(s) in the system in order to improve the design.  For example, it was observed in Chapter 4 that the least reliable component in a series system has the biggest effect on the system reliability.  In this case, if the reliability of the system is to be improved, then the efforts can best be concentrated on improving the reliability of that component first.   In simple systems such as a series system, it is easy to identify the weak components.  However, in more complex systems this becomes quite a difficult task.  For complex systems, the analyst needs a mathematical approach that will provide the means of identifying and quantifying the importance of each component in the system.&lt;br /&gt;
&lt;br /&gt;
Using reliability importance measures is one method of identifying the relative importance of each component in a system with respect to the overall reliability of the system.  The reliability importance,  &amp;lt;math&amp;gt;{{I}_{R}}&amp;lt;/math&amp;gt; , of component  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;  in a system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  components is given by [[Appendix D: Weibull References | Leemis [17]]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{i}}}   \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{s}}&amp;lt;/math&amp;gt;  is the system reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  is the component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The value of the reliability importance given by Eqn. (importance) depends both on the reliability of a component and its corresponding position in the system.  In Chapter 4 we observed that for a simple series system (three components in series with reliabilities of 0.7, 0.8 and 0.9) the rate of increase of the system reliability was greatest when the least reliable component was improved.  In other words, it was observed that Component 1 had the largest reliability importance in the system relative to the other two components (see Figure Ch6fig1).  The same conclusion can be drawn by using Eqn. (importance) and obtaining the reliability importance in terms of a value for each component.&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Using BlockSim, the reliability importance values for these components can be calculated with Eqn. (importance).  Using the plot option and selecting the Static Reliability Importance plot type, Figure fig1a can be obtained.  Note that the time input required to create this plot is irrelevant for this example because the components are static.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The values shown in Figure fig1a for each component were obtained using Eqn. (importance).  The reliability equation for this series system was given by: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Taking the partial derivative of Eqn. (imp ex) with respect to  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{I}_{{{R}_{1}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{1}}}= &amp;amp; {{R}_{2}}{{R}_{3}} \\ &lt;br /&gt;
= &amp;amp; 0.8\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.72  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus the reliability importance of Component 1 is  &amp;lt;math&amp;gt;{{I}_{{{R}_{1}}}}=&amp;lt;/math&amp;gt;  0.72.  The reliability importance values for Components 2 and 3 are obtained in a similar manner.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.1.png|thumb|center|300px|Rate of change of system reliability when increasing the reliability of each component.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:6.2.gif|thumb|center|300px|Static reliability importance plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Time-Dependent Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
\The same concept applies if the components have a time-varying reliability.  That is, if  &amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt; , then one could compute  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  at any time  &amp;lt;math&amp;gt;x,&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}{{(t)}_{_{t=x}}}.&amp;lt;/math&amp;gt;   This is quantified in Eqn. (importance time). &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)=\frac{\partial {{R}_{s}}(t)}{\partial {{R}_{i}}(t)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In turn, this can be viewed as either a static plot (at a given time) or as time-varying plot, as illustrated in the next figures.  Specifically, Figures Ch6fig3, Ch6fig4 and Ch6fig5 present the analysis for three components configured reliability-wise in series following a Weibull distribution with  &amp;lt;math&amp;gt;\beta =3&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=1,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\eta }_{2}}=2,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=3,000&amp;lt;/math&amp;gt; .  Figure Ch6fig3 shows a bar chart of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  while Figure Ch6fig4 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  in BlockSim&#039;s tableau chart format.  In this chart, the area of the square is  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt; .  Lastly, Figure Ch6fig5 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  vs. time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that a system has failure modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; .  Furthermore, assume that failure of the entire system will occur if:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  occurs.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  occur.&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, assume the following failure probabilities for each mode.&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  have a mean time to occurrence of 1,000 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=1,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  has a mean time to occurrence of 100 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=100).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a mean time to occurrence of 700,000, 1,000,000 and 2,000,000 hours respectively (i.e. exponential with  &amp;lt;math&amp;gt;MTT{{F}_{B}}=700,000&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;MTT{{F}_{C}}=1,000,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;MTT{{F}_{F}}=2,000,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examine the mode importance for operating times of 100 and 500 hours.&lt;br /&gt;
&lt;br /&gt;
[[Image:6.3.gif|thumb|center|400px|Static Reliability Importance plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.4.png|thumb|center|400px|Static Reliability Importance tableau plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.5.png|thumb|center|400px|Reliability Importance vs. time plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
The RBD for this example is (from Chapter 4, Example 18):&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6ex1.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig6 illustrates  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt; .  It can be seen that even though  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a much rarer rate of occurrence, they are much more significant at 100 hours.  By 500 hours,  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt; , the effects of the lower reliability components become greatly pronounced and thus they become more important, as can be seen in Figure Ch6fig7.  Finally, the behavior of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  can be observed in Figure Ch6fig8.  Note that not all lines are plainly visible in Figure Ch6fig8 due to overlap.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Reliability Allocation=&lt;br /&gt;
&lt;br /&gt;
In the process of  developing a new product, the engineer is often faced with the task of designing a system that conforms to a set of reliability specifications.  The engineer is given the goal for the system and must then develop a design that will achieve the desired reliability of the system, while performing all of the system&#039;s intended functions at a minimum cost. This involves a balancing act of determining how to allocate reliability to the components in the system so the system will meet its reliability goal while at the same time ensuring that the system meets all of the other associated performance specifications.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.6.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.7.gif|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.8.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
The simplest method for allocating reliability is to distribute the reliabilities uniformly among all components. For example, suppose a system with five components in series has a reliability objective of 90% for a given operating time. The uniform allocation of the objective to all components would require each component to have a reliability of 98% for the specified operating time, since  &amp;lt;math&amp;gt;{{0.98}^{5}}\tilde{=}0.90&amp;lt;/math&amp;gt;. While this manner of allocation is easy to calculate, it is generally not the best way to allocate reliability for a system. The optimum method of allocating reliability would take into account the cost or relative difficulty of improving the reliability of different subsystems or components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability optimization process begins with the development of a model that represents the entire system.  This is accomplished with the construction of a system reliability block diagram that represents the reliability relationships of the components in the system.  From this model, the system reliability impact of different component modifications can be estimated and considered alongside the costs that would be incurred in the process of making those modifications.  It is then possible to perform an optimization analysis for this problem, finding the best combination of component reliability improvements that meet or exceed the performance goals at the lowest cost.&lt;br /&gt;
&lt;br /&gt;
===Importance Measures and FMEA/FMECA===&lt;br /&gt;
&lt;br /&gt;
Traditional Failure Mode and Effects analysis (FMEA/FMECA) relies on Risk Priority Numbers (RPNs) or criticality calculations to identify and prioritize the significance/importance of different failure modes.  The RPN methodology (and to some extent, the criticality methodology) tend to be subjective.  When conducting these types of analyses, one may wish to incorporate more quantitative metrics, such as the importance measure presented here and/or the RS FCI and RS DECI for repairable systems (which are discussed in later chapters).  ReliaSoft&#039;s Xfmea software can be used to export an FMEA/FMECA analysis to BlockSim.  The documentation that accompanies Xfmea provides more information on FMEA/FMECA, including both methods of risk assessment.&lt;br /&gt;
&lt;br /&gt;
=Improving Reliability=&lt;br /&gt;
Reliability engineers are very often called upon to make decisions as to whether to improve a certain component or components in order to achieve a minimum required system reliability.  There are two approaches to improving the reliability of a system: fault avoidance and fault tolerance.  Fault avoidance is achieved by using high-quality and high-reliability components and is usually less expensive than fault tolerance.  Fault tolerance, on the other hand, is achieved by redundancy.  Redundancy can result in increased design complexity and increased costs through additional weight, space, etc.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before deciding whether to improve the reliability of a system by fault tolerance or fault avoidance, a reliability assessment for each component in the system should be made.  Once the reliability values for the components have been quantified, an analysis can be performed in order to determine if that system&#039;s reliability goal will be met.  If it becomes apparent that the system&#039;s reliability will not be adequate to meet the desired goal at the specified mission duration, steps can be taken to determine the best way to improve the system&#039;s reliability so that it will reach the desired target.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a system with three components connected reliability-wise in series.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 70%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 80% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 90%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 85%, is required for this system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}=50.4%&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, this is far short of the system&#039;s required reliability performance.  It is apparent that the reliability of the system&#039;s constituent components will need to be increased in order for the system to meet its goal.  First, we will try increasing the reliability of one component at a time to see whether the reliability goal can be achieved.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig9 shows that even by raising the individual component reliability to a hypothetical value of 1 (100% reliability, which implies that the component will never fail), the overall system reliability goal will not be met by improving the reliability of just one component.  The next logical step would be to try to increase the reliability of two components.  The question now becomes: which two?  One might also suggest increasing the reliability of all three components.  A basis for making such decisions needs to be found in order to avoid the ``trial and error&#039;&#039; aspect of altering the system&#039;s components randomly in an attempt to achieve the system reliability goal.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.9.gif|thumb|center|400px|Change in system reliability of a three-unit series system due to increasing the reliability of just one component.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As we have seen, the reliability goal for the preceding example could not be achieved by increasing the reliability of just one component.  There are cases, however, where increasing the reliability of one component results in achieving the system reliability goal.  Consider, for example, a system with three components connected reliability-wise in parallel.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 60%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 70% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 80%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 99%, is required for this system.  The initial system reliability is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-(1-0.6)\cdot (1-0.7)\cdot (1-0.8)=0.976&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current system reliability is inadequate to meet the goal.  Once again, we can try to meet the system reliability goal by raising the reliability of just one of the three components in the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From Figure fig10, it can be seen that the reliability goal can be reached by improving Component 1, Component 2 or Component 3.  The reliability engineer is now faced with another dilemma:  which component&#039;s reliability should be improved? This presents a new aspect to the problem of allocating the reliability of the system.  Since we know that the system reliability goal can be achieved by increasing at least one unit, the question becomes one of how to do this most efficiently and cost effectively.  We will need more information to make an informed decision as to how to go about improving the system&#039;s reliability.  How much does each component need to be improved for the system to meet its goal?  How feasible is it to improve the reliability of each component?  Would it actually be more efficient to slightly raise the reliability of two or three components rather than radically improving only one?&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to answer these questions, we must introduce another variable into the problem &amp;lt;math&amp;gt;:\ \ \ &amp;lt;/math&amp;gt; cost.  Cost does not necessarily have to be in dollars.  It could be described in terms of non-monetary resources, such as time.  By associating cost values to the reliabilities of the system&#039;s components, we can find an optimum design that will provide the required reliability at a minimum cost.&lt;br /&gt;
&lt;br /&gt;
===Cost/Penalty Function===&lt;br /&gt;
&lt;br /&gt;
There is always a cost associated with changing a design due to change of vendors, use of higher-quality materials, retooling costs, administrative fees, etc.  The cost as a function of the reliability for each component must be quantified before attempting to improve the reliability.  Otherwise, the design changes may result in a system that is needlessly expensive or overdesigned.  Developing the ``cost of reliability&#039;&#039; relationship will give the engineer an understanding of which components to improve and how to best concentrate the effort and allocate resources in doing so.  The first step will be to obtain a relationship between the cost of improvement and reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.10.png|thumb|center|400px|Meeting a reliability goal requirement by increasing a component&#039;s reliability]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The preferred approach would be to formulate the cost function from actual cost data.  This can be done from past experience.  If a reliability growth program is in place, the costs associated with each stage of improvement can also be quantified.  Defining the different costs associated with different vendors or different component models is also useful in formulating a model of component cost as a function of reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, there are many cases where no such information is available.  For this reason, a general (default) behavior model of the cost versus the component&#039;s reliability was developed for performing reliability optimization in BlockSim.  The objective of this function is to model an overall cost behavior for all types of components.  Of course, it is impossible to formulate a model that will be precisely applicable to every situation; but the proposed relationship is general enough to cover most applications.  In addition to the default model formulation, BlockSim does allow the definition of user-defined cost models.&lt;br /&gt;
&lt;br /&gt;
====Quantifying the Cost/Penalty Function====&lt;br /&gt;
&lt;br /&gt;
One needs to quantify a cost function for each component,  &amp;lt;math&amp;gt;{{C}_{i}}&amp;lt;/math&amp;gt; , in terms of the reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , of each component, or:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}=f({{R}_{i}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This function should:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the current reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Current}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the maximum possible reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Max}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Allow for different levels of difficulty (or cost) in increasing the reliability of each component.  It can take into account:&amp;lt;br&amp;gt;&lt;br /&gt;
::o	design issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	supplier issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	state of technology.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	time-to-market issues, etc.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus, for the cost function to comply with these needs, the following conditions should be adhered to:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be constrained by the minimum and maximum reliabilities of each component (i.e. reliability must be less than one and greater than the current reliability of the component or at least greater than zero).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should not be linear, but rather quantify the fact that it is incrementally harder to improve reliability.  For example, it is considerably easier to increase the reliability from 90% to 91% than to increase it from 99.99% to 99.999%, even though the increase is larger in the first case.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be asymptotic to the maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The following default cost function (also used in BlockSim) adheres to all of these conditions and acts like a penalty function for increasing a component&#039;s reliability.  Furthermore, an exponential behavior for the cost is assumed since it should get exponentially more difficult to increase the reliability. See Mettas [21]. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})&amp;lt;/math&amp;gt;  is the penalty (or cost) function as a function of component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  is the feasibility (or cost index) of improving a component&#039;s reliability relative to the other components in the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{min,i}}&amp;lt;/math&amp;gt;  is the current reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{max,i}}&amp;lt;/math&amp;gt;  is the maximum achievable reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
Note that this penalty function is dimensionless.  It essentially acts as a weighting factor that describes the difficulty in increasing the component reliability from its current value, relative to the other components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examining the cost function given by Eqn. (Default Cost), the following observations can be made:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability departs from the minimum or current value of reliability.  It is assumed that the reliabilities for the components will not take values any lower than they already have.  Depending on the optimization, a component&#039;s reliability may not need to be increased from its current value but it will not drop any lower.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability approaches the maximum achievable reliability.  This is a reliability value that is approached asymptotically as the cost increases but is never actually reached.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost is a function of the range of improvement, which is the difference between the component&#039;s initial reliability and the corresponding maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The exponent in Eqn. (Default Cost) approaches infinity as the component&#039;s reliability approaches its maximum achievable value.  This means that it is easier to increase the reliability of a component from a lower initial value.  For example, it is easier to increase a component&#039;s reliability from 70% to 75% than increasing its reliability from 90% to 95%.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====The Feasibility Term,  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
The feasibility term in Eqn. (Default Cost) is a constant (or an equation parameter) that represents the difficulty in increasing a component&#039;s reliability relative to the rest of the components in the system.  Depending on the design complexity, technological limitations, etc., certain components can be very hard to improve.  Clearly, the more difficult it is to improve the reliability of the component, the greater the cost.  Figure feasplot illustrates the behavior of the function defined in Eqn. (Default Cost) for different values of  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; .  It can be seen that the lower the feasibility value, the more rapidly the cost function approaches infinity.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Several methods can be used to obtain a feasibility value.  Weighting factors for allocating reliability have been proposed by many authors and can be used to quantify feasibility.  These weights depend on certain factors of influence, such as the complexity of the component, the state of the art, the operational profile, the criticality, etc.  Engineering judgment based on past experience, supplier quality, supplier availability and other factors can also be used in determining a feasibility value.  Overall, the assignment of a feasibility value is going to be a subjective process.  Of course, this problem is negated if the relationship between the cost and the reliability for each component is known because one can use regression methods to estimate the parameter value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.11.gif|thumb|center|400px|Behavior of the cost function for different feasibility values.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Maximum Achievable Reliability====&lt;br /&gt;
&lt;br /&gt;
For the purposes of reliability optimization, we also need to define a limiting reliability that a component will approach, but not reach.  The costs near the maximum achievable reliability are very high and the actual value for the maximum reliability is usually dictated by technological or financial constraints.  In deciding on a value to use for the maximum achievable reliability, the current state of the art of the component in question and other similar factors will have to be considered.  In the end, a realistic estimation based on engineering judgment and experience will be necessary to assign a value to this input.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that the time associated with this maximum achievable reliability is the same as that of the overall system reliability goal.  Almost any component can achieve a very high reliability value, provided the mission time is short enough.  For example, a component with an exponential distribution and a failure rate of one failure per hour has a reliability that drops below 1% for missions greater than five hours.  However, it can achieve a reliability of 99.9% as long as the mission is no longer than four seconds.  For the purposes of optimization in BlockSim, the reliability values of the components are associated with the time for which the system reliability goal is specified.  For example, if the problem is to achieve a system goal of 99% reliability at 1,000 hours, the maximum achievable reliability values entered for the individual components would be the maximum reliability that each component could attain for a mission of 1,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As the component reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , approaches the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function approaches infinity.  The maximum achievable reliability acts as a scale parameter for the cost function.  By decreasing  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function is compressed between  &amp;lt;math&amp;gt;{{R}_{i,min}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , as shown in Figure oldfig5.&lt;br /&gt;
 &lt;br /&gt;
====Cost Function====&lt;br /&gt;
Once the cost functions for the individual components have been determined, it becomes necessary to develop an expression for the overall system cost.  This takes the form of:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{s}}({{R}_{G}})={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+...+{{C}_{n}}({{R}_{n}}),i=1,2,...,n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, the cost of the system is simply the sum of the costs of its components.  This is regardless of the form of the individual component cost functions.  They can be of the general behavior model in BlockSim or they can be user-defined.   Once the overall cost function for the system has been defined, the problem becomes one of minimizing the cost function while remaining within the constraints defined by the target system reliability and the reliability ranges for the components.  The latter constraints in this case are defined by the minimum and maximum reliability values for the individual components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.12.png|thumb|center|400px|Effect5 of the maximum achievable reliability on the cost function.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim employs a nonlinear programming technique to minimize the system cost function.  The system has a minimum (current) and theoretical maximum reliability value that is defined by the minimum and maximum reliabilities of the components and by the way the system is configured.  That is, the structural properties of the system are accounted for in the determination of the optimum solution.  For example, the optimization for a system of three units in series will be different from the optimization for a system consisting of the same three units in parallel.  The optimization occurs by varying the reliability values of the components within their respective constraints of maximum and minimum reliability in a way that the overall system goal is achieved.  Obviously, there can be any number of different combinations of component reliability values that might achieve the system goal.  The optimization routine essentially finds the combination that results in the lowest overall system cost. &lt;br /&gt;
&lt;br /&gt;
====Determining the Optimum Allocation Scheme====&lt;br /&gt;
&lt;br /&gt;
To determine the optimum reliability allocation, the analyst first determines the system reliability equation (the objective function).  As an example, and again for a trivial system with three components in series, this would be:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a target reliability of 90% is sought, then Eqn. (optAlloc) is recast as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The objective now is to solve for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  so that the equality in Eqn. (optAlloc90) is satisfied.  To obtain an optimum solution, we also need to use our cost functions (i.e. define the total allocation costs) as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
With the cost equation defined, then the optimum values for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  are the values that satisfy the reliability requirement, Eqn. (optAlloc90), at the minimum cost, Eqn. (optcost).  BlockSim uses this methodology during the optimization task.&lt;br /&gt;
&lt;br /&gt;
====Defining a Feasibility Policy in BlockSim====&lt;br /&gt;
&lt;br /&gt;
In BlockSim you can choose to use the default feasibility function, as defined by Eqn. (Default Cost), or use your own function.  Figure BSfvalues illustrates the use of the default values using the slider control. Figure BSFcustom shows the use of an associated feasibility policy to create a user-defined cost function.  When defining your own cost function, you should be aware of/adhere to the following guidelines:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Because the cost functions are evaluated relative to each other, they should be correlated.  In other words, if one function evaluates to 10,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=10&amp;lt;/math&amp;gt;  for one block and 20 for another,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=20&amp;lt;/math&amp;gt; , then the implication is that there is a 1 to 2 cost relation.  &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Do not mix your own function with the software&#039;s default functions unless you have verified that your cost functions are defined and correlated to the default cost functions, as defined by Eqn. (Default Cost).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Your function should adhere to the guidelines presented earlier.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Lastly, and since the evaluation is relative, it is preferable to use the pre-defined functions unless you have a compelling reason (or data) to do otherwise.  The last section in this chapter describes cases where user-defined functions are preferred.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.13.png|thumb|center|300px|Setting the default feasibility function in BlockSim with the feasibility slider. Note that the feasibility slider displays values, &#039;&#039;SV&#039;&#039;, from 1 to 9 when moved by the user, with SV=9 being the hardest. The relationship between &#039;&#039;f&#039;&#039; and &#039;&#039;SV&#039;&#039; is ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.14.png|thumb|center|400px|Setting a user-defined feasibility function in BlockSim utilizing an assiciated feasibility policy. Any user-defined equation can be entered as a function of &#039;&#039;R.&#039;&#039;]]&lt;br /&gt;
&lt;br /&gt;
=Implementing the Optimization=&lt;br /&gt;
&lt;br /&gt;
As was mentioned earlier, there are two different methods of implementing the changes suggested by the reliability optimization routine: fault tolerance and fault avoidance.  When the optimized component reliabilities have been determined, it does not matter which of the two methods is employed to realize the optimum reliability for the component in question.  For example, suppose we have determined that a component must have its reliability for a certain mission time raised from 50% to 75%.  The engineer must now decide how to go about implementing the increase in reliability.  If the engineer decides to do this via fault avoidance, another component must be found (or the existing component must be redesigned) so that it will perform the same function with a higher reliability.  On the other hand, if the engineer decides to go the fault tolerance route, the optimized reliability can be achieved merely by placing a second identical component in parallel with the first one.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, the method of implementing the reliability optimization is going to be related to the cost function and this is something the reliability engineer must take into account when deciding on what type of cost function is used for the optimization.  In fact, if we take a closer look at the fault tolerance scheme, we can see some parallels with the general behavior cost model included in BlockSim.  For example, consider a system that consists of a single unit.  The cost of that unit, including all associated mounting and hardware costs, is one dollar.  The reliability of this unit for a given mission time is 30%.  It has been determined that this is inadequate and that a second component is to be added in parallel to increase the reliability.  Thus, the reliability for the two-unit parallel system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-{{(1-0.3)}^{2}}=0.51\text{ or }51%&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
So, the reliability has increased by a value of 21% and the cost has increased by one dollar.  In a similar fashion, we can continue to add more units in parallel, thus increasing the reliability and the cost.  We now have an array of reliability values and the associated costs that we can use to develop a cost function for this fault tolerance scheme.  Figure costredundant shows the relationship between cost and reliability for this example.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As can be seen, this looks quite similar to the general behavior cost model presented earlier.  In fact, a standard regression analysis available in Weibull++ indicates that an exponential model fits this cost model quite well.   The function is given by the following equation, where  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is the cost in dollars and  &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt;  is the fractional reliability value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C(R)=0.3756\cdot {{e}^{3.1972\cdot R}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.15.gif|thumb|center|400px|Cost function for redundant parallel units.]]&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
&lt;br /&gt;
Consider a system consisting of three components connected reliability-wise in series.  Assume the objective reliability for the system is 90% for a mission time of 100 hours.  Five cases will be considered for the allocation problem. See Mettas [21].&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1 - All three components are identical with times-to-failure that are described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.318 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 312 hours. All three components have the same feasibility value of Moderate (5).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2 - Same as in Case 1, but Component 1 has a feasibility of Easy, Component 2 has a feasibility of Moderate and Component 3 has a feasibility of Hard.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3 - Component 1 has 70% reliability, Component 2 has 80% reliability and Component 3 has 90% reliability, all for a mission duration of 100 hours.  All three components have the same feasibility of Easy.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 4 - Component 1 has 70% reliability and Easy feasibility, Component 2 has 80% reliability and Moderate feasibility, and Component 3 has 90% reliability and Hard feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 5 - Component 1 has 70% reliability and Hard feasibility, Component 2 has 80% reliability and Easy feasibility and Component 3 has 90% reliability and Moderate feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In all cases, the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , for each component is 99.9% for a mission duration of 100 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.16.gif|thumb|center|300px|Optimization inputs in BlockSim&#039;s Analytical QCP for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Solution====&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Case 1&#039;&#039;&#039; - The reliability equation for Case 1 is: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the equality desired is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}(t=100)\cdot {{R}_{2}}(t=100)\cdot {{R}_{3}}(t=100)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{1,2,3}}={{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cost or feasibility function is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{1,2,3}}({{R}_{1,2,3}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
And where  &amp;lt;math&amp;gt;{{R}_{\max _{1,2,3}^{}}}=0.999&amp;lt;/math&amp;gt;  (arbitrarily set),  &amp;lt;math&amp;gt;{{R}_{\min _{1,2,3}^{}}}&amp;lt;/math&amp;gt;  computed from the reliability function of each component at the time of interest,  &amp;lt;math&amp;gt;t=100&amp;lt;/math&amp;gt; , or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{\min _{1,2,3}^{}}}= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( \tfrac{100}{312} \right)}^{1.318}}}} \\ &lt;br /&gt;
= &amp;amp; 0.79995  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  obtained from: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
f= &amp;amp; \left( 1-\frac{5}{10} \right) \\ &lt;br /&gt;
= &amp;amp; 0.5  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The solution,  &amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}&amp;lt;/math&amp;gt; , is the one that satisfies Eqn. (exbjective2) while minimizing Eqn. (exonstraint).  In this case (and since all the components are identical), the target reliability is found to be: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}(t=100)=0.9655&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Figures QCPOpt and OptResults show related BlockSim screens.  Based on this, each component&#039;s reliability should be at least 96.55% at 100 hours in order for the system&#039;s reliability to be 90% at 100 hours.  Note the column labeled N.E.P.U. in the Results Panel shown in Figure OptResults.  This stands for &amp;quot;Number of Equivalent Parallel Units&amp;quot; and represents the number of redundant units that would be required to bring that particular component up to the recommended reliability.  In the case where the fault tolerance approach is to be implemented, the N.E.P.U value should be rounded up to an integer.  Therefore, some manipulation by the engineer is required in order to ensure that the chosen integer values will yield the required system reliability goal (or exceed it).  In addition, further cost analysis should be performed in order to account for the costs of adding redundancy to the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Additionally, and when the results have been obtained, the engineer may wish to re-scale the components based on their distribution parameters instead of the fixed reliability value.  In the case of these components, one may wish to re-scale the scale parameter of the distribution ,  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt; , for the components, or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{100}{\eta } \right)}^{1.318}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.17.png|thumb|center|300px|Optimization results for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Which yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\eta }_{{{O}_{i}}}}=1269.48&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The Parameter Experimenter in BlockSim can also be used for this (Figure paramexper).  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The results from the other cases can be obtained in a similar fashion.  The results for Cases 1 through 5 are summarized next.&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{matrix}&lt;br /&gt;
   {} &amp;amp; Case 1 &amp;amp; Case 2 &amp;amp; Case 3 &amp;amp; Case 4 &amp;amp; Case 5  \\&lt;br /&gt;
   Component 1 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9874} &amp;amp; \text{0}\text{.9552} &amp;amp; \text{0}\text{.9790} &amp;amp; \text{0}\text{.9295}  \\&lt;br /&gt;
   Component 2 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9633} &amp;amp; \text{0}\text{.9649} &amp;amp; \text{0}\text{.9553} &amp;amp; \text{0}\text{.9884}  \\&lt;br /&gt;
   Component 3 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9463} &amp;amp; \text{0}\text{.9765} &amp;amp; \text{0}\text{.9624} &amp;amp; \text{0}\text{.9797}  \\&lt;br /&gt;
\end{matrix}&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 2&#039;&#039;&#039; - It can be seen that the highest reliability was allocated to Component 1 with the Easy feasibility.  The lowest reliability was assigned to Component 3 with the Hard feasibility.  This makes sense in that an optimized reliability scheme will call for the greatest reliability changes in those components that are the easiest to change.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 3&#039;&#039;&#039; - The components were different but had the same feasibility values.  &lt;br /&gt;
&lt;br /&gt;
[[Image:BS6.18.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, all three components have the same opportunity for improvement.  This case differs from Cases 1 and 2 since there are two factors, not present previously, that will affect the outcome of the allocation in this case.  First, each component in this case has a different reliability importance (impact of a component on the system&#039;s reliability); whereas in Cases 1 and 2, all three components were identical and had the same reliability importance.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure relimp shows the reliability importance for each component, where it can be seen that Component 1 has the greatest reliability importance and Component 3 has the smallest (this reliability importance also applies in Cases 4 and 5).  This indicates that the reliability of Component 1 should be significantly increased because it has the biggest impact on the overall system reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In addition, each component&#039;s cost function in Case 3 also depends on the difference between each component&#039;s initial reliability and its corresponding maximum achievable reliability.  (In Cases 1 and 2 this was not an issue because the components were identical.)  The greater this difference, the greater the cost of improving the reliability of a particular component relative to the other two components.  This difference between the initial reliability of a component and its maximum achievable reliability is called the range of improvement for that component.  Since all three components have the same maximum achievable reliability, Component 1, with the largest range for improvement, is the most cost efficient component to improve.  The improvement ranges for all three components are illustrated in Figure Rangeofimprovement. At the same time, however, there is a reliability value between the initial and the maximum achievable reliability beyond which it becomes cost prohibitive to improve any further.  This reliability value is dictated by the feasibility value.  From the table of results, it can be seen that in Case 3 there was a 25.52% improvement for Component 1, 16.49% for Component 2 and 7.65% for Component 3.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.19.png|thumb|center|300px|Reliability importance for Example 2, Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.20.png|thumb|center|300px|Range of improvement for each component for Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 4&#039;&#039;&#039; - As opposed to Case 3, Component 1 was assigned an even greater increase of 27.9%, with Components 2 and 3 receiving lesser increases than in Case 3, of 15.53% and 6.24% respectively.  This is due to the fact that Component 1 has an Easy feasibility and Component 3 has a Hard feasibility, which means that it is more difficult to increase the reliability of Component 3 than to increase the reliability of Component 1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 5&#039;&#039;&#039; - The feasibility values here are reversed with Component 1 having a Hard feasibility and Component 3 an Easy feasibility.  The recommended increase in Component 1&#039;s reliability is less compared to its increase for Cases 3 and 4.  Note, however, that Components 2 and 3 still received a smaller increase in reliability than Component 1 because their ranges of improvement are smaller.  In other words, Component 3 was assigned the smallest increase in reliability in Cases 3, 4 and 5 because its initial reliability is very close to its maximum achievable reliability. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Setting Specifications=&lt;br /&gt;
&lt;br /&gt;
This methodology could also be used to arrive at initial specifications for a set of components.  In the prior examples, we assumed a current reliability for the components.  One could repeat these steps by choosing an arbitrary (lower) initial reliability for each component, thus allowing the algorithm to travel up to the target.  When doing this, it is important to keep in mind the fact that both the distance from the target (the distance from the initial arbitrary value and the target value) for each component is also a significant contributor to the final results, as presented in the prior example.  If one wishes to arrive at the results using only the cost functions then it may be advantageous to set equal initial reliabilities for all components.&lt;br /&gt;
&lt;br /&gt;
=Other Notes on User-Defined Cost Functions=&lt;br /&gt;
&lt;br /&gt;
The optimization method in BlockSim is a very powerful tool for allocating reliability to the components of a system while minimizing an overall cost of improvement.  The default cost function in BlockSim was derived in order to model a general relationship between the cost and the component reliability.  However, if actual cost information is available, then one can use the cost data instead of using the default function.  Additionally, one can also view the feasibility in the default function as a measure of the difficulty in increasing the reliability of the component relative to the rest of the components to be optimized, assuming that they also follow the same cost function with the corresponding feasibility values.  If fault tolerance is a viable option, a reliability cost function for adding parallel units can be developed as demonstrated previously.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Another method for developing a reliability cost function would be to obtain different samples of components from different suppliers and test the samples to determine the reliability of each sample type.  From this data, a curve could be fitted through standard regression techniques and an equation defining the cost as a function of reliability could be developed.  Figure RGplot shows such a curve.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, and in cases where a reliability growth program is in place, the simplest way of obtaining a relationship between cost and reliability is by associating a cost to each development stage of the growth process.  Reliability growth models such as the Crow (AMSAA), Duane, Gompertz and Logistic models can be used to describe the cost as a function of reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.21.png|thumb|center|300px|Typical reliability growth curve generated using ReliaSoft&#039;s Reliability Growth software.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a reliability growth model has been successfully implemented, the development costs over the respective development time stages can be applied to the growth model, resulting in equations that describe reliability/cost relationships.  These equations can then be entered into BlockSim as user-defined cost functions (feasibility policies).  The only potential drawback to using growth model data is the lack of flexibility in applying the optimum results.  Making the cost projection for future stages of the project would require the assumption that development costs will be accrued at a similar rate in the future, which may not always be a valid assumption.  Also, if the optimization result suggests using a high reliability value for a component, it may take more time than is allotted for that project to attain the required reliability given the current reliability growth of the project.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15609</id>
		<title>Reliability Importance and Optimized Reliability Allocation (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Reliability_Importance_and_Optimized_Reliability_Allocation_(Analytical)&amp;diff=15609"/>
		<updated>2012-02-13T23:00:00Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Static Reliability Importance */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|6}}&lt;br /&gt;
&lt;br /&gt;
=Component Reliability Importance=&lt;br /&gt;
===Static Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Once the reliability of a system has been determined, engineers are often faced with the task of identifying the least reliable component(s) in the system in order to improve the design.  For example, it was observed in Chapter 4 that the least reliable component in a series system has the biggest effect on the system reliability.  In this case, if the reliability of the system is to be improved, then the efforts can best be concentrated on improving the reliability of that component first.   In simple systems such as a series system, it is easy to identify the weak components.  However, in more complex systems this becomes quite a difficult task.  For complex systems, the analyst needs a mathematical approach that will provide the means of identifying and quantifying the importance of each component in the system.&lt;br /&gt;
&lt;br /&gt;
Using reliability importance measures is one method of identifying the relative importance of each component in a system with respect to the overall reliability of the system.  The reliability importance,  &amp;lt;math&amp;gt;{{I}_{R}}&amp;lt;/math&amp;gt; , of component  &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;  in a system of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  components is given by [[Appendix D: Weibull References | Leemis [17]]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{i}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{s}}&amp;lt;/math&amp;gt;  is the system reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  is the component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The value of the reliability importance given by Eqn. (importance) depends both on the reliability of a component and its corresponding position in the system.  In Chapter 4 we observed that for a simple series system (three components in series with reliabilities of 0.7, 0.8 and 0.9) the rate of increase of the system reliability was greatest when the least reliable component was improved.  In other words, it was observed that Component 1 had the largest reliability importance in the system relative to the other two components (see Figure Ch6fig1).  The same conclusion can be drawn by using Eqn. (importance) and obtaining the reliability importance in terms of a value for each component.&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
Using BlockSim, the reliability importance values for these components can be calculated with Eqn. (importance).  Using the plot option and selecting the Static Reliability Importance plot type, Figure fig1a can be obtained.  Note that the time input required to create this plot is irrelevant for this example because the components are static.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The values shown in Figure fig1a for each component were obtained using Eqn. (importance).  The reliability equation for this series system was given by: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Taking the partial derivative of Eqn. (imp ex) with respect to  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{I}_{{{R}_{1}}}}=\frac{\partial {{R}_{s}}}{\partial {{R}_{1}}}= &amp;amp; {{R}_{2}}{{R}_{3}} \\ &lt;br /&gt;
= &amp;amp; 0.8\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.72  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus the reliability importance of Component 1 is  &amp;lt;math&amp;gt;{{I}_{{{R}_{1}}}}=&amp;lt;/math&amp;gt;  0.72.  The reliability importance values for Components 2 and 3 are obtained in a similar manner.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.1.png|thumb|center|300px|Rate of change of system reliability when increasing the reliability of each component.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:6.2.gif|thumb|center|300px|Static reliability importance plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Time-Dependent Reliability Importance===&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
\The same concept applies if the components have a time-varying reliability.  That is, if  &amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt; , then one could compute  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  at any time  &amp;lt;math&amp;gt;x,&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}{{(t)}_{_{t=x}}}.&amp;lt;/math&amp;gt;   This is quantified in Eqn. (importance time). &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)=\frac{\partial {{R}_{s}}(t)}{\partial {{R}_{i}}(t)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In turn, this can be viewed as either a static plot (at a given time) or as time-varying plot, as illustrated in the next figures.  Specifically, Figures Ch6fig3, Ch6fig4 and Ch6fig5 present the analysis for three components configured reliability-wise in series following a Weibull distribution with  &amp;lt;math&amp;gt;\beta =3&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=1,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;{{\eta }_{2}}=2,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=3,000&amp;lt;/math&amp;gt; .  Figure Ch6fig3 shows a bar chart of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  while Figure Ch6fig4 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt;  in BlockSim&#039;s tableau chart format.  In this chart, the area of the square is  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}&amp;lt;/math&amp;gt; .  Lastly, Figure Ch6fig5 shows the  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  vs. time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Assume that a system has failure modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; .  Furthermore, assume that failure of the entire system will occur if:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  occurs.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  occur.&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, assume the following failure probabilities for each mode.&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;  have a mean time to occurrence of 1,000 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=1,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Mode  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  has a mean time to occurrence of 100 hours (i.e. exponential with  &amp;lt;math&amp;gt;MTTF=100).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Modes  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a mean time to occurrence of 700,000, 1,000,000 and 2,000,000 hours respectively (i.e. exponential with  &amp;lt;math&amp;gt;MTT{{F}_{B}}=700,000&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;MTT{{F}_{C}}=1,000,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;MTT{{F}_{F}}=2,000,000).&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examine the mode importance for operating times of 100 and 500 hours.&lt;br /&gt;
&lt;br /&gt;
[[Image:6.3.gif|thumb|center|400px|Static Reliability Importance plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.4.png|thumb|center|400px|Static Reliability Importance tableau plot at &#039;&#039;t&#039;&#039;=1,000.]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.5.png|thumb|center|400px|Reliability Importance vs. time plot.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
The RBD for this example is (from Chapter 4, Example 18):&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6ex1.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig6 illustrates  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt; .  It can be seen that even though  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;  have a much rarer rate of occurrence, they are much more significant at 100 hours.  By 500 hours,  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt; , the effects of the lower reliability components become greatly pronounced and thus they become more important, as can be seen in Figure Ch6fig7.  Finally, the behavior of  &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;  can be observed in Figure Ch6fig8.  Note that not all lines are plainly visible in Figure Ch6fig8 due to overlap.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Reliability Allocation=&lt;br /&gt;
&lt;br /&gt;
In the process of  developing a new product, the engineer is often faced with the task of designing a system that conforms to a set of reliability specifications.  The engineer is given the goal for the system and must then develop a design that will achieve the desired reliability of the system, while performing all of the system&#039;s intended functions at a minimum cost. This involves a balancing act of determining how to allocate reliability to the components in the system so the system will meet its reliability goal while at the same time ensuring that the system meets all of the other associated performance specifications.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.6.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=100)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.7.gif|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t=500)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
[[Image:6.8.png|thumb|center|400px|Plot of &amp;lt;math&amp;gt;{{I}_{{{R}_{i}}}}(t)&amp;lt;/math&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
The simplest method for allocating reliability is to distribute the reliabilities uniformly among all components. For example, suppose a system with five components in series has a reliability objective of 90% for a given operating time. The uniform allocation of the objective to all components would require each component to have a reliability of 98% for the specified operating time, since  &amp;lt;math&amp;gt;{{0.98}^{5}}\tilde{=}0.90&amp;lt;/math&amp;gt;. While this manner of allocation is easy to calculate, it is generally not the best way to allocate reliability for a system. The optimum method of allocating reliability would take into account the cost or relative difficulty of improving the reliability of different subsystems or components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability optimization process begins with the development of a model that represents the entire system.  This is accomplished with the construction of a system reliability block diagram that represents the reliability relationships of the components in the system.  From this model, the system reliability impact of different component modifications can be estimated and considered alongside the costs that would be incurred in the process of making those modifications.  It is then possible to perform an optimization analysis for this problem, finding the best combination of component reliability improvements that meet or exceed the performance goals at the lowest cost.&lt;br /&gt;
&lt;br /&gt;
===Importance Measures and FMEA/FMECA===&lt;br /&gt;
&lt;br /&gt;
Traditional Failure Mode and Effects analysis (FMEA/FMECA) relies on Risk Priority Numbers (RPNs) or criticality calculations to identify and prioritize the significance/importance of different failure modes.  The RPN methodology (and to some extent, the criticality methodology) tend to be subjective.  When conducting these types of analyses, one may wish to incorporate more quantitative metrics, such as the importance measure presented here and/or the RS FCI and RS DECI for repairable systems (which are discussed in later chapters).  ReliaSoft&#039;s Xfmea software can be used to export an FMEA/FMECA analysis to BlockSim.  The documentation that accompanies Xfmea provides more information on FMEA/FMECA, including both methods of risk assessment.&lt;br /&gt;
&lt;br /&gt;
=Improving Reliability=&lt;br /&gt;
Reliability engineers are very often called upon to make decisions as to whether to improve a certain component or components in order to achieve a minimum required system reliability.  There are two approaches to improving the reliability of a system: fault avoidance and fault tolerance.  Fault avoidance is achieved by using high-quality and high-reliability components and is usually less expensive than fault tolerance.  Fault tolerance, on the other hand, is achieved by redundancy.  Redundancy can result in increased design complexity and increased costs through additional weight, space, etc.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before deciding whether to improve the reliability of a system by fault tolerance or fault avoidance, a reliability assessment for each component in the system should be made.  Once the reliability values for the components have been quantified, an analysis can be performed in order to determine if that system&#039;s reliability goal will be met.  If it becomes apparent that the system&#039;s reliability will not be adequate to meet the desired goal at the specified mission duration, steps can be taken to determine the best way to improve the system&#039;s reliability so that it will reach the desired target.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a system with three components connected reliability-wise in series.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 70%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 80% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 90%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 85%, is required for this system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}=50.4%&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, this is far short of the system&#039;s required reliability performance.  It is apparent that the reliability of the system&#039;s constituent components will need to be increased in order for the system to meet its goal.  First, we will try increasing the reliability of one component at a time to see whether the reliability goal can be achieved.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure Ch6fig9 shows that even by raising the individual component reliability to a hypothetical value of 1 (100% reliability, which implies that the component will never fail), the overall system reliability goal will not be met by improving the reliability of just one component.  The next logical step would be to try to increase the reliability of two components.  The question now becomes: which two?  One might also suggest increasing the reliability of all three components.  A basis for making such decisions needs to be found in order to avoid the ``trial and error&#039;&#039; aspect of altering the system&#039;s components randomly in an attempt to achieve the system reliability goal.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS6.9.gif|thumb|center|400px|Change in system reliability of a three-unit series system due to increasing the reliability of just one component.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As we have seen, the reliability goal for the preceding example could not be achieved by increasing the reliability of just one component.  There are cases, however, where increasing the reliability of one component results in achieving the system reliability goal.  Consider, for example, a system with three components connected reliability-wise in parallel.  The reliabilities for each component for a given time are:  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  = 60%,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  = 70% and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  = 80%.  A reliability goal,  &amp;lt;math&amp;gt;{{R}_{G}}&amp;lt;/math&amp;gt;  = 99%, is required for this system.  The initial system reliability is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-(1-0.6)\cdot (1-0.7)\cdot (1-0.8)=0.976&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The current system reliability is inadequate to meet the goal.  Once again, we can try to meet the system reliability goal by raising the reliability of just one of the three components in the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From Figure fig10, it can be seen that the reliability goal can be reached by improving Component 1, Component 2 or Component 3.  The reliability engineer is now faced with another dilemma:  which component&#039;s reliability should be improved? This presents a new aspect to the problem of allocating the reliability of the system.  Since we know that the system reliability goal can be achieved by increasing at least one unit, the question becomes one of how to do this most efficiently and cost effectively.  We will need more information to make an informed decision as to how to go about improving the system&#039;s reliability.  How much does each component need to be improved for the system to meet its goal?  How feasible is it to improve the reliability of each component?  Would it actually be more efficient to slightly raise the reliability of two or three components rather than radically improving only one?&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to answer these questions, we must introduce another variable into the problem &amp;lt;math&amp;gt;:\ \ \ &amp;lt;/math&amp;gt; cost.  Cost does not necessarily have to be in dollars.  It could be described in terms of non-monetary resources, such as time.  By associating cost values to the reliabilities of the system&#039;s components, we can find an optimum design that will provide the required reliability at a minimum cost.&lt;br /&gt;
&lt;br /&gt;
===Cost/Penalty Function===&lt;br /&gt;
&lt;br /&gt;
There is always a cost associated with changing a design due to change of vendors, use of higher-quality materials, retooling costs, administrative fees, etc.  The cost as a function of the reliability for each component must be quantified before attempting to improve the reliability.  Otherwise, the design changes may result in a system that is needlessly expensive or overdesigned.  Developing the ``cost of reliability&#039;&#039; relationship will give the engineer an understanding of which components to improve and how to best concentrate the effort and allocate resources in doing so.  The first step will be to obtain a relationship between the cost of improvement and reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.10.png|thumb|center|400px|Meeting a reliability goal requirement by increasing a component&#039;s reliability]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The preferred approach would be to formulate the cost function from actual cost data.  This can be done from past experience.  If a reliability growth program is in place, the costs associated with each stage of improvement can also be quantified.  Defining the different costs associated with different vendors or different component models is also useful in formulating a model of component cost as a function of reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, there are many cases where no such information is available.  For this reason, a general (default) behavior model of the cost versus the component&#039;s reliability was developed for performing reliability optimization in BlockSim.  The objective of this function is to model an overall cost behavior for all types of components.  Of course, it is impossible to formulate a model that will be precisely applicable to every situation; but the proposed relationship is general enough to cover most applications.  In addition to the default model formulation, BlockSim does allow the definition of user-defined cost models.&lt;br /&gt;
&lt;br /&gt;
====Quantifying the Cost/Penalty Function====&lt;br /&gt;
&lt;br /&gt;
One needs to quantify a cost function for each component,  &amp;lt;math&amp;gt;{{C}_{i}}&amp;lt;/math&amp;gt; , in terms of the reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , of each component, or:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}=f({{R}_{i}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This function should:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the current reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Current}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Look at the maximum possible reliability of the component,  &amp;lt;math&amp;gt;{{R}_{Max}}&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Allow for different levels of difficulty (or cost) in increasing the reliability of each component.  It can take into account:&amp;lt;br&amp;gt;&lt;br /&gt;
::o	design issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	supplier issues.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	state of technology.&amp;lt;br&amp;gt;&lt;br /&gt;
::o	time-to-market issues, etc.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Thus, for the cost function to comply with these needs, the following conditions should be adhered to:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be constrained by the minimum and maximum reliabilities of each component (i.e. reliability must be less than one and greater than the current reliability of the component or at least greater than zero).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should not be linear, but rather quantify the fact that it is incrementally harder to improve reliability.  For example, it is considerably easier to increase the reliability from 90% to 91% than to increase it from 99.99% to 99.999%, even though the increase is larger in the first case.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The function should be asymptotic to the maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The following default cost function (also used in BlockSim) adheres to all of these conditions and acts like a penalty function for increasing a component&#039;s reliability.  Furthermore, an exponential behavior for the cost is assumed since it should get exponentially more difficult to increase the reliability. See Mettas [21]. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})&amp;lt;/math&amp;gt;  is the penalty (or cost) function as a function of component reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  is the feasibility (or cost index) of improving a component&#039;s reliability relative to the other components in the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{min,i}}&amp;lt;/math&amp;gt;  is the current reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{max,i}}&amp;lt;/math&amp;gt;  is the maximum achievable reliability at the time at which the optimization is to be performed.&amp;lt;br&amp;gt;&lt;br /&gt;
Note that this penalty function is dimensionless.  It essentially acts as a weighting factor that describes the difficulty in increasing the component reliability from its current value, relative to the other components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Examining the cost function given by Eqn. (Default Cost), the following observations can be made:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability departs from the minimum or current value of reliability.  It is assumed that the reliabilities for the components will not take values any lower than they already have.  Depending on the optimization, a component&#039;s reliability may not need to be increased from its current value but it will not drop any lower.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost increases as the allocated reliability approaches the maximum achievable reliability.  This is a reliability value that is approached asymptotically as the cost increases but is never actually reached.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The cost is a function of the range of improvement, which is the difference between the component&#039;s initial reliability and the corresponding maximum achievable reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The exponent in Eqn. (Default Cost) approaches infinity as the component&#039;s reliability approaches its maximum achievable value.  This means that it is easier to increase the reliability of a component from a lower initial value.  For example, it is easier to increase a component&#039;s reliability from 70% to 75% than increasing its reliability from 90% to 95%.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====The Feasibility Term,  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;====&lt;br /&gt;
&lt;br /&gt;
The feasibility term in Eqn. (Default Cost) is a constant (or an equation parameter) that represents the difficulty in increasing a component&#039;s reliability relative to the rest of the components in the system.  Depending on the design complexity, technological limitations, etc., certain components can be very hard to improve.  Clearly, the more difficult it is to improve the reliability of the component, the greater the cost.  Figure feasplot illustrates the behavior of the function defined in Eqn. (Default Cost) for different values of  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; .  It can be seen that the lower the feasibility value, the more rapidly the cost function approaches infinity.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Several methods can be used to obtain a feasibility value.  Weighting factors for allocating reliability have been proposed by many authors and can be used to quantify feasibility.  These weights depend on certain factors of influence, such as the complexity of the component, the state of the art, the operational profile, the criticality, etc.  Engineering judgment based on past experience, supplier quality, supplier availability and other factors can also be used in determining a feasibility value.  Overall, the assignment of a feasibility value is going to be a subjective process.  Of course, this problem is negated if the relationship between the cost and the reliability for each component is known because one can use regression methods to estimate the parameter value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.11.gif|thumb|center|400px|Behavior of the cost function for different feasibility values.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Maximum Achievable Reliability====&lt;br /&gt;
&lt;br /&gt;
For the purposes of reliability optimization, we also need to define a limiting reliability that a component will approach, but not reach.  The costs near the maximum achievable reliability are very high and the actual value for the maximum reliability is usually dictated by technological or financial constraints.  In deciding on a value to use for the maximum achievable reliability, the current state of the art of the component in question and other similar factors will have to be considered.  In the end, a realistic estimation based on engineering judgment and experience will be necessary to assign a value to this input.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that the time associated with this maximum achievable reliability is the same as that of the overall system reliability goal.  Almost any component can achieve a very high reliability value, provided the mission time is short enough.  For example, a component with an exponential distribution and a failure rate of one failure per hour has a reliability that drops below 1% for missions greater than five hours.  However, it can achieve a reliability of 99.9% as long as the mission is no longer than four seconds.  For the purposes of optimization in BlockSim, the reliability values of the components are associated with the time for which the system reliability goal is specified.  For example, if the problem is to achieve a system goal of 99% reliability at 1,000 hours, the maximum achievable reliability values entered for the individual components would be the maximum reliability that each component could attain for a mission of 1,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As the component reliability,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , approaches the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function approaches infinity.  The maximum achievable reliability acts as a scale parameter for the cost function.  By decreasing  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , the cost function is compressed between  &amp;lt;math&amp;gt;{{R}_{i,min}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , as shown in Figure oldfig5.&lt;br /&gt;
 &lt;br /&gt;
====Cost Function====&lt;br /&gt;
Once the cost functions for the individual components have been determined, it becomes necessary to develop an expression for the overall system cost.  This takes the form of:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{s}}({{R}_{G}})={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+...+{{C}_{n}}({{R}_{n}}),i=1,2,...,n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, the cost of the system is simply the sum of the costs of its components.  This is regardless of the form of the individual component cost functions.  They can be of the general behavior model in BlockSim or they can be user-defined.   Once the overall cost function for the system has been defined, the problem becomes one of minimizing the cost function while remaining within the constraints defined by the target system reliability and the reliability ranges for the components.  The latter constraints in this case are defined by the minimum and maximum reliability values for the individual components.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.12.png|thumb|center|400px|Effect5 of the maximum achievable reliability on the cost function.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim employs a nonlinear programming technique to minimize the system cost function.  The system has a minimum (current) and theoretical maximum reliability value that is defined by the minimum and maximum reliabilities of the components and by the way the system is configured.  That is, the structural properties of the system are accounted for in the determination of the optimum solution.  For example, the optimization for a system of three units in series will be different from the optimization for a system consisting of the same three units in parallel.  The optimization occurs by varying the reliability values of the components within their respective constraints of maximum and minimum reliability in a way that the overall system goal is achieved.  Obviously, there can be any number of different combinations of component reliability values that might achieve the system goal.  The optimization routine essentially finds the combination that results in the lowest overall system cost. &lt;br /&gt;
&lt;br /&gt;
====Determining the Optimum Allocation Scheme====&lt;br /&gt;
&lt;br /&gt;
To determine the optimum reliability allocation, the analyst first determines the system reliability equation (the objective function).  As an example, and again for a trivial system with three components in series, this would be:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a target reliability of 90% is sought, then Eqn. (optAlloc) is recast as:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The objective now is to solve for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  so that the equality in Eqn. (optAlloc90) is satisfied.  To obtain an optimum solution, we also need to use our cost functions (i.e. define the total allocation costs) as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
With the cost equation defined, then the optimum values for  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  are the values that satisfy the reliability requirement, Eqn. (optAlloc90), at the minimum cost, Eqn. (optcost).  BlockSim uses this methodology during the optimization task.&lt;br /&gt;
&lt;br /&gt;
====Defining a Feasibility Policy in BlockSim====&lt;br /&gt;
&lt;br /&gt;
In BlockSim you can choose to use the default feasibility function, as defined by Eqn. (Default Cost), or use your own function.  Figure BSfvalues illustrates the use of the default values using the slider control. Figure BSFcustom shows the use of an associated feasibility policy to create a user-defined cost function.  When defining your own cost function, you should be aware of/adhere to the following guidelines:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Because the cost functions are evaluated relative to each other, they should be correlated.  In other words, if one function evaluates to 10,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=10&amp;lt;/math&amp;gt;  for one block and 20 for another,  &amp;lt;math&amp;gt;{{C}_{i}}({{R}_{i}})=20&amp;lt;/math&amp;gt; , then the implication is that there is a 1 to 2 cost relation.  &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Do not mix your own function with the software&#039;s default functions unless you have verified that your cost functions are defined and correlated to the default cost functions, as defined by Eqn. (Default Cost).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Your function should adhere to the guidelines presented earlier.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Lastly, and since the evaluation is relative, it is preferable to use the pre-defined functions unless you have a compelling reason (or data) to do otherwise.  The last section in this chapter describes cases where user-defined functions are preferred.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.13.png|thumb|center|300px|Setting the default feasibility function in BlockSim with the feasibility slider. Note that the feasibility slider displays values, &#039;&#039;SV&#039;&#039;, from 1 to 9 when moved by the user, with SV=9 being the hardest. The relationship between &#039;&#039;f&#039;&#039; and &#039;&#039;SV&#039;&#039; is ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.14.png|thumb|center|400px|Setting a user-defined feasibility function in BlockSim utilizing an assiciated feasibility policy. Any user-defined equation can be entered as a function of &#039;&#039;R.&#039;&#039;]]&lt;br /&gt;
&lt;br /&gt;
=Implementing the Optimization=&lt;br /&gt;
&lt;br /&gt;
As was mentioned earlier, there are two different methods of implementing the changes suggested by the reliability optimization routine: fault tolerance and fault avoidance.  When the optimized component reliabilities have been determined, it does not matter which of the two methods is employed to realize the optimum reliability for the component in question.  For example, suppose we have determined that a component must have its reliability for a certain mission time raised from 50% to 75%.  The engineer must now decide how to go about implementing the increase in reliability.  If the engineer decides to do this via fault avoidance, another component must be found (or the existing component must be redesigned) so that it will perform the same function with a higher reliability.  On the other hand, if the engineer decides to go the fault tolerance route, the optimized reliability can be achieved merely by placing a second identical component in parallel with the first one.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Obviously, the method of implementing the reliability optimization is going to be related to the cost function and this is something the reliability engineer must take into account when deciding on what type of cost function is used for the optimization.  In fact, if we take a closer look at the fault tolerance scheme, we can see some parallels with the general behavior cost model included in BlockSim.  For example, consider a system that consists of a single unit.  The cost of that unit, including all associated mounting and hardware costs, is one dollar.  The reliability of this unit for a given mission time is 30%.  It has been determined that this is inadequate and that a second component is to be added in parallel to increase the reliability.  Thus, the reliability for the two-unit parallel system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}=1-{{(1-0.3)}^{2}}=0.51\text{ or }51%&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
So, the reliability has increased by a value of 21% and the cost has increased by one dollar.  In a similar fashion, we can continue to add more units in parallel, thus increasing the reliability and the cost.  We now have an array of reliability values and the associated costs that we can use to develop a cost function for this fault tolerance scheme.  Figure costredundant shows the relationship between cost and reliability for this example.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As can be seen, this looks quite similar to the general behavior cost model presented earlier.  In fact, a standard regression analysis available in Weibull++ indicates that an exponential model fits this cost model quite well.   The function is given by the following equation, where  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is the cost in dollars and  &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt;  is the fractional reliability value.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;C(R)=0.3756\cdot {{e}^{3.1972\cdot R}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.15.gif|thumb|center|400px|Cost function for redundant parallel units.]]&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
&lt;br /&gt;
Consider a system consisting of three components connected reliability-wise in series.  Assume the objective reliability for the system is 90% for a mission time of 100 hours.  Five cases will be considered for the allocation problem. See Mettas [21].&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1 - All three components are identical with times-to-failure that are described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.318 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 312 hours. All three components have the same feasibility value of Moderate (5).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2 - Same as in Case 1, but Component 1 has a feasibility of Easy, Component 2 has a feasibility of Moderate and Component 3 has a feasibility of Hard.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3 - Component 1 has 70% reliability, Component 2 has 80% reliability and Component 3 has 90% reliability, all for a mission duration of 100 hours.  All three components have the same feasibility of Easy.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 4 - Component 1 has 70% reliability and Easy feasibility, Component 2 has 80% reliability and Moderate feasibility, and Component 3 has 90% reliability and Hard feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 5 - Component 1 has 70% reliability and Hard feasibility, Component 2 has 80% reliability and Easy feasibility and Component 3 has 90% reliability and Moderate feasibility, all for a mission duration of 100 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In all cases, the maximum achievable reliability,  &amp;lt;math&amp;gt;{{R}_{i,max}}&amp;lt;/math&amp;gt; , for each component is 99.9% for a mission duration of 100 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.16.gif|thumb|center|300px|Optimization inputs in BlockSim&#039;s Analytical QCP for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Solution====&lt;br /&gt;
 &lt;br /&gt;
&#039;&#039;&#039;Case 1&#039;&#039;&#039; - The reliability equation for Case 1 is: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{_{S}}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the equality desired is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{R}_{1}}(t=100)\cdot {{R}_{2}}(t=100)\cdot {{R}_{3}}(t=100)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{1,2,3}}={{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The cost or feasibility function is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{T}}={{C}_{1}}({{R}_{1}})+{{C}_{2}}({{R}_{2}})+{{C}_{3}}({{R}_{3}})&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{C}_{1,2,3}}({{R}_{1,2,3}})={{e}^{(1-f)\cdot \tfrac{{{R}_{i}}-{{R}_{\min ,i}}}{{{R}_{\max ,i}}-{{R}_{i}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
And where  &amp;lt;math&amp;gt;{{R}_{\max _{1,2,3}^{}}}=0.999&amp;lt;/math&amp;gt;  (arbitrarily set),  &amp;lt;math&amp;gt;{{R}_{\min _{1,2,3}^{}}}&amp;lt;/math&amp;gt;  computed from the reliability function of each component at the time of interest,  &amp;lt;math&amp;gt;t=100&amp;lt;/math&amp;gt; , or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{\min _{1,2,3}^{}}}= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( \tfrac{100}{312} \right)}^{1.318}}}} \\ &lt;br /&gt;
= &amp;amp; 0.79995  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And  &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;  obtained from: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
f= &amp;amp; \left( 1-\frac{5}{10} \right) \\ &lt;br /&gt;
= &amp;amp; 0.5  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The solution,  &amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}&amp;lt;/math&amp;gt; , is the one that satisfies Eqn. (exbjective2) while minimizing Eqn. (exonstraint).  In this case (and since all the components are identical), the target reliability is found to be: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{{{O}_{i}}}}(t=100)=0.9655&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Figures QCPOpt and OptResults show related BlockSim screens.  Based on this, each component&#039;s reliability should be at least 96.55% at 100 hours in order for the system&#039;s reliability to be 90% at 100 hours.  Note the column labeled N.E.P.U. in the Results Panel shown in Figure OptResults.  This stands for &amp;quot;Number of Equivalent Parallel Units&amp;quot; and represents the number of redundant units that would be required to bring that particular component up to the recommended reliability.  In the case where the fault tolerance approach is to be implemented, the N.E.P.U value should be rounded up to an integer.  Therefore, some manipulation by the engineer is required in order to ensure that the chosen integer values will yield the required system reliability goal (or exceed it).  In addition, further cost analysis should be performed in order to account for the costs of adding redundancy to the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Additionally, and when the results have been obtained, the engineer may wish to re-scale the components based on their distribution parameters instead of the fixed reliability value.  In the case of these components, one may wish to re-scale the scale parameter of the distribution ,  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt; , for the components, or:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
0.9655= &amp;amp; {{e}^{-{{\left( \tfrac{100}{\eta } \right)}^{1.318}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.17.png|thumb|center|300px|Optimization results for Example 2, Case 1.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Which yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{\eta }_{{{O}_{i}}}}=1269.48&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The Parameter Experimenter in BlockSim can also be used for this (Figure paramexper).  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The results from the other cases can be obtained in a similar fashion.  The results for Cases 1 through 5 are summarized next.&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{matrix}&lt;br /&gt;
   {} &amp;amp; Case 1 &amp;amp; Case 2 &amp;amp; Case 3 &amp;amp; Case 4 &amp;amp; Case 5  \\&lt;br /&gt;
   Component 1 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9874} &amp;amp; \text{0}\text{.9552} &amp;amp; \text{0}\text{.9790} &amp;amp; \text{0}\text{.9295}  \\&lt;br /&gt;
   Component 2 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9633} &amp;amp; \text{0}\text{.9649} &amp;amp; \text{0}\text{.9553} &amp;amp; \text{0}\text{.9884}  \\&lt;br /&gt;
   Component 3 &amp;amp; \text{0}\text{.9655} &amp;amp; \text{0}\text{.9463} &amp;amp; \text{0}\text{.9765} &amp;amp; \text{0}\text{.9624} &amp;amp; \text{0}\text{.9797}  \\&lt;br /&gt;
\end{matrix}&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 2&#039;&#039;&#039; - It can be seen that the highest reliability was allocated to Component 1 with the Easy feasibility.  The lowest reliability was assigned to Component 3 with the Hard feasibility.  This makes sense in that an optimized reliability scheme will call for the greatest reliability changes in those components that are the easiest to change.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 3&#039;&#039;&#039; - The components were different but had the same feasibility values.  &lt;br /&gt;
&lt;br /&gt;
[[Image:BS6.18.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In other words, all three components have the same opportunity for improvement.  This case differs from Cases 1 and 2 since there are two factors, not present previously, that will affect the outcome of the allocation in this case.  First, each component in this case has a different reliability importance (impact of a component on the system&#039;s reliability); whereas in Cases 1 and 2, all three components were identical and had the same reliability importance.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure relimp shows the reliability importance for each component, where it can be seen that Component 1 has the greatest reliability importance and Component 3 has the smallest (this reliability importance also applies in Cases 4 and 5).  This indicates that the reliability of Component 1 should be significantly increased because it has the biggest impact on the overall system reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In addition, each component&#039;s cost function in Case 3 also depends on the difference between each component&#039;s initial reliability and its corresponding maximum achievable reliability.  (In Cases 1 and 2 this was not an issue because the components were identical.)  The greater this difference, the greater the cost of improving the reliability of a particular component relative to the other two components.  This difference between the initial reliability of a component and its maximum achievable reliability is called the range of improvement for that component.  Since all three components have the same maximum achievable reliability, Component 1, with the largest range for improvement, is the most cost efficient component to improve.  The improvement ranges for all three components are illustrated in Figure Rangeofimprovement. At the same time, however, there is a reliability value between the initial and the maximum achievable reliability beyond which it becomes cost prohibitive to improve any further.  This reliability value is dictated by the feasibility value.  From the table of results, it can be seen that in Case 3 there was a 25.52% improvement for Component 1, 16.49% for Component 2 and 7.65% for Component 3.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.19.png|thumb|center|300px|Reliability importance for Example 2, Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:6.20.png|thumb|center|300px|Range of improvement for each component for Cases 3, 4, and 5.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 4&#039;&#039;&#039; - As opposed to Case 3, Component 1 was assigned an even greater increase of 27.9%, with Components 2 and 3 receiving lesser increases than in Case 3, of 15.53% and 6.24% respectively.  This is due to the fact that Component 1 has an Easy feasibility and Component 3 has a Hard feasibility, which means that it is more difficult to increase the reliability of Component 3 than to increase the reliability of Component 1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Case 5&#039;&#039;&#039; - The feasibility values here are reversed with Component 1 having a Hard feasibility and Component 3 an Easy feasibility.  The recommended increase in Component 1&#039;s reliability is less compared to its increase for Cases 3 and 4.  Note, however, that Components 2 and 3 still received a smaller increase in reliability than Component 1 because their ranges of improvement are smaller.  In other words, Component 3 was assigned the smallest increase in reliability in Cases 3, 4 and 5 because its initial reliability is very close to its maximum achievable reliability. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Setting Specifications=&lt;br /&gt;
&lt;br /&gt;
This methodology could also be used to arrive at initial specifications for a set of components.  In the prior examples, we assumed a current reliability for the components.  One could repeat these steps by choosing an arbitrary (lower) initial reliability for each component, thus allowing the algorithm to travel up to the target.  When doing this, it is important to keep in mind the fact that both the distance from the target (the distance from the initial arbitrary value and the target value) for each component is also a significant contributor to the final results, as presented in the prior example.  If one wishes to arrive at the results using only the cost functions then it may be advantageous to set equal initial reliabilities for all components.&lt;br /&gt;
&lt;br /&gt;
=Other Notes on User-Defined Cost Functions=&lt;br /&gt;
&lt;br /&gt;
The optimization method in BlockSim is a very powerful tool for allocating reliability to the components of a system while minimizing an overall cost of improvement.  The default cost function in BlockSim was derived in order to model a general relationship between the cost and the component reliability.  However, if actual cost information is available, then one can use the cost data instead of using the default function.  Additionally, one can also view the feasibility in the default function as a measure of the difficulty in increasing the reliability of the component relative to the rest of the components to be optimized, assuming that they also follow the same cost function with the corresponding feasibility values.  If fault tolerance is a viable option, a reliability cost function for adding parallel units can be developed as demonstrated previously.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Another method for developing a reliability cost function would be to obtain different samples of components from different suppliers and test the samples to determine the reliability of each sample type.  From this data, a curve could be fitted through standard regression techniques and an equation defining the cost as a function of reliability could be developed.  Figure RGplot shows such a curve.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, and in cases where a reliability growth program is in place, the simplest way of obtaining a relationship between cost and reliability is by associating a cost to each development stage of the growth process.  Reliability growth models such as the Crow (AMSAA), Duane, Gompertz and Logistic models can be used to describe the cost as a function of reliability.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:6.21.png|thumb|center|300px|Typical reliability growth curve generated using ReliaSoft&#039;s Reliability Growth software.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If a reliability growth model has been successfully implemented, the development costs over the respective development time stages can be applied to the growth model, resulting in equations that describe reliability/cost relationships.  These equations can then be entered into BlockSim as user-defined cost functions (feasibility policies).  The only potential drawback to using growth model data is the lack of flexibility in applying the optimum results.  Making the cost projection for future stages of the project would require the assumption that development costs will be accrued at a similar rate in the future, which may not always be a valid assumption.  Also, if the optimization result suggests using a high reliability value for a component, it may take more time than is allotted for that project to attain the required reliability given the current reliability growth of the project.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15582</id>
		<title>Time-Dependent System Reliability (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15582"/>
		<updated>2012-02-13T22:31:31Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Example 6 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|5}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the previous chapter, different system configuration types were examined, as well as different methods for obtaining the system&#039;s reliability function analytically.  Because the reliabilities in the problems presented were treated as probabilities (e.g.  &amp;lt;math&amp;gt;P(A)&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; ), the reliability values and equations presented were referred to as static (not time-dependent).  Thus, in the prior chapter, the life distributions of the components were not incorporated in the process of calculating the system reliability.  In this chapter, time dependency in the reliability function will be introduced.  We will develop the models necessary to observe the reliability over the life of the system, instead of at just one point in time.  In addition, performance measures such as failure rate, MTTF and warranty time will be estimated for the entire system.  The methods of obtaining the reliability function analytically remain identical to the ones presented in the previous chapter, with the exception that the reliabilities will be functions of time.  In other words, instead of dealing with  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , we will use  &amp;lt;math&amp;gt;{{R}_{i}}(t)&amp;lt;/math&amp;gt; .  All examples in this chapter assume that no repairs are performed on the components.  &lt;br /&gt;
&lt;br /&gt;
=Analytical Life Predictions=&lt;br /&gt;
The analytical approach presented in the prior chapter involved the determination of a mathematical expression that describes the reliability of the system, expressed in terms of the reliabilities of its components.  So far we have estimated only static system reliability (at a fixed time).  For example, in the case of a system with three components in series, the system&#039;s reliability equation was given by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}  \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The values of  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  were given for a common time and the reliability of the system was estimated for that time.  However, since the component failure characteristics can be described by distributions, the system reliability is actually time-dependent.  In this case, Eqn. (1) can be rewritten as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability of the system for any mission time can now be estimated.  Assuming a Weibull life distribution for each component, Eqn. (1) can now be expressed in terms of each component&#039;s reliability function, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the same manner, any life distribution can be substituted into the system reliability equation.  Suppose that the times-to-failure of the first component are described with a Weibull distribution, the times-to-failure of the second component with an exponential distribution and the times-to-failure of the third component with a normal distribution.  Then Eqn. (1) can be written as:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot \left[ 1-\Phi \left( \frac{t-{{\mu }_{3}}}{{{\sigma }_{3}}} \right) \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It can be seen that the biggest challenge is in obtaining the system&#039;s reliability function in terms of component reliabilities, which has already been discussed in depth.  Once this has been achieved, calculating the reliability of the system for any mission duration is just a matter of substituting the corresponding component reliability functions into the system reliability equation.&lt;br /&gt;
===Advantages of the Analytical Method===&lt;br /&gt;
The primary advantage of the analytical solution is that it produces a mathematical expression that describes the reliability of the system.  Once the system&#039;s reliability function has been determined, other calculations can then be performed to obtain metrics of interest for the system. Such calculations include:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of warranty periods.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s failure rate.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s MTTF.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, optimization and reliability allocation techniques can be used to aid engineers in their design improvement efforts.  Another advantage in using analytical techniques is the ability to perform static calculations and analyze systems with a mixture of static and time-dependent components.  Finally, the reliability importance of components over time can be calculated with this methodology.&lt;br /&gt;
===Disadvantages of the Analytical Method===&lt;br /&gt;
The biggest disadvantage of the analytical method is that formulations can become very complicated.  The more complicated a system is, the larger and more difficult it will be to analytically formulate an expression for the system&#039;s reliability.  For particularly detailed systems this process can be quite time-consuming, even with the use of computers.  Furthermore, when the maintainability of the system or some of its components must be taken into consideration, analytical solutions become intractable.  In these situations, the use of simulation methods may be more advantageous than attempting to develop a solution analytically.  Simulation methods are presented in later chapters.&lt;br /&gt;
===Looking at a Simple &#039;&#039;Complex&#039;&#039; System Analytically===&lt;br /&gt;
&lt;br /&gt;
The complexity involved in an analytical solution can be best illustrated by looking at the simple &#039;&#039;complex&#039;&#039; system with 15 components, as shown in Figure 5.1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1.png|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.1 An RBD of a complex system.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system reliability for this system (computed using BlockSim) is shown next.  The first solution is provided using BlockSim&#039;s symbolic solution.  In symbolic mode, BlockSim breaks the equation into segments, identified by tokens, that need to be substituted into the final system equation for a complete solution.  This creates algebraic solutions that are more compact than if the substitutions were made.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; D2\cdot D3\cdot {{R}_{L}} \\ &lt;br /&gt;
D3= &amp;amp; +{{R}_{K}}\cdot IK \\ &lt;br /&gt;
IK= &amp;amp; +{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}} \\ &lt;br /&gt;
D2 = &amp;amp; +{{R}_{A}}\cdot {{R}_{E}}\cdot IE \\ &lt;br /&gt;
IE = &amp;amp; -D1\cdot {{R}_{M}}\cdot {{R}_{N}}+{{R}_{M}}\cdot {{R}_{N}}+D1 \\ &lt;br /&gt;
D1 = &amp;amp; +{{R}_{D}}\cdot ID \\ &lt;br /&gt;
ID = &amp;amp; -{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting the terms yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; {{R}_{A}}\cdot {{R}_{E}}\cdot {{R}_{L}}\cdot {{R}_{K}} \\ &lt;br /&gt;
&amp;amp; \cdot \{({{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})\cdot {{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}}-{{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}\} \\ &lt;br /&gt;
&amp;amp; \cdot \{{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}}\}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
BlockSim&#039;s automatic algebraic simplification would yield the following format for the above solution: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; (({{R}_{A}}\cdot {{R}_{E}}(-({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})){{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})))) \\ &lt;br /&gt;
&amp;amp; ({{R}_{K}}({{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; +RI\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}})){{R}_{L}})  \ (eqn 2)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this equation, each  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  represents the reliability function of a block.  For example, if  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  has a Weibull distribution, then each  &amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;  and so forth.  Substitution of each component&#039;s reliability function in Eqn.2 will result in an analytical expression for the system reliability as a function of time, or  &amp;lt;math&amp;gt;{{R}_{s}}(t)&amp;lt;/math&amp;gt; , which is the same as  &amp;lt;math&amp;gt;(1-cd{{f}_{System}}).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Obtaining Other Functions of Interest===&lt;br /&gt;
Once the system reliability equation (or the cumulative density function,  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt; ) has been determined, other functions and metrics of interest can be derived.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Consider the following simple system:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-2.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.2 Simple two-component system. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Furthermore, assume that component 1 follows an exponential distribution with a mean of 10,000 (&amp;lt;math&amp;gt;\mu =10,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;\lambda =1/10,000)&amp;lt;/math&amp;gt;  and component 2 follows a Weibull distribution with  &amp;lt;math&amp;gt;\beta =6&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt; .  The reliability equation of this system is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}  \ (eqn 3) &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{F}_{S}}(t)= &amp;amp; 1-({{R}_{1}}(t)\cdot {{R}_{2}}(t)) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \right) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
{{analytical system pdf}}&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability====&lt;br /&gt;
Conditional reliability is the probability of a system successfully completing another mission following the successful completion of a previous mission.  The time of the previous mission and the time for the mission to be undertaken must be taken into account for conditional reliability calculations.  The system&#039;s conditional reliability function is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(T,t)=\frac{R(T+t)}{R(T)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Eqn.6 gives the reliability for a new mission of duration  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  having already accumulated  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  hours of operation up to the start of this new mission. The system is evaluated to assure that it will start the next mission successfully.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2, the reliability for mission of  &amp;lt;math&amp;gt;t=1,000&amp;lt;/math&amp;gt;  hours, having an age of  &amp;lt;math&amp;gt;T=500&amp;lt;/math&amp;gt;  hours, is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(T=500,t=1000)= &amp;amp; \frac{R(T+t)}{R(T)} \\ &lt;br /&gt;
= &amp;amp; \frac{R(1500)}{R(500)} \\ &lt;br /&gt;
= &amp;amp; \frac{{{e}^{-\tfrac{1500}{10,000}}}\cdot {{e}^{-{{\left( \tfrac{1500}{10,000} \right)}^{6}}}}}{{{e}^{-\tfrac{500}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{500}{10,000} \right)}^{6}}}}} \\ &lt;br /&gt;
= &amp;amp; 0.9048=90.48%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.3.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.4.png|thumb|center|400px|]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability for Components====&lt;br /&gt;
&lt;br /&gt;
Now in this formulation, it was assumed that the accumulated age was equivalent for both units. That is, both started life at zero and aged to 500.  It is possible to consider an individual component that has already accumulated some age (used component) in the same formulation.  To illustrate this, assume that component 2 started life with an age of T=100.  Then the reliability equation of the system, as given in Eqn.3, would need to be modified to include a conditional term for 2, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{R}_{1}}(t)\cdot \frac{{{R}_{2}}({{T}_{2}}+t)}{{{R}_{2}}({{T}_{2}})} \ (eqn 7)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In BlockSim, the start age input box may be used to specify a starting age greater than zero.&lt;br /&gt;
{{system failure rate analytical}}&lt;br /&gt;
&lt;br /&gt;
====System Mean Life (Mean Time To Failure)====&lt;br /&gt;
The mean life (or mean time to failure, MTTF) can be obtained by integrating the system reliability function from zero to infinity: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;MTTF=\int_{0}^{\infty }{{R}_{s}}\left( t \right)dt   \ (eqn 10)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mean time is a performance index and does not provide any information about the behavior of the failure distribution of the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)dt \\ &lt;br /&gt;
= &amp;amp; 5978.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Warranty Period and BX Life====&lt;br /&gt;
Sometimes it is desirable to know the time value associated with a certain reliability.  Warranty periods are often calculated by determining what percentage of the failure population can be covered financially and estimating the time at which this portion of the population will fail.  Similarly, engineering specifications may call for a certain BX life, which also represents a time period during which a certain proportion of the population will fail.  For example, the B10 life is the time in which 10% of the population will fail.  &lt;br /&gt;
This is obtained by setting  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  to the desired value and solving for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}\left( t \right)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To compute the time by which reliability would be equal to 90%, Eqn.11 is recast as follows and solved for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this case,  &amp;lt;math&amp;gt;t=1053.59&amp;lt;/math&amp;gt; .  Equivalently, the B10 life for this system is also  &amp;lt;math&amp;gt;1053.59&amp;lt;/math&amp;gt; .&lt;br /&gt;
Except for some trivial cases, a closed form solution for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  cannot be obtained.   Thus, it is necessary to solve for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  using numerical methods.  BlockSim uses numerical methods.&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
Consider a system consisting of three exponential units in series with the following failure rates (in failures per hour):  &amp;lt;math&amp;gt;{{\lambda }_{1}}&amp;lt;/math&amp;gt;  = 0.0002,  &amp;lt;math&amp;gt;{{\lambda }_{2}}&amp;lt;/math&amp;gt;  = 0.0005 and  &amp;lt;math&amp;gt;{{\lambda }_{3}}&amp;lt;/math&amp;gt;  = 0.0001.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the reliability equation for the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the reliability of the system after 150 hours of operation?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s failure rate equation.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the MTTF for the system?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What should the warranty period be for a 90% reliability?&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The analytical expression for the reliability of the system is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\lambda }_{1}}t}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot {{e}^{-{{\lambda }_{1}}t}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
At 150 hours of operation, the reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{e}^{-(0.0002+0.0005+0.0001)150}} \\ &lt;br /&gt;
= &amp;amp; 0.8869\text{ or }88.69%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.12 is taken with respect to time, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; -\frac{d[{{R}_{s}}(t)]}{dt} \\ &lt;br /&gt;
= &amp;amp; -\frac{d\left[ {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}} \right]}{dt} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
:•	The system&#039;s failure rate can now be obtained simply by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn. 13 by the system&#039;s reliability function given in Eqn.12, and:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}\left( t \right)= &amp;amp; \frac{{{f}_{s}}\left( t \right)}{{{R}_{s}}\left( t \right)} \\ &lt;br /&gt;
= &amp;amp; \frac{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}}{{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}) \\ &lt;br /&gt;
= &amp;amp; 0.0008\text{ }fr/hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Combining Eqn.10 and Eqn. 12, the system&#039;s MTTF can be obtained:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }{{R}_{s}}\left( t \right)dt \\ &lt;br /&gt;
= &amp;amp; \int_{0}^{\infty }{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}dt \\ &lt;br /&gt;
= &amp;amp; \frac{1}{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})} \\ &lt;br /&gt;
= &amp;amp; 1250\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Solving Eqn. 12 with respect to time will yield the corresponding warranty period for a 90% reliability.  In this case, the system reliability equation is simple and a closed form solution exists.  The warranty time can now be found by solving:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
t= &amp;amp; -\frac{\ln (R)}{{{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}} \\ &lt;br /&gt;
= &amp;amp; -\frac{\ln (0.9)}{0.0008} \\ &lt;br /&gt;
= &amp;amp; 131.7\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the warranty period should be 132 hours.&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
Consider the system shown in Figure 5.5.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.5.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.5 Complex bridge system in Example 2. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Components  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  through  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  are Weibull distributed with  &amp;lt;math&amp;gt;\beta =1.2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1230&amp;lt;/math&amp;gt;  hours.  The starting and ending blocks cannot fail.  &amp;lt;br&amp;gt;&lt;br /&gt;
Determine the following:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability equation for the system and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s failure rate equation and the corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The warranty time for a 90% reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability for a 200-hour mission, if it is known that the system has already successfully operated for 200 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The first step is to obtain the reliability function for the system.  The methods described in the previous chapter can be employed, such as the event space or path-tracing methods.  Using BlockSim, the following reliability equation is obtained:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; ({{R}_{Start}}\cdot {{R}_{End}}(2{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}))  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that since the starting and ending blocks cannot fail,  &amp;lt;math&amp;gt;{{R}_{Start}}=1&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{End}}=1,&amp;lt;/math&amp;gt;  Eqn.14 can be reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; 2\cdot {{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  is the reliability equation for Component A, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{B}}&amp;lt;/math&amp;gt;  is the reliability equation for Component  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; , etc.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the components in this example are identical, the system reliability equation can be further reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2R{{(t)}^{2}}+2R{{(t)}^{3}}-5R{{(t)}^{4}}+2R{{(t)}^{5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Or, in terms of the failure distribution:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.6.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.6.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.6 Reliability plot for the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.18 is taken with respect to time, resulting in: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; 4\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
&amp;amp; -20\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  can now be plotted for different time values,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , as shown in Figure 5.7.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system&#039;s failure rate can now be obtained by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn.19 by the system&#039;s reliability function given in Eqn.18, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}(t)= &amp;amp; \frac{4\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}} \\ &lt;br /&gt;
&amp;amp; +\frac{-20\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.8.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.7.gif|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.7 &#039;&#039;pdf&#039;&#039; plot for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of the system is obtained by integrating the system&#039;s reliability function given by Eqn. 18 from time zero to infinity, as given by Eqn. 10.  Using BlockSim&#039;s Analytical QCP, an  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of 1007.8 hours is calculated, as shown in Figure 5.9.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The warranty time can be obtained by solving Eqn. 18 with respect to time for a system reliability  &amp;lt;math&amp;gt;{{R}_{s}}=0.9&amp;lt;/math&amp;gt; .  Using the Analytical QCP and selecting the &amp;lt;br&amp;gt;&lt;br /&gt;
Warranty Time option, a time of 372.72 hours is obtained, as shown in Figure 5.10.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.8.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.8 Failure rate for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.9.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.9 MTTF of the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.10.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.10 Time at which &#039;&#039;R&#039;&#039;=0.9 or 90% for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.11.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.11 Conditional reliability calculation for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, the conditional reliability can be obtained using Eqn.6 and Eqn.18, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(200,200)= &amp;amp; \frac{R(400)}{R(200)} \\ &lt;br /&gt;
= &amp;amp; \frac{0.883825}{0.975321} \\ &lt;br /&gt;
= &amp;amp; 0.906189  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated using BlockSim&#039;s Analytical QCP, as shown in Figure 5.11.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Approximating the System CDF=&lt;br /&gt;
&lt;br /&gt;
In many cases, it is valuable to fit a distribution that represents the system&#039;s times-to-failure.  This can be useful when the system is part of a larger assembly and may be used for repeated calculations or in calculations for other systems.  In cases such as this, it can be useful to characterize the system&#039;s behavior by fitting a distribution to the overall system and calculating parameters for this distribution.   This is equivalent to fitting a single distribution to describe  &amp;lt;math&amp;gt;{{R}_{S}}(t&amp;lt;/math&amp;gt; ).  In essence, it is like reducing the entire system to a component in order to simplify calculations.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
To compute an approximate reliability function for this system,  &amp;lt;math&amp;gt;{{R}_{A}}(t)\simeq {{R}_{S}}(t)&amp;lt;/math&amp;gt; , one would compute  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  pairs of reliability and time values and then fit a single distribution to the data, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 10,396.7)=10% \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 9,361.9)=20% \\ &lt;br /&gt;
&amp;amp; ... \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 1,053.6)=90%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A single distribution,  &amp;lt;math&amp;gt;{{R}_{A}}(t)&amp;lt;/math&amp;gt; , that approximates  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  can now be computed from these pairs using life data analysis methods.  If using the Weibull++ software, one would enter the values as free form data.&lt;br /&gt;
&lt;br /&gt;
===Example 3===&lt;br /&gt;
Compute a single Weibull distribution approximation for the system in Example 2.&lt;br /&gt;
====Solution to Example 3====&lt;br /&gt;
The system in the previous example, shown in Figure 5.5, can be approximated by use of a 2-parameter Weibull distribution with  &amp;lt;math&amp;gt;\beta =2.02109&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1123.51&amp;lt;/math&amp;gt; .  In BlockSim, this is accomplished by representing the entire system as one distribution by going to the Distribution Fit window Figure 5.12. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.13.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.12 Representing a system with a distribution.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.12.PNG|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.13 Distribution Fitting window.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
by clicking the Distribution Fit Window, the Distribution Estimator window will appear (Figure 5.13).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this window you can select a distribution to represent the data. BlockSim will then generate a number of system failure times based on the system&#039;s reliability function. The system&#039;s reliability function can be used to solve for a time value associated with that unreliability value. The distribution of the generated time values can then be fitted to a probability distribution function.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a value of  &amp;lt;math&amp;gt;F(t)=0.11&amp;lt;/math&amp;gt; .  Using the system&#039;s reliability equation and solving for time, the corresponding time-to-failure for a 0.11 unreliability can be calculated.  &amp;lt;br&amp;gt;&lt;br /&gt;
For the system of Example 2, the time for a 0.11 unreliability is 389.786 hours.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
When enough points have been generated, the selected distribution will be fitted to this data set and the distribution&#039;s parameters will be returned.  In addition, if ReliaSoft&#039;s Weibull++ is installed, the generated data can be viewed/analyzed using a Weibull++ instance, as shown in Figure 5.14.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.14.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.14 Using Weibull++ to calculate distribution parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
It is recommended that the analyst examine the fit to ascertain the applicability of the approximation.&lt;br /&gt;
&lt;br /&gt;
=Duty Cycle=&lt;br /&gt;
&lt;br /&gt;
Components of a system may not operate continuously during a system&#039;s mission, or may be subjected to loads greater or lesser than the rated loads during system operation.  To model this, a factor called the Duty Cycle ( &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt; ) is used.  The duty cycle may also be used to account for changes in environmental stress, such as temperature changes, that may effect the operation of a component.  The duty cycle is a positive value, with a default value of 1 representing continuous operation at rated load, and any values other than 1 representing other load values with respect to the rated load value (or total operating time).   A duty cycle value higher than 1 indicates a load in excess of the rated value.  A duty cycle value lower than 1 indicates that the component is operating at a load lower than the rated load or not operating continuously during the system&#039;s mission.  For instance, a duty cycle of 0.5 may be used for a component that operates only half of the time during the system&#039;s mission.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The reliability metrics for a component with a duty cycle are calculated as follows. Let  &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt;  represent the duty cycle during a particular mission of the component,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  represent the mission time and  &amp;lt;math&amp;gt;{t}&#039;&amp;lt;/math&amp;gt;  represent the accumulated age. Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{t}&#039;={{d}_{c}}\times t&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability equation for the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R({t}&#039;)=R({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The component &#039;&#039;pdf&#039;&#039; is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;f({t}&#039;)=-\frac{d(R({t}&#039;))}{dt}=-\frac{d(R({{d}_{c}}\times t))}{dt}={{d}_{c}}f({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The failure rate of the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\lambda ({t}&#039;)=\frac{f({t}&#039;)}{R({t}&#039;)}=\frac{{{d}_{c}}f({{d}_{c}}\times t)}{R({{d}_{c}}\times t)}={{d}_{c}}\lambda ({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Example 4===&lt;br /&gt;
Consider a computer system with three components: a processor, a hard drive and a CD drive in series as shown next.  Assume that all three components follow a Weibull failure distribution with the parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}=1.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=5000&amp;lt;/math&amp;gt;  for the processor,  &amp;lt;math&amp;gt;{{\beta }_{2}}=2.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{2}}=3000&amp;lt;/math&amp;gt;  for the hard drive, and  &amp;lt;math&amp;gt;{{\beta }_{3}}=2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=4000&amp;lt;/math&amp;gt;  for the CD drive.  Determine the reliability of the computer system after one year (365 days) of operation, assuming that the CD drive is used only 30% of the time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5ex4.png|thumb|center|300px|]]&lt;br /&gt;
&lt;br /&gt;
====Solution to Example 4====&lt;br /&gt;
The reliability of the processor after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  {{R}_{processor}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{5000} \right)}^{1.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\text{ or }98.05%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the hard drive after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{harddrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{3000} \right)}^{2.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9948\text{ or }99.48%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the CD drive after 365 days of operation (taking into account the 30% operation using a duty cycle of 0.3) is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{CDdrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{{{d}_{c}}\times 365}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{0.3\times 365}{4000} \right)}^{2}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9993\text{ or }99.93%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.15.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.15 Result for the computer system reliability.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
Thus the reliability of the computer system after 365 days of operation is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{s}}(365)= &amp;amp; {{R}_{processor}}(365)\cdot {{R}_{harddrive}}(365)\cdot {{R}_{CDdrive}}(365) \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\cdot 0.9948\cdot 0.9993 \\ &lt;br /&gt;
	  = &amp;amp; 0.9747\text{ or }97.47%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;	&lt;br /&gt;
&amp;lt;br&amp;gt;	&lt;br /&gt;
This result can be obtained in BlockSim as shown in Figure DutyCycleExResults.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Load Sharing=&lt;br /&gt;
As presented in earlier chapters, a reliability block diagram (RBD) allows you to graphically represent how the components within a system are reliability-wise connected.  In most cases, independence is assumed across the components within the system.  For example, the failure of component A does not affect the failure of component B.  However, if a system consists of components that are sharing a load, then the assumption of independence no longer holds true.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If one component fails, then the component(s) that are still operating will have to assume the failed unit&#039;s portion of the load.  Therefore, the reliabilities of the surviving unit(s) will change.  Calculating the system reliability is no longer an easy proposition.  In the case of load sharing components, the change of the failure distributions of the surviving components must be known in order to determine the system&#039;s reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate this, consider the a system of two units connected reliability-wise in parallel (Figure 5.16).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-16.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.16 Two units connected reliability-wise in parallel.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Assume that the units must supply an output of 8 volts and that if both units are operational, each unit is to supply 50% of the total output.  If one of the units fails, then the surviving unit supplies 100%.  Furthermore, assume that having to supply the entire load has a negative impact on the reliability characteristics of the surviving unit.  Since the reliability characteristics of the unit change based on whether both or only one is operating, a life distribution along with a life-stress relationship (as discussed in Chapter 3) will be needed to model each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the steps needed, we will create the model starting from raw data.  Assume that a total of 20 units were tested to failure at 7, 10 and 15 volts.  The test data set is presented in the next table.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-17.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
For this example, Units 1 and 2 are the same component.  Therefore, only one set of data was collected.  However, it is possible that the load sharing components in a system may not be the same.  If that were the case, data would need to be collected for each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The data set in Table 1 was analyzed using ReliaSoft&#039;s ALTA software (as shown in Figure 5.17) with the Inverse Power Law as the underlying life-stress relationship and Weibull as the life distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated model parameters,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , are shown next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
\beta = &amp;amp; 1.9239 \\ &lt;br /&gt;
K= &amp;amp; 3.2387\times {{10}^{-7}} \\ &lt;br /&gt;
n= &amp;amp; 3.4226  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:Or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( 3.2387\times {{10}^{-7}}S_{1}^{3.4226}t \right)}^{1.9239}}}}  \ (eqn 20)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1}}(t,{{S}_{1}})=\beta KS_{1}^{n}{{\left( KS_{1}^{n}t \right)}^{\beta -1}}{{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}}  \ (eqn 21 )&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And for this case:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{R}_{2}}(t,{{S}_{2}})  \\ &lt;br /&gt;
{{f}_{1}}(t,{{S}_{1}})= &amp;amp; {{f}_{2}}(t,{{S}_{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.17.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.17 Using ALTA to calculate component parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 5.18 shows a plot of Eqn.20.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Now that the failure properties have been determined using the test data, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t,S)= &amp;amp; {{R}_{1}}(t,{{S}_{1}})\cdot {{R}_{2}}(t,{{S}_{2}}) \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{1}}\left( x,{{S}_{1}} \right)\cdot {{R}_{2}}(x,{{S}_{2}})\cdot \left( \frac{{{R}_{2}}({{t}_{1e}}+(t-x),S)}{{{R}_{2}}({{t}_{1e}},S)} \right)dx \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{2}}\left( x,{{S}_{2}} \right)\cdot {{R}_{1}}(x,{{S}_{1}})\cdot \left( \frac{{{R}_{1}}({{t}_{2e}}+(t-x),S)}{{{R}_{1}}({{t}_{2e}},S)} \right)dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:Where: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{S}_{1}}= &amp;amp; {{P}_{1}}S \\ &lt;br /&gt;
{{S}_{2}}= &amp;amp; {{P}_{2}}S  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-18.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig. 5.18 Reliability curves for different voltage output conditions. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.png|thumb|center|395px]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  is the total load (or required output).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{P}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{P}_{2}}&amp;lt;/math&amp;gt;  are the portion of the total load that each unit supports when both units are operational.  In this case,  &amp;lt;math&amp;gt;{{P}_{1}}={{P}_{2}}=0.5=50%.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{S}_{2}}&amp;lt;/math&amp;gt;  represent the portions of the load that Unit 1 and Unit 2 must support when both units are operational.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{{{1}_{e}}}}&amp;lt;/math&amp;gt;  is the equivalent operating time for Unit 1 if it had been operating at  &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  instead of  &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt; .  A graphical representation of the equivalent time is shown in Figure 5.19, where the curve marked by L represents the low stress (load) and the curve marked by H represents the high stress (load).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  can be calculated by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t)= &amp;amp; {{R}_{1}}({{t}_{1e}}) \\ &lt;br /&gt;
{{e}^{-{{(tKS_{1}^{n})}^{\beta }}}}= &amp;amp; {{e}^{-{{({{t}_{1e}}K{{S}^{n}})}^{\beta }}}} \\ &lt;br /&gt;
tS_{1}^{n}= &amp;amp; {{t}_{1e}}{{S}^{n}} \\ &lt;br /&gt;
{{t}_{1e}}= &amp;amp; t{{\left( \frac{{{S}_{1}}}{S} \right)}^{n}},\text{     }{{S}_{1}}={{P}_{1}}S \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{1e}}=tP_{1}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  can be calculated the same way, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{2}}(t)= &amp;amp; {{R}_{2}}({{t}_{2e}}) \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{2e}}=tP_{2}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this example, the reliability equations for Unit 1 and Unit 2 are the same since they are the same type of component and demonstrate the same failure properties.  In addition, the total output is divided equally between the two units (when both units are operating), so  &amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  will also be the same.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The next step is to determine the reliability of the system after 8,760 hours,  &amp;lt;math&amp;gt;R(t=8,760)&amp;lt;/math&amp;gt; .  Using Eqn. 22 the system reliability is found to be:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t=8760)= &amp;amp; 0.8567 \\ &lt;br /&gt;
= &amp;amp; 85.67%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Load Sharing in BlockSim===&lt;br /&gt;
BlockSim uses this formulation when computing reliabilities of units in a load sharing configuration.  When using the System Reliability Equation window, BlockSim returns a single token for the reliability of units in a load sharing configuration (as well as in the case of standby redundancy, discussed in the next section).  As an example, consider the following RBD with Unit 1 in series with a container that includes two load sharing units.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.2.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
BlockSim will return the system equation as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{System}}=+{{R}_{LS}}\cdot {{R}_{1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{LS}}&amp;lt;/math&amp;gt;  implies a form similar to Eqn. 22.  BlockSim allows for  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -out-of- &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  units in a load sharing configuration.&lt;br /&gt;
&lt;br /&gt;
===Example 5===&lt;br /&gt;
A component has five possible failure modes,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , and the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are interdependent.  The system will fail if mode  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  occurs, mode  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  occurs or two out of the three  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes occur.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  each have a Weibull distribution, with a  &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt;  and 15,000 respectively.  Events  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  each have an exponential distribution with a mean of 10,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If any  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  event occurs (i.e.  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt; ), the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are more likely to occur.  Specifically, the mean times of the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are halved.  Determine the reliability at 1000 hours for this component.&lt;br /&gt;
====Solution to Example 5====&lt;br /&gt;
The first step is to create the RBD.  Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and a load sharing container with the  &amp;lt;math&amp;gt;{{B}_{i}}&amp;lt;/math&amp;gt;  modes must be drawn in series, as illustrated next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.3.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The next step is to define the properties for each block, including those for the container.  Setting the failure distributions for modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is simple.  The more difficult part is setting the properties for the container and the contained blocks.  Based on the problem statement, the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are in a 2-out-of-3 load sharing redundancy.  When all three are working (i.e. when no  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode has occurred), each block has an exponential distribution with &amp;lt;math&amp;gt;\mu=10,000&amp;lt;/math&amp;gt;.  If one  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode occurs, then the two surviving units have an exponential distribution with  &amp;lt;math&amp;gt;\mu =5,000.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assume a Power Life-Stress relationship for the components.  Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\mu }_{1}}= &amp;amp; \frac{1}{KV_{1}^{n}}  \ (eqn 23)\\ &lt;br /&gt;
{{\mu }_{2}}= &amp;amp; \frac{1}{KV_{2}^{n}}  \ (eqn 24)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{1}}=10,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{V}_{1}}=1&amp;lt;/math&amp;gt;  in Eqn. 23 and casting it in terms of  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
10,000= &amp;amp; \frac{1}{K}  \ (eqn 25) \\ &lt;br /&gt;
K = &amp;amp; \frac{1}{10,000}=0.0001  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{2}}=5,000&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{V}_{2}}=1.5&amp;lt;/math&amp;gt;  (because if one fails, then each survivor takes on an additional 0.5 units of load) and Eqn. 25 for  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  in Eqn.24 yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
5,000= &amp;amp; \frac{1}{0.0001\cdot {{(1.5)}^{n}}} \\ &lt;br /&gt;
0.5= &amp;amp; {{(1.5)}^{-n}} \\ &lt;br /&gt;
\ln (0.5)= &amp;amp; -n\ln (1.5) \\ &lt;br /&gt;
n = &amp;amp; 1.7095  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This also could have been computed in ALTA, as shown in Figure 5.20, or with the Load &amp;amp; Life Parameter Experimenter in BlockSim, as shown in Figure 5.21.&lt;br /&gt;
 &lt;br /&gt;
At this point, the parameters for the load sharing units have been computed and can be set, as shown in Figure 5.22.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The next step is to set the weight proportionality factor.  This factor defines the portion of the load that the particular item carries while operating, as well as the load that shifts to the remaining units upon failure of the item.  To illustrate, assume three units (1, 2 and 3) are in a load sharing container with weight proportionality factors of 1, 2 and 3 respectively (and a 1-out-of-3 requirement).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Unit 1 carries  &amp;lt;math&amp;gt;\left( \tfrac{1}{1+2+3} \right)=0.166&amp;lt;/math&amp;gt;  or 16.6% of the total load.&lt;br /&gt;
:•	Unit 2 carries  &amp;lt;math&amp;gt;\left( \tfrac{2}{1+2+3} \right)=0.333&amp;lt;/math&amp;gt;  or 33.3% of the total load.&lt;br /&gt;
:•	Unit 3 carries  &amp;lt;math&amp;gt;\left( \tfrac{3}{1+2+3} \right)=0.50&amp;lt;/math&amp;gt;  or 50% of the total load.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The actual load on each unit then becomes the product of the entire load defined for the container times the portion carried by that unit.  For example, if the container load is 100 lbs, then the portion assigned to Unit 1 will be  &amp;lt;math&amp;gt;100\cdot 0.166=16.6&amp;lt;/math&amp;gt;  lbs.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the current example, all units share the same load and thus have equal weight proportionality factors.  Because these factors are relative, if the same number is used for all three items then the results will be the same.  Thus, weight proportional factor is set equal to 1 for each item.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5.20.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.20 Calculation performed in ALTA.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.21_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.21 Quick Parameter Estimator &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.21_3.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.22 Quick Parameter Estimator results &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.23_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.23 Defining Weight Proportional Factor. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The last properties that need to be defined are the total load and the 2-out-of-3 redundancy.  The total load is dependent on how the parameters were computed.  In this case, total load was assumed to be 3 when the parameters were computed (i.e. the load per item was 1 when all worked and 1.5 when two worked).  This is defined at the container level, set No. of Paths required = 3.&lt;br /&gt;
When all of the parameters have been specified in BlockSim, the reliability at 1,000 hours can be determined.  From the Analytical QCP, this is found to be 98.57%.&lt;br /&gt;
&lt;br /&gt;
=Standby Components=&lt;br /&gt;
&lt;br /&gt;
In the previous section, the case of a system with load sharing components was presented.  This is a form of redundancy with dependent components. That is, the failure of one component affects the failure of the other(s).  This section presents another form of redundancy: standby redundancy.  In standby redundancy the redundant components are set to be under a lighter load condition (or no load) while not needed and under the operating load when they are activated.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In standby redundancy the components are set to have two states: an active state and a standby state.  Components in standby redundancy have two failure distributions, one for each state.  When in the standby state, they have a quiescent (or dormant) failure distribution and when operating, they have an active failure distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case that both quiescent and active failure distributions are the same, the units are in a simple parallel configuration (also called a hot standby configuration).  When the rate of failure of the standby component is lower in quiescent mode than in active mode, that is called a warm standby configuration.  When the rate of failure of the standby component is zero in quiescent mode (i.e. the component cannot fail when in standby), that is called a cold standby configuration.  &lt;br /&gt;
&lt;br /&gt;
===Simple Standby Configuration===&lt;br /&gt;
&lt;br /&gt;
Consider two components in a standby configuration.  Component 1 is the active component with a Weibull failure distribution with parameters  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Component 2 is the standby component.  When Component 2 is operating, it also has a Weibull failure distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Furthermore, assume the following cases for the quiescent distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1:  The quiescent distribution is the same as the active distribution (hot standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2:  The quiescent distribution is a Weibull  distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 2000 (warm standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3: The component cannot fail in quiescent mode (cold standby).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this case, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be obtained using the following equation:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(t)={{R}_{1}}(t)+\underset{0}{\overset{t}{\mathop \int }}\,{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x)\cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}dx   \ (eqn 26)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode, such that: &amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
&amp;lt;br&amp;gt;  &lt;br /&gt;
&amp;lt;math&amp;gt;{{R}_{2;SB}}(x)={{R}_{2;A}}({{t}_{e}})  \ (eqn 27)&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Eqn. 27 can be solved for  &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  and substituted into Eqn.26.&lt;br /&gt;
Figure 5.24 illustrates the example as entered in BlockSim using a standby container.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5.24.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.24 Standby container.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The active and standby blocks are within a container, which is used to specify standby redundancy.  Since the standby component has two distributions (active and quiescent), the Block Properties window of the standby block has two pages for specifying each one.  Figures 5.24 and 5.26 illustrate these pages.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system reliability results for 1000 hours are given in the following table:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-24.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Note that even though the  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  value for the quiescent distribution is the same as in the active distribution, it is possible that the two can be different. That is, the failure modes present during the quiescent mode could be different from the modes present during the active mode.  In that sense, the two distribution types can be different as well (e.g. lognormal when quiescent and Weibull when active).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In many cases when considering standby systems, a switching device may also be present that switches from the failed active component to the standby component.  The reliability of the switch can also be incorporated into Eqn.26, as presented in the next section.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim&#039;s System Reliability Equation window returns a single token for the reliability of units in a standby configuration.  This is the same as the load sharing case presented in the previous section. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image: Fig 5.25.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.25 Defining the active failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.26.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.26 Defining the quiescent failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Reliability of Standby Systems with a Switching Device===&lt;br /&gt;
&lt;br /&gt;
In many cases when dealing with standby systems, a switching device is present that will switch to the standby component when the active component fails.  Therefore, the failure properties of the switch must also be included in the analysis.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.26.2.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In most cases when the reliability of a switch is to be included in the analysis, two probabilities can be considered.  The first and most common one is the probability of the switch performing the action (i.e. switching) when requested to do so.  This is called Switch Probability per Request in BlockSim and is expressed as a static probability (e.g. 90%).  The second probability is the quiescent reliability of the switch.  This is the reliability of the switch as it ages (e.g. the switch might wear out with age due to corrosion, material degradation, etc.). Thus it is possible for the switch to fail before the active component fails.  However, a switch failure does not cause the system to fail, but rather causes the system to fail only if the switch is needed and the switch has failed.  For example, if the active component does not fail until the mission end time and the switch fails, then the system does not fail.  However, if the active component fails and the switch has also failed, then the system cannot be switched to the standby component and it therefore fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In analyzing standby components with a switching device, either or both failure probabilities (during the switching or while waiting to switch) can be considered for the switch, since each probability can represent different failure modes.  For example, the switch probability per request may represent software-related issues or the probability of detecting the failure of an active component, and the quiescent probability may represent wear-out type failures of the switch.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the formulation, consider the previous example that assumes perfect switching.  To examine the effects of including an imperfect switch, assume that when the active component fails there is a 90% probability that the switch will switch from the active component to the standby component.  In addition, assume that the switch can also fail due to a wear-out failure mode described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.7 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 5000.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , is given by the following equation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t)= &amp;amp; {{R}_{1}}(t) \\ &lt;br /&gt;
&amp;amp; +\underset{0}{\overset{t}{\mathop \int }}\,\{{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x) \\ &lt;br /&gt;
&amp;amp; \cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}\cdot {{R}_{SW;Q}}(x)\cdot {{R}_{SW;REQ}}(x)\}dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;Q}}&amp;lt;/math&amp;gt;  is the quiescent reliability of the switch.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;REQ}}&amp;lt;/math&amp;gt;  is the switch probability per request.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
This problem can be solved in BlockSim by including these probabilities in the container&#039;s properties, as shown in Figures 5.27 and 5.28.  In BlockSim, the standby container is acting as the switch.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.28_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.27 Standby container (switch) failure probabilities while attempting to switch &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.27.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.28 Standby container (switch) failure distribution while waiting to switch &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that there are additional properties that can be specified in BlockSim for a switch, such as Switch Restart Probability, No. of Restarts and Switch Delay Time.  In many applications, the switch is re-tested (or re-cycled) if it fails to switch the first time.  In these cases, it might be possible that it switches in the second or third, or  &amp;lt;math&amp;gt;{{n}^{th}}&amp;lt;/math&amp;gt; attempt.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The Switch Restart Probability specifies each additional attempt&#039;s probability of successfully switching and the Finite Restarts specifies the total number of attempts.  Note that the Switch Restart Probability specifies the probability of success of each trial (or attempt).  The probability of success of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  consecutive trials is calculated by BlockSim using the binomial distribution and this probability is then incorporated into Eqn.28.  The Switch Delay Time property is related to repairable systems and is considered in BlockSim only when using simulation.  When using the analytical solution (i.e. for a non-repairable system), this property is ignored.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving the analytical solution (as given by Eqn. 28), the following results are obtained.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-30.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From the table above, it can be seen that the presence of a switching device has a significant effect on the reliability of a standby system.  It is therefore important when modeling standby redundancy to incorporate the switching device reliability properties.  It should be noted that this methodology is not the same as treating the switching device as another series component with the standby subsystem.  This would be valid only if the failure of the switch resulted in the failure of system (e.g. switch failing open).  In Eqn.28, the Switch Probability per Request and quiescent probability are present only in the second term of the equation.  Treating these two failure modes as a series configuration with the standby subsystem would imply that they are also present when the active component is functioning (i.e. first term of Eqn.28).  This is invalid and would result in the underestimation of the reliability of the system.  In other words, these two failure modes become significant only when the active component fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As an example, and if we consider the warm standby case, the reliability of the system without the switch is 70.57% at 1000 hours.  If the system was modeled so that the switching device was in series with the warm standby subsystem, the result would have been:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(1000)= &amp;amp; {{R}_{Standby}}(1000)\cdot {{R}_{sw,Q(1000)}}\cdot {{R}_{sw,req}} \\ &lt;br /&gt;
= &amp;amp; 0.7057\cdot 0.9372\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.5952  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case where a switch failure mode causes the standby subsystem to fail, then this mode can be modeled as an individual block in series with the standby subsystem.&lt;br /&gt;
&lt;br /&gt;
===Example 6===&lt;br /&gt;
Consider a car with four new tires and a full-size spare.  Assume the following failure characteristics:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires follow a Weibull distribution with a &amp;lt;math&amp;gt;\beta=4&amp;lt;/math&amp;gt; and an  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  40,000 miles while on the car due to wear.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires also have a probability of failing due to puncture or other causes.  For this, assume a constant rate for this occurrence with a probability of 1 every 50,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	When not on the car (i.e. is a spare), a tire&#039;s probability of failing also has a Weibull distribution with a  &amp;lt;math&amp;gt;\beta =&amp;lt;/math&amp;gt;  2 and  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  120,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Assume a mission of 1,000 miles.  If a tire fails during this trip, it will be replaced with the spare.  However, the spare will not be repaired during the trip.  In other words, the trip will continue with the spare on the car and if the spare fails the system will fail.  Determine the probability of system failure.&lt;br /&gt;
====Solution to Example 6====&lt;br /&gt;
Active failure distribution for tires:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to wear-out, Weibull  &amp;lt;math&amp;gt;\beta =4&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =40,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to random puncture, exponential  &amp;lt;math&amp;gt;\mu =50,000.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	The quiescent failure distribution is a Weibull distribution with &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =120,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The block diagram for each tire has two blocks in series, one block representing the wear-out mode and the other the random puncture mode, as shown next:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:small5.gif|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
There are five tires, four active and one standby (represented in the diagram by a standby container with a 4-out-of-5 requirement), as shown next: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BStirecontainer.png|thumb|center|400px|]]&lt;br /&gt;
 &lt;br /&gt;
For the standby Wear block, set the active failure and the quiescent distributions, but for the Puncture block, set only the active puncture distribution (because the tire cannot fail due to puncture while stored).  Using BlockSim, the probability of system failure is found to be 0.003 or 0.3%. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Note Regarding Numerical Integration Solutions=&lt;br /&gt;
&lt;br /&gt;
Load sharing and standby solutions in BlockSim are performed using numerical integration routines.  As with any numerical analysis routine, the solution error depends on the number of iterations performed, the step size chosen and related factors, plus the behavior of the underlying function.  By default, BlockSim uses a certain set of preset factors.  In general, these defaults are sufficient for most problems.  If a higher precision or verification of the precision for a specific problem is required, BlockSim&#039;s preset options can be modified and/or the integration error can be assessed using the Integration Parameters... option for each container.  For more details, you can refer to the documentation on the Algorithm Setup window in the BlockSim 7 User&#039;s Guide.&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15580</id>
		<title>Time-Dependent System Reliability (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15580"/>
		<updated>2012-02-13T22:30:12Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Reliability of Standby Systems with a Switching Device */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|5}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the previous chapter, different system configuration types were examined, as well as different methods for obtaining the system&#039;s reliability function analytically.  Because the reliabilities in the problems presented were treated as probabilities (e.g.  &amp;lt;math&amp;gt;P(A)&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; ), the reliability values and equations presented were referred to as static (not time-dependent).  Thus, in the prior chapter, the life distributions of the components were not incorporated in the process of calculating the system reliability.  In this chapter, time dependency in the reliability function will be introduced.  We will develop the models necessary to observe the reliability over the life of the system, instead of at just one point in time.  In addition, performance measures such as failure rate, MTTF and warranty time will be estimated for the entire system.  The methods of obtaining the reliability function analytically remain identical to the ones presented in the previous chapter, with the exception that the reliabilities will be functions of time.  In other words, instead of dealing with  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , we will use  &amp;lt;math&amp;gt;{{R}_{i}}(t)&amp;lt;/math&amp;gt; .  All examples in this chapter assume that no repairs are performed on the components.  &lt;br /&gt;
&lt;br /&gt;
=Analytical Life Predictions=&lt;br /&gt;
The analytical approach presented in the prior chapter involved the determination of a mathematical expression that describes the reliability of the system, expressed in terms of the reliabilities of its components.  So far we have estimated only static system reliability (at a fixed time).  For example, in the case of a system with three components in series, the system&#039;s reliability equation was given by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}  \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The values of  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  were given for a common time and the reliability of the system was estimated for that time.  However, since the component failure characteristics can be described by distributions, the system reliability is actually time-dependent.  In this case, Eqn. (1) can be rewritten as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability of the system for any mission time can now be estimated.  Assuming a Weibull life distribution for each component, Eqn. (1) can now be expressed in terms of each component&#039;s reliability function, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the same manner, any life distribution can be substituted into the system reliability equation.  Suppose that the times-to-failure of the first component are described with a Weibull distribution, the times-to-failure of the second component with an exponential distribution and the times-to-failure of the third component with a normal distribution.  Then Eqn. (1) can be written as:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot \left[ 1-\Phi \left( \frac{t-{{\mu }_{3}}}{{{\sigma }_{3}}} \right) \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It can be seen that the biggest challenge is in obtaining the system&#039;s reliability function in terms of component reliabilities, which has already been discussed in depth.  Once this has been achieved, calculating the reliability of the system for any mission duration is just a matter of substituting the corresponding component reliability functions into the system reliability equation.&lt;br /&gt;
===Advantages of the Analytical Method===&lt;br /&gt;
The primary advantage of the analytical solution is that it produces a mathematical expression that describes the reliability of the system.  Once the system&#039;s reliability function has been determined, other calculations can then be performed to obtain metrics of interest for the system. Such calculations include:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of warranty periods.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s failure rate.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s MTTF.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, optimization and reliability allocation techniques can be used to aid engineers in their design improvement efforts.  Another advantage in using analytical techniques is the ability to perform static calculations and analyze systems with a mixture of static and time-dependent components.  Finally, the reliability importance of components over time can be calculated with this methodology.&lt;br /&gt;
===Disadvantages of the Analytical Method===&lt;br /&gt;
The biggest disadvantage of the analytical method is that formulations can become very complicated.  The more complicated a system is, the larger and more difficult it will be to analytically formulate an expression for the system&#039;s reliability.  For particularly detailed systems this process can be quite time-consuming, even with the use of computers.  Furthermore, when the maintainability of the system or some of its components must be taken into consideration, analytical solutions become intractable.  In these situations, the use of simulation methods may be more advantageous than attempting to develop a solution analytically.  Simulation methods are presented in later chapters.&lt;br /&gt;
===Looking at a Simple &#039;&#039;Complex&#039;&#039; System Analytically===&lt;br /&gt;
&lt;br /&gt;
The complexity involved in an analytical solution can be best illustrated by looking at the simple &#039;&#039;complex&#039;&#039; system with 15 components, as shown in Figure 5.1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1.png|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.1 An RBD of a complex system.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system reliability for this system (computed using BlockSim) is shown next.  The first solution is provided using BlockSim&#039;s symbolic solution.  In symbolic mode, BlockSim breaks the equation into segments, identified by tokens, that need to be substituted into the final system equation for a complete solution.  This creates algebraic solutions that are more compact than if the substitutions were made.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; D2\cdot D3\cdot {{R}_{L}} \\ &lt;br /&gt;
D3= &amp;amp; +{{R}_{K}}\cdot IK \\ &lt;br /&gt;
IK= &amp;amp; +{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}} \\ &lt;br /&gt;
D2 = &amp;amp; +{{R}_{A}}\cdot {{R}_{E}}\cdot IE \\ &lt;br /&gt;
IE = &amp;amp; -D1\cdot {{R}_{M}}\cdot {{R}_{N}}+{{R}_{M}}\cdot {{R}_{N}}+D1 \\ &lt;br /&gt;
D1 = &amp;amp; +{{R}_{D}}\cdot ID \\ &lt;br /&gt;
ID = &amp;amp; -{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting the terms yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; {{R}_{A}}\cdot {{R}_{E}}\cdot {{R}_{L}}\cdot {{R}_{K}} \\ &lt;br /&gt;
&amp;amp; \cdot \{({{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})\cdot {{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}}-{{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}\} \\ &lt;br /&gt;
&amp;amp; \cdot \{{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}}\}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
BlockSim&#039;s automatic algebraic simplification would yield the following format for the above solution: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; (({{R}_{A}}\cdot {{R}_{E}}(-({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})){{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})))) \\ &lt;br /&gt;
&amp;amp; ({{R}_{K}}({{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; +RI\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}})){{R}_{L}})  \ (eqn 2)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this equation, each  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  represents the reliability function of a block.  For example, if  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  has a Weibull distribution, then each  &amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;  and so forth.  Substitution of each component&#039;s reliability function in Eqn.2 will result in an analytical expression for the system reliability as a function of time, or  &amp;lt;math&amp;gt;{{R}_{s}}(t)&amp;lt;/math&amp;gt; , which is the same as  &amp;lt;math&amp;gt;(1-cd{{f}_{System}}).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Obtaining Other Functions of Interest===&lt;br /&gt;
Once the system reliability equation (or the cumulative density function,  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt; ) has been determined, other functions and metrics of interest can be derived.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Consider the following simple system:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-2.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.2 Simple two-component system. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Furthermore, assume that component 1 follows an exponential distribution with a mean of 10,000 (&amp;lt;math&amp;gt;\mu =10,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;\lambda =1/10,000)&amp;lt;/math&amp;gt;  and component 2 follows a Weibull distribution with  &amp;lt;math&amp;gt;\beta =6&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt; .  The reliability equation of this system is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}  \ (eqn 3) &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{F}_{S}}(t)= &amp;amp; 1-({{R}_{1}}(t)\cdot {{R}_{2}}(t)) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \right) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
{{analytical system pdf}}&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability====&lt;br /&gt;
Conditional reliability is the probability of a system successfully completing another mission following the successful completion of a previous mission.  The time of the previous mission and the time for the mission to be undertaken must be taken into account for conditional reliability calculations.  The system&#039;s conditional reliability function is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(T,t)=\frac{R(T+t)}{R(T)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Eqn.6 gives the reliability for a new mission of duration  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  having already accumulated  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  hours of operation up to the start of this new mission. The system is evaluated to assure that it will start the next mission successfully.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2, the reliability for mission of  &amp;lt;math&amp;gt;t=1,000&amp;lt;/math&amp;gt;  hours, having an age of  &amp;lt;math&amp;gt;T=500&amp;lt;/math&amp;gt;  hours, is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(T=500,t=1000)= &amp;amp; \frac{R(T+t)}{R(T)} \\ &lt;br /&gt;
= &amp;amp; \frac{R(1500)}{R(500)} \\ &lt;br /&gt;
= &amp;amp; \frac{{{e}^{-\tfrac{1500}{10,000}}}\cdot {{e}^{-{{\left( \tfrac{1500}{10,000} \right)}^{6}}}}}{{{e}^{-\tfrac{500}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{500}{10,000} \right)}^{6}}}}} \\ &lt;br /&gt;
= &amp;amp; 0.9048=90.48%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.3.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.4.png|thumb|center|400px|]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability for Components====&lt;br /&gt;
&lt;br /&gt;
Now in this formulation, it was assumed that the accumulated age was equivalent for both units. That is, both started life at zero and aged to 500.  It is possible to consider an individual component that has already accumulated some age (used component) in the same formulation.  To illustrate this, assume that component 2 started life with an age of T=100.  Then the reliability equation of the system, as given in Eqn.3, would need to be modified to include a conditional term for 2, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{R}_{1}}(t)\cdot \frac{{{R}_{2}}({{T}_{2}}+t)}{{{R}_{2}}({{T}_{2}})} \ (eqn 7)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In BlockSim, the start age input box may be used to specify a starting age greater than zero.&lt;br /&gt;
{{system failure rate analytical}}&lt;br /&gt;
&lt;br /&gt;
====System Mean Life (Mean Time To Failure)====&lt;br /&gt;
The mean life (or mean time to failure, MTTF) can be obtained by integrating the system reliability function from zero to infinity: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;MTTF=\int_{0}^{\infty }{{R}_{s}}\left( t \right)dt   \ (eqn 10)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mean time is a performance index and does not provide any information about the behavior of the failure distribution of the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)dt \\ &lt;br /&gt;
= &amp;amp; 5978.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Warranty Period and BX Life====&lt;br /&gt;
Sometimes it is desirable to know the time value associated with a certain reliability.  Warranty periods are often calculated by determining what percentage of the failure population can be covered financially and estimating the time at which this portion of the population will fail.  Similarly, engineering specifications may call for a certain BX life, which also represents a time period during which a certain proportion of the population will fail.  For example, the B10 life is the time in which 10% of the population will fail.  &lt;br /&gt;
This is obtained by setting  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  to the desired value and solving for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}\left( t \right)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To compute the time by which reliability would be equal to 90%, Eqn.11 is recast as follows and solved for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this case,  &amp;lt;math&amp;gt;t=1053.59&amp;lt;/math&amp;gt; .  Equivalently, the B10 life for this system is also  &amp;lt;math&amp;gt;1053.59&amp;lt;/math&amp;gt; .&lt;br /&gt;
Except for some trivial cases, a closed form solution for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  cannot be obtained.   Thus, it is necessary to solve for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  using numerical methods.  BlockSim uses numerical methods.&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
Consider a system consisting of three exponential units in series with the following failure rates (in failures per hour):  &amp;lt;math&amp;gt;{{\lambda }_{1}}&amp;lt;/math&amp;gt;  = 0.0002,  &amp;lt;math&amp;gt;{{\lambda }_{2}}&amp;lt;/math&amp;gt;  = 0.0005 and  &amp;lt;math&amp;gt;{{\lambda }_{3}}&amp;lt;/math&amp;gt;  = 0.0001.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the reliability equation for the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the reliability of the system after 150 hours of operation?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s failure rate equation.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the MTTF for the system?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What should the warranty period be for a 90% reliability?&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The analytical expression for the reliability of the system is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\lambda }_{1}}t}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot {{e}^{-{{\lambda }_{1}}t}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
At 150 hours of operation, the reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{e}^{-(0.0002+0.0005+0.0001)150}} \\ &lt;br /&gt;
= &amp;amp; 0.8869\text{ or }88.69%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.12 is taken with respect to time, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; -\frac{d[{{R}_{s}}(t)]}{dt} \\ &lt;br /&gt;
= &amp;amp; -\frac{d\left[ {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}} \right]}{dt} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
:•	The system&#039;s failure rate can now be obtained simply by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn. 13 by the system&#039;s reliability function given in Eqn.12, and:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}\left( t \right)= &amp;amp; \frac{{{f}_{s}}\left( t \right)}{{{R}_{s}}\left( t \right)} \\ &lt;br /&gt;
= &amp;amp; \frac{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}}{{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}) \\ &lt;br /&gt;
= &amp;amp; 0.0008\text{ }fr/hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Combining Eqn.10 and Eqn. 12, the system&#039;s MTTF can be obtained:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }{{R}_{s}}\left( t \right)dt \\ &lt;br /&gt;
= &amp;amp; \int_{0}^{\infty }{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}dt \\ &lt;br /&gt;
= &amp;amp; \frac{1}{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})} \\ &lt;br /&gt;
= &amp;amp; 1250\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Solving Eqn. 12 with respect to time will yield the corresponding warranty period for a 90% reliability.  In this case, the system reliability equation is simple and a closed form solution exists.  The warranty time can now be found by solving:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
t= &amp;amp; -\frac{\ln (R)}{{{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}} \\ &lt;br /&gt;
= &amp;amp; -\frac{\ln (0.9)}{0.0008} \\ &lt;br /&gt;
= &amp;amp; 131.7\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the warranty period should be 132 hours.&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
Consider the system shown in Figure 5.5.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.5.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.5 Complex bridge system in Example 2. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Components  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  through  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  are Weibull distributed with  &amp;lt;math&amp;gt;\beta =1.2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1230&amp;lt;/math&amp;gt;  hours.  The starting and ending blocks cannot fail.  &amp;lt;br&amp;gt;&lt;br /&gt;
Determine the following:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability equation for the system and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s failure rate equation and the corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The warranty time for a 90% reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability for a 200-hour mission, if it is known that the system has already successfully operated for 200 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The first step is to obtain the reliability function for the system.  The methods described in the previous chapter can be employed, such as the event space or path-tracing methods.  Using BlockSim, the following reliability equation is obtained:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; ({{R}_{Start}}\cdot {{R}_{End}}(2{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}))  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that since the starting and ending blocks cannot fail,  &amp;lt;math&amp;gt;{{R}_{Start}}=1&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{End}}=1,&amp;lt;/math&amp;gt;  Eqn.14 can be reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; 2\cdot {{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  is the reliability equation for Component A, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{B}}&amp;lt;/math&amp;gt;  is the reliability equation for Component  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; , etc.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the components in this example are identical, the system reliability equation can be further reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2R{{(t)}^{2}}+2R{{(t)}^{3}}-5R{{(t)}^{4}}+2R{{(t)}^{5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Or, in terms of the failure distribution:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.6.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.6.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.6 Reliability plot for the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.18 is taken with respect to time, resulting in: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; 4\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
&amp;amp; -20\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  can now be plotted for different time values,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , as shown in Figure 5.7.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system&#039;s failure rate can now be obtained by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn.19 by the system&#039;s reliability function given in Eqn.18, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}(t)= &amp;amp; \frac{4\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}} \\ &lt;br /&gt;
&amp;amp; +\frac{-20\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.8.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.7.gif|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.7 &#039;&#039;pdf&#039;&#039; plot for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of the system is obtained by integrating the system&#039;s reliability function given by Eqn. 18 from time zero to infinity, as given by Eqn. 10.  Using BlockSim&#039;s Analytical QCP, an  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of 1007.8 hours is calculated, as shown in Figure 5.9.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The warranty time can be obtained by solving Eqn. 18 with respect to time for a system reliability  &amp;lt;math&amp;gt;{{R}_{s}}=0.9&amp;lt;/math&amp;gt; .  Using the Analytical QCP and selecting the &amp;lt;br&amp;gt;&lt;br /&gt;
Warranty Time option, a time of 372.72 hours is obtained, as shown in Figure 5.10.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.8.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.8 Failure rate for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.9.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.9 MTTF of the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.10.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.10 Time at which &#039;&#039;R&#039;&#039;=0.9 or 90% for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.11.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.11 Conditional reliability calculation for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, the conditional reliability can be obtained using Eqn.6 and Eqn.18, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(200,200)= &amp;amp; \frac{R(400)}{R(200)} \\ &lt;br /&gt;
= &amp;amp; \frac{0.883825}{0.975321} \\ &lt;br /&gt;
= &amp;amp; 0.906189  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated using BlockSim&#039;s Analytical QCP, as shown in Figure 5.11.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Approximating the System CDF=&lt;br /&gt;
&lt;br /&gt;
In many cases, it is valuable to fit a distribution that represents the system&#039;s times-to-failure.  This can be useful when the system is part of a larger assembly and may be used for repeated calculations or in calculations for other systems.  In cases such as this, it can be useful to characterize the system&#039;s behavior by fitting a distribution to the overall system and calculating parameters for this distribution.   This is equivalent to fitting a single distribution to describe  &amp;lt;math&amp;gt;{{R}_{S}}(t&amp;lt;/math&amp;gt; ).  In essence, it is like reducing the entire system to a component in order to simplify calculations.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
To compute an approximate reliability function for this system,  &amp;lt;math&amp;gt;{{R}_{A}}(t)\simeq {{R}_{S}}(t)&amp;lt;/math&amp;gt; , one would compute  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  pairs of reliability and time values and then fit a single distribution to the data, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 10,396.7)=10% \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 9,361.9)=20% \\ &lt;br /&gt;
&amp;amp; ... \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 1,053.6)=90%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A single distribution,  &amp;lt;math&amp;gt;{{R}_{A}}(t)&amp;lt;/math&amp;gt; , that approximates  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  can now be computed from these pairs using life data analysis methods.  If using the Weibull++ software, one would enter the values as free form data.&lt;br /&gt;
&lt;br /&gt;
===Example 3===&lt;br /&gt;
Compute a single Weibull distribution approximation for the system in Example 2.&lt;br /&gt;
====Solution to Example 3====&lt;br /&gt;
The system in the previous example, shown in Figure 5.5, can be approximated by use of a 2-parameter Weibull distribution with  &amp;lt;math&amp;gt;\beta =2.02109&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1123.51&amp;lt;/math&amp;gt; .  In BlockSim, this is accomplished by representing the entire system as one distribution by going to the Distribution Fit window Figure 5.12. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.13.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.12 Representing a system with a distribution.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.12.PNG|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.13 Distribution Fitting window.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
by clicking the Distribution Fit Window, the Distribution Estimator window will appear (Figure 5.13).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this window you can select a distribution to represent the data. BlockSim will then generate a number of system failure times based on the system&#039;s reliability function. The system&#039;s reliability function can be used to solve for a time value associated with that unreliability value. The distribution of the generated time values can then be fitted to a probability distribution function.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a value of  &amp;lt;math&amp;gt;F(t)=0.11&amp;lt;/math&amp;gt; .  Using the system&#039;s reliability equation and solving for time, the corresponding time-to-failure for a 0.11 unreliability can be calculated.  &amp;lt;br&amp;gt;&lt;br /&gt;
For the system of Example 2, the time for a 0.11 unreliability is 389.786 hours.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
When enough points have been generated, the selected distribution will be fitted to this data set and the distribution&#039;s parameters will be returned.  In addition, if ReliaSoft&#039;s Weibull++ is installed, the generated data can be viewed/analyzed using a Weibull++ instance, as shown in Figure 5.14.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.14.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.14 Using Weibull++ to calculate distribution parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
It is recommended that the analyst examine the fit to ascertain the applicability of the approximation.&lt;br /&gt;
&lt;br /&gt;
=Duty Cycle=&lt;br /&gt;
&lt;br /&gt;
Components of a system may not operate continuously during a system&#039;s mission, or may be subjected to loads greater or lesser than the rated loads during system operation.  To model this, a factor called the Duty Cycle ( &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt; ) is used.  The duty cycle may also be used to account for changes in environmental stress, such as temperature changes, that may effect the operation of a component.  The duty cycle is a positive value, with a default value of 1 representing continuous operation at rated load, and any values other than 1 representing other load values with respect to the rated load value (or total operating time).   A duty cycle value higher than 1 indicates a load in excess of the rated value.  A duty cycle value lower than 1 indicates that the component is operating at a load lower than the rated load or not operating continuously during the system&#039;s mission.  For instance, a duty cycle of 0.5 may be used for a component that operates only half of the time during the system&#039;s mission.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The reliability metrics for a component with a duty cycle are calculated as follows. Let  &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt;  represent the duty cycle during a particular mission of the component,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  represent the mission time and  &amp;lt;math&amp;gt;{t}&#039;&amp;lt;/math&amp;gt;  represent the accumulated age. Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{t}&#039;={{d}_{c}}\times t&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability equation for the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R({t}&#039;)=R({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The component &#039;&#039;pdf&#039;&#039; is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;f({t}&#039;)=-\frac{d(R({t}&#039;))}{dt}=-\frac{d(R({{d}_{c}}\times t))}{dt}={{d}_{c}}f({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The failure rate of the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\lambda ({t}&#039;)=\frac{f({t}&#039;)}{R({t}&#039;)}=\frac{{{d}_{c}}f({{d}_{c}}\times t)}{R({{d}_{c}}\times t)}={{d}_{c}}\lambda ({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Example 4===&lt;br /&gt;
Consider a computer system with three components: a processor, a hard drive and a CD drive in series as shown next.  Assume that all three components follow a Weibull failure distribution with the parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}=1.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=5000&amp;lt;/math&amp;gt;  for the processor,  &amp;lt;math&amp;gt;{{\beta }_{2}}=2.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{2}}=3000&amp;lt;/math&amp;gt;  for the hard drive, and  &amp;lt;math&amp;gt;{{\beta }_{3}}=2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=4000&amp;lt;/math&amp;gt;  for the CD drive.  Determine the reliability of the computer system after one year (365 days) of operation, assuming that the CD drive is used only 30% of the time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5ex4.png|thumb|center|300px|]]&lt;br /&gt;
&lt;br /&gt;
====Solution to Example 4====&lt;br /&gt;
The reliability of the processor after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  {{R}_{processor}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{5000} \right)}^{1.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\text{ or }98.05%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the hard drive after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{harddrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{3000} \right)}^{2.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9948\text{ or }99.48%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the CD drive after 365 days of operation (taking into account the 30% operation using a duty cycle of 0.3) is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{CDdrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{{{d}_{c}}\times 365}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{0.3\times 365}{4000} \right)}^{2}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9993\text{ or }99.93%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.15.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.15 Result for the computer system reliability.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
Thus the reliability of the computer system after 365 days of operation is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{s}}(365)= &amp;amp; {{R}_{processor}}(365)\cdot {{R}_{harddrive}}(365)\cdot {{R}_{CDdrive}}(365) \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\cdot 0.9948\cdot 0.9993 \\ &lt;br /&gt;
	  = &amp;amp; 0.9747\text{ or }97.47%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;	&lt;br /&gt;
&amp;lt;br&amp;gt;	&lt;br /&gt;
This result can be obtained in BlockSim as shown in Figure DutyCycleExResults.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Load Sharing=&lt;br /&gt;
As presented in earlier chapters, a reliability block diagram (RBD) allows you to graphically represent how the components within a system are reliability-wise connected.  In most cases, independence is assumed across the components within the system.  For example, the failure of component A does not affect the failure of component B.  However, if a system consists of components that are sharing a load, then the assumption of independence no longer holds true.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If one component fails, then the component(s) that are still operating will have to assume the failed unit&#039;s portion of the load.  Therefore, the reliabilities of the surviving unit(s) will change.  Calculating the system reliability is no longer an easy proposition.  In the case of load sharing components, the change of the failure distributions of the surviving components must be known in order to determine the system&#039;s reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate this, consider the a system of two units connected reliability-wise in parallel (Figure 5.16).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-16.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.16 Two units connected reliability-wise in parallel.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Assume that the units must supply an output of 8 volts and that if both units are operational, each unit is to supply 50% of the total output.  If one of the units fails, then the surviving unit supplies 100%.  Furthermore, assume that having to supply the entire load has a negative impact on the reliability characteristics of the surviving unit.  Since the reliability characteristics of the unit change based on whether both or only one is operating, a life distribution along with a life-stress relationship (as discussed in Chapter 3) will be needed to model each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the steps needed, we will create the model starting from raw data.  Assume that a total of 20 units were tested to failure at 7, 10 and 15 volts.  The test data set is presented in the next table.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-17.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
For this example, Units 1 and 2 are the same component.  Therefore, only one set of data was collected.  However, it is possible that the load sharing components in a system may not be the same.  If that were the case, data would need to be collected for each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The data set in Table 1 was analyzed using ReliaSoft&#039;s ALTA software (as shown in Figure 5.17) with the Inverse Power Law as the underlying life-stress relationship and Weibull as the life distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated model parameters,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , are shown next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
\beta = &amp;amp; 1.9239 \\ &lt;br /&gt;
K= &amp;amp; 3.2387\times {{10}^{-7}} \\ &lt;br /&gt;
n= &amp;amp; 3.4226  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:Or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( 3.2387\times {{10}^{-7}}S_{1}^{3.4226}t \right)}^{1.9239}}}}  \ (eqn 20)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1}}(t,{{S}_{1}})=\beta KS_{1}^{n}{{\left( KS_{1}^{n}t \right)}^{\beta -1}}{{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}}  \ (eqn 21 )&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And for this case:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{R}_{2}}(t,{{S}_{2}})  \\ &lt;br /&gt;
{{f}_{1}}(t,{{S}_{1}})= &amp;amp; {{f}_{2}}(t,{{S}_{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.17.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.17 Using ALTA to calculate component parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 5.18 shows a plot of Eqn.20.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Now that the failure properties have been determined using the test data, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t,S)= &amp;amp; {{R}_{1}}(t,{{S}_{1}})\cdot {{R}_{2}}(t,{{S}_{2}}) \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{1}}\left( x,{{S}_{1}} \right)\cdot {{R}_{2}}(x,{{S}_{2}})\cdot \left( \frac{{{R}_{2}}({{t}_{1e}}+(t-x),S)}{{{R}_{2}}({{t}_{1e}},S)} \right)dx \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{2}}\left( x,{{S}_{2}} \right)\cdot {{R}_{1}}(x,{{S}_{1}})\cdot \left( \frac{{{R}_{1}}({{t}_{2e}}+(t-x),S)}{{{R}_{1}}({{t}_{2e}},S)} \right)dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:Where: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{S}_{1}}= &amp;amp; {{P}_{1}}S \\ &lt;br /&gt;
{{S}_{2}}= &amp;amp; {{P}_{2}}S  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-18.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig. 5.18 Reliability curves for different voltage output conditions. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.png|thumb|center|395px]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  is the total load (or required output).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{P}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{P}_{2}}&amp;lt;/math&amp;gt;  are the portion of the total load that each unit supports when both units are operational.  In this case,  &amp;lt;math&amp;gt;{{P}_{1}}={{P}_{2}}=0.5=50%.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{S}_{2}}&amp;lt;/math&amp;gt;  represent the portions of the load that Unit 1 and Unit 2 must support when both units are operational.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{{{1}_{e}}}}&amp;lt;/math&amp;gt;  is the equivalent operating time for Unit 1 if it had been operating at  &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  instead of  &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt; .  A graphical representation of the equivalent time is shown in Figure 5.19, where the curve marked by L represents the low stress (load) and the curve marked by H represents the high stress (load).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  can be calculated by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t)= &amp;amp; {{R}_{1}}({{t}_{1e}}) \\ &lt;br /&gt;
{{e}^{-{{(tKS_{1}^{n})}^{\beta }}}}= &amp;amp; {{e}^{-{{({{t}_{1e}}K{{S}^{n}})}^{\beta }}}} \\ &lt;br /&gt;
tS_{1}^{n}= &amp;amp; {{t}_{1e}}{{S}^{n}} \\ &lt;br /&gt;
{{t}_{1e}}= &amp;amp; t{{\left( \frac{{{S}_{1}}}{S} \right)}^{n}},\text{     }{{S}_{1}}={{P}_{1}}S \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{1e}}=tP_{1}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  can be calculated the same way, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{2}}(t)= &amp;amp; {{R}_{2}}({{t}_{2e}}) \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{2e}}=tP_{2}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this example, the reliability equations for Unit 1 and Unit 2 are the same since they are the same type of component and demonstrate the same failure properties.  In addition, the total output is divided equally between the two units (when both units are operating), so  &amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  will also be the same.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The next step is to determine the reliability of the system after 8,760 hours,  &amp;lt;math&amp;gt;R(t=8,760)&amp;lt;/math&amp;gt; .  Using Eqn. 22 the system reliability is found to be:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t=8760)= &amp;amp; 0.8567 \\ &lt;br /&gt;
= &amp;amp; 85.67%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Load Sharing in BlockSim===&lt;br /&gt;
BlockSim uses this formulation when computing reliabilities of units in a load sharing configuration.  When using the System Reliability Equation window, BlockSim returns a single token for the reliability of units in a load sharing configuration (as well as in the case of standby redundancy, discussed in the next section).  As an example, consider the following RBD with Unit 1 in series with a container that includes two load sharing units.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.2.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
BlockSim will return the system equation as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{System}}=+{{R}_{LS}}\cdot {{R}_{1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{LS}}&amp;lt;/math&amp;gt;  implies a form similar to Eqn. 22.  BlockSim allows for  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -out-of- &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  units in a load sharing configuration.&lt;br /&gt;
&lt;br /&gt;
===Example 5===&lt;br /&gt;
A component has five possible failure modes,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , and the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are interdependent.  The system will fail if mode  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  occurs, mode  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  occurs or two out of the three  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes occur.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  each have a Weibull distribution, with a  &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt;  and 15,000 respectively.  Events  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  each have an exponential distribution with a mean of 10,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If any  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  event occurs (i.e.  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt; ), the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are more likely to occur.  Specifically, the mean times of the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are halved.  Determine the reliability at 1000 hours for this component.&lt;br /&gt;
====Solution to Example 5====&lt;br /&gt;
The first step is to create the RBD.  Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and a load sharing container with the  &amp;lt;math&amp;gt;{{B}_{i}}&amp;lt;/math&amp;gt;  modes must be drawn in series, as illustrated next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.3.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The next step is to define the properties for each block, including those for the container.  Setting the failure distributions for modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is simple.  The more difficult part is setting the properties for the container and the contained blocks.  Based on the problem statement, the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are in a 2-out-of-3 load sharing redundancy.  When all three are working (i.e. when no  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode has occurred), each block has an exponential distribution with &amp;lt;math&amp;gt;\mu=10,000&amp;lt;/math&amp;gt;.  If one  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode occurs, then the two surviving units have an exponential distribution with  &amp;lt;math&amp;gt;\mu =5,000.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assume a Power Life-Stress relationship for the components.  Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\mu }_{1}}= &amp;amp; \frac{1}{KV_{1}^{n}}  \ (eqn 23)\\ &lt;br /&gt;
{{\mu }_{2}}= &amp;amp; \frac{1}{KV_{2}^{n}}  \ (eqn 24)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{1}}=10,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{V}_{1}}=1&amp;lt;/math&amp;gt;  in Eqn. 23 and casting it in terms of  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
10,000= &amp;amp; \frac{1}{K}  \ (eqn 25) \\ &lt;br /&gt;
K = &amp;amp; \frac{1}{10,000}=0.0001  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{2}}=5,000&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{V}_{2}}=1.5&amp;lt;/math&amp;gt;  (because if one fails, then each survivor takes on an additional 0.5 units of load) and Eqn. 25 for  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  in Eqn.24 yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
5,000= &amp;amp; \frac{1}{0.0001\cdot {{(1.5)}^{n}}} \\ &lt;br /&gt;
0.5= &amp;amp; {{(1.5)}^{-n}} \\ &lt;br /&gt;
\ln (0.5)= &amp;amp; -n\ln (1.5) \\ &lt;br /&gt;
n = &amp;amp; 1.7095  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This also could have been computed in ALTA, as shown in Figure 5.20, or with the Load &amp;amp; Life Parameter Experimenter in BlockSim, as shown in Figure 5.21.&lt;br /&gt;
 &lt;br /&gt;
At this point, the parameters for the load sharing units have been computed and can be set, as shown in Figure 5.22.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The next step is to set the weight proportionality factor.  This factor defines the portion of the load that the particular item carries while operating, as well as the load that shifts to the remaining units upon failure of the item.  To illustrate, assume three units (1, 2 and 3) are in a load sharing container with weight proportionality factors of 1, 2 and 3 respectively (and a 1-out-of-3 requirement).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Unit 1 carries  &amp;lt;math&amp;gt;\left( \tfrac{1}{1+2+3} \right)=0.166&amp;lt;/math&amp;gt;  or 16.6% of the total load.&lt;br /&gt;
:•	Unit 2 carries  &amp;lt;math&amp;gt;\left( \tfrac{2}{1+2+3} \right)=0.333&amp;lt;/math&amp;gt;  or 33.3% of the total load.&lt;br /&gt;
:•	Unit 3 carries  &amp;lt;math&amp;gt;\left( \tfrac{3}{1+2+3} \right)=0.50&amp;lt;/math&amp;gt;  or 50% of the total load.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The actual load on each unit then becomes the product of the entire load defined for the container times the portion carried by that unit.  For example, if the container load is 100 lbs, then the portion assigned to Unit 1 will be  &amp;lt;math&amp;gt;100\cdot 0.166=16.6&amp;lt;/math&amp;gt;  lbs.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the current example, all units share the same load and thus have equal weight proportionality factors.  Because these factors are relative, if the same number is used for all three items then the results will be the same.  Thus, weight proportional factor is set equal to 1 for each item.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5.20.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.20 Calculation performed in ALTA.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.21_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.21 Quick Parameter Estimator &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.21_3.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.22 Quick Parameter Estimator results &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.23_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.23 Defining Weight Proportional Factor. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The last properties that need to be defined are the total load and the 2-out-of-3 redundancy.  The total load is dependent on how the parameters were computed.  In this case, total load was assumed to be 3 when the parameters were computed (i.e. the load per item was 1 when all worked and 1.5 when two worked).  This is defined at the container level, set No. of Paths required = 3.&lt;br /&gt;
When all of the parameters have been specified in BlockSim, the reliability at 1,000 hours can be determined.  From the Analytical QCP, this is found to be 98.57%.&lt;br /&gt;
&lt;br /&gt;
=Standby Components=&lt;br /&gt;
&lt;br /&gt;
In the previous section, the case of a system with load sharing components was presented.  This is a form of redundancy with dependent components. That is, the failure of one component affects the failure of the other(s).  This section presents another form of redundancy: standby redundancy.  In standby redundancy the redundant components are set to be under a lighter load condition (or no load) while not needed and under the operating load when they are activated.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In standby redundancy the components are set to have two states: an active state and a standby state.  Components in standby redundancy have two failure distributions, one for each state.  When in the standby state, they have a quiescent (or dormant) failure distribution and when operating, they have an active failure distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case that both quiescent and active failure distributions are the same, the units are in a simple parallel configuration (also called a hot standby configuration).  When the rate of failure of the standby component is lower in quiescent mode than in active mode, that is called a warm standby configuration.  When the rate of failure of the standby component is zero in quiescent mode (i.e. the component cannot fail when in standby), that is called a cold standby configuration.  &lt;br /&gt;
&lt;br /&gt;
===Simple Standby Configuration===&lt;br /&gt;
&lt;br /&gt;
Consider two components in a standby configuration.  Component 1 is the active component with a Weibull failure distribution with parameters  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Component 2 is the standby component.  When Component 2 is operating, it also has a Weibull failure distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Furthermore, assume the following cases for the quiescent distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1:  The quiescent distribution is the same as the active distribution (hot standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2:  The quiescent distribution is a Weibull  distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 2000 (warm standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3: The component cannot fail in quiescent mode (cold standby).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this case, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be obtained using the following equation:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(t)={{R}_{1}}(t)+\underset{0}{\overset{t}{\mathop \int }}\,{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x)\cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}dx   \ (eqn 26)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode, such that: &amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
&amp;lt;br&amp;gt;  &lt;br /&gt;
&amp;lt;math&amp;gt;{{R}_{2;SB}}(x)={{R}_{2;A}}({{t}_{e}})  \ (eqn 27)&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Eqn. 27 can be solved for  &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  and substituted into Eqn.26.&lt;br /&gt;
Figure 5.24 illustrates the example as entered in BlockSim using a standby container.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5.24.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.24 Standby container.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The active and standby blocks are within a container, which is used to specify standby redundancy.  Since the standby component has two distributions (active and quiescent), the Block Properties window of the standby block has two pages for specifying each one.  Figures 5.24 and 5.26 illustrate these pages.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system reliability results for 1000 hours are given in the following table:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-24.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Note that even though the  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  value for the quiescent distribution is the same as in the active distribution, it is possible that the two can be different. That is, the failure modes present during the quiescent mode could be different from the modes present during the active mode.  In that sense, the two distribution types can be different as well (e.g. lognormal when quiescent and Weibull when active).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In many cases when considering standby systems, a switching device may also be present that switches from the failed active component to the standby component.  The reliability of the switch can also be incorporated into Eqn.26, as presented in the next section.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim&#039;s System Reliability Equation window returns a single token for the reliability of units in a standby configuration.  This is the same as the load sharing case presented in the previous section. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image: Fig 5.25.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.25 Defining the active failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.26.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.26 Defining the quiescent failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Reliability of Standby Systems with a Switching Device===&lt;br /&gt;
&lt;br /&gt;
In many cases when dealing with standby systems, a switching device is present that will switch to the standby component when the active component fails.  Therefore, the failure properties of the switch must also be included in the analysis.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.26.2.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In most cases when the reliability of a switch is to be included in the analysis, two probabilities can be considered.  The first and most common one is the probability of the switch performing the action (i.e. switching) when requested to do so.  This is called Switch Probability per Request in BlockSim and is expressed as a static probability (e.g. 90%).  The second probability is the quiescent reliability of the switch.  This is the reliability of the switch as it ages (e.g. the switch might wear out with age due to corrosion, material degradation, etc.). Thus it is possible for the switch to fail before the active component fails.  However, a switch failure does not cause the system to fail, but rather causes the system to fail only if the switch is needed and the switch has failed.  For example, if the active component does not fail until the mission end time and the switch fails, then the system does not fail.  However, if the active component fails and the switch has also failed, then the system cannot be switched to the standby component and it therefore fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In analyzing standby components with a switching device, either or both failure probabilities (during the switching or while waiting to switch) can be considered for the switch, since each probability can represent different failure modes.  For example, the switch probability per request may represent software-related issues or the probability of detecting the failure of an active component, and the quiescent probability may represent wear-out type failures of the switch.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the formulation, consider the previous example that assumes perfect switching.  To examine the effects of including an imperfect switch, assume that when the active component fails there is a 90% probability that the switch will switch from the active component to the standby component.  In addition, assume that the switch can also fail due to a wear-out failure mode described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.7 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 5000.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , is given by the following equation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t)= &amp;amp; {{R}_{1}}(t) \\ &lt;br /&gt;
&amp;amp; +\underset{0}{\overset{t}{\mathop \int }}\,\{{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x) \\ &lt;br /&gt;
&amp;amp; \cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}\cdot {{R}_{SW;Q}}(x)\cdot {{R}_{SW;REQ}}(x)\}dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;Q}}&amp;lt;/math&amp;gt;  is the quiescent reliability of the switch.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;REQ}}&amp;lt;/math&amp;gt;  is the switch probability per request.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
This problem can be solved in BlockSim by including these probabilities in the container&#039;s properties, as shown in Figures 5.27 and 5.28.  In BlockSim, the standby container is acting as the switch.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.28_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.27 Standby container (switch) failure probabilities while attempting to switch &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.27.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.28 Standby container (switch) failure distribution while waiting to switch &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that there are additional properties that can be specified in BlockSim for a switch, such as Switch Restart Probability, No. of Restarts and Switch Delay Time.  In many applications, the switch is re-tested (or re-cycled) if it fails to switch the first time.  In these cases, it might be possible that it switches in the second or third, or  &amp;lt;math&amp;gt;{{n}^{th}}&amp;lt;/math&amp;gt; attempt.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The Switch Restart Probability specifies each additional attempt&#039;s probability of successfully switching and the Finite Restarts specifies the total number of attempts.  Note that the Switch Restart Probability specifies the probability of success of each trial (or attempt).  The probability of success of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  consecutive trials is calculated by BlockSim using the binomial distribution and this probability is then incorporated into Eqn.28.  The Switch Delay Time property is related to repairable systems and is considered in BlockSim only when using simulation.  When using the analytical solution (i.e. for a non-repairable system), this property is ignored.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving the analytical solution (as given by Eqn. 28), the following results are obtained.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-30.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From the table above, it can be seen that the presence of a switching device has a significant effect on the reliability of a standby system.  It is therefore important when modeling standby redundancy to incorporate the switching device reliability properties.  It should be noted that this methodology is not the same as treating the switching device as another series component with the standby subsystem.  This would be valid only if the failure of the switch resulted in the failure of system (e.g. switch failing open).  In Eqn.28, the Switch Probability per Request and quiescent probability are present only in the second term of the equation.  Treating these two failure modes as a series configuration with the standby subsystem would imply that they are also present when the active component is functioning (i.e. first term of Eqn.28).  This is invalid and would result in the underestimation of the reliability of the system.  In other words, these two failure modes become significant only when the active component fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As an example, and if we consider the warm standby case, the reliability of the system without the switch is 70.57% at 1000 hours.  If the system was modeled so that the switching device was in series with the warm standby subsystem, the result would have been:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(1000)= &amp;amp; {{R}_{Standby}}(1000)\cdot {{R}_{sw,Q(1000)}}\cdot {{R}_{sw,req}} \\ &lt;br /&gt;
= &amp;amp; 0.7057\cdot 0.9372\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.5952  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case where a switch failure mode causes the standby subsystem to fail, then this mode can be modeled as an individual block in series with the standby subsystem.&lt;br /&gt;
&lt;br /&gt;
===Example 6===&lt;br /&gt;
Consider a car with four new tires and a full-size spare.  Assume the following failure characteristics:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires follow a Weibull distribution with a  ..  and an  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  40,000 miles while on the car due to wear.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires also have a probability of failing due to puncture or other causes.  For this, assume a constant rate for this occurrence with a probability of 1 every 50,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	When not on the car (i.e. is a spare), a tire&#039;s probability of failing also has a Weibull distribution with a  &amp;lt;math&amp;gt;\beta =&amp;lt;/math&amp;gt;  2 and  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  120,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Assume a mission of 1,000 miles.  If a tire fails during this trip, it will be replaced with the spare.  However, the spare will not be repaired during the trip.  In other words, the trip will continue with the spare on the car and if the spare fails the system will fail.  Determine the probability of system failure.&lt;br /&gt;
====Solution to Example 6====&lt;br /&gt;
Active failure distribution for tires:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to wear-out, Weibull  &amp;lt;math&amp;gt;\beta =4&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =40,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to random puncture, exponential  &amp;lt;math&amp;gt;\mu =50,000.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	The quiescent failure distribution is a Weibull distribution with &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =120,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The block diagram for each tire has two blocks in series, one block representing the wear-out mode and the other the random puncture mode, as shown next:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:small5.gif|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
There are five tires, four active and one standby (represented in the diagram by a standby container with a 4-out-of-5 requirement), as shown next: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BStirecontainer.png|thumb|center|400px|]]&lt;br /&gt;
 &lt;br /&gt;
For the standby Wear block, set the active failure and the quiescent distributions, but for the Puncture block, set only the active puncture distribution (because the tire cannot fail due to puncture while stored).  Using BlockSim, the probability of system failure is found to be 0.003 or 0.3%. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Note Regarding Numerical Integration Solutions=&lt;br /&gt;
&lt;br /&gt;
Load sharing and standby solutions in BlockSim are performed using numerical integration routines.  As with any numerical analysis routine, the solution error depends on the number of iterations performed, the step size chosen and related factors, plus the behavior of the underlying function.  By default, BlockSim uses a certain set of preset factors.  In general, these defaults are sufficient for most problems.  If a higher precision or verification of the precision for a specific problem is required, BlockSim&#039;s preset options can be modified and/or the integration error can be assessed using the Integration Parameters... option for each container.  For more details, you can refer to the documentation on the Algorithm Setup window in the BlockSim 7 User&#039;s Guide.&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15576</id>
		<title>Time-Dependent System Reliability (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15576"/>
		<updated>2012-02-13T22:25:42Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Reliability of Standby Systems with a Switching Device */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|5}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the previous chapter, different system configuration types were examined, as well as different methods for obtaining the system&#039;s reliability function analytically.  Because the reliabilities in the problems presented were treated as probabilities (e.g.  &amp;lt;math&amp;gt;P(A)&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; ), the reliability values and equations presented were referred to as static (not time-dependent).  Thus, in the prior chapter, the life distributions of the components were not incorporated in the process of calculating the system reliability.  In this chapter, time dependency in the reliability function will be introduced.  We will develop the models necessary to observe the reliability over the life of the system, instead of at just one point in time.  In addition, performance measures such as failure rate, MTTF and warranty time will be estimated for the entire system.  The methods of obtaining the reliability function analytically remain identical to the ones presented in the previous chapter, with the exception that the reliabilities will be functions of time.  In other words, instead of dealing with  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , we will use  &amp;lt;math&amp;gt;{{R}_{i}}(t)&amp;lt;/math&amp;gt; .  All examples in this chapter assume that no repairs are performed on the components.  &lt;br /&gt;
&lt;br /&gt;
=Analytical Life Predictions=&lt;br /&gt;
The analytical approach presented in the prior chapter involved the determination of a mathematical expression that describes the reliability of the system, expressed in terms of the reliabilities of its components.  So far we have estimated only static system reliability (at a fixed time).  For example, in the case of a system with three components in series, the system&#039;s reliability equation was given by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}  \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The values of  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  were given for a common time and the reliability of the system was estimated for that time.  However, since the component failure characteristics can be described by distributions, the system reliability is actually time-dependent.  In this case, Eqn. (1) can be rewritten as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability of the system for any mission time can now be estimated.  Assuming a Weibull life distribution for each component, Eqn. (1) can now be expressed in terms of each component&#039;s reliability function, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the same manner, any life distribution can be substituted into the system reliability equation.  Suppose that the times-to-failure of the first component are described with a Weibull distribution, the times-to-failure of the second component with an exponential distribution and the times-to-failure of the third component with a normal distribution.  Then Eqn. (1) can be written as:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot \left[ 1-\Phi \left( \frac{t-{{\mu }_{3}}}{{{\sigma }_{3}}} \right) \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It can be seen that the biggest challenge is in obtaining the system&#039;s reliability function in terms of component reliabilities, which has already been discussed in depth.  Once this has been achieved, calculating the reliability of the system for any mission duration is just a matter of substituting the corresponding component reliability functions into the system reliability equation.&lt;br /&gt;
===Advantages of the Analytical Method===&lt;br /&gt;
The primary advantage of the analytical solution is that it produces a mathematical expression that describes the reliability of the system.  Once the system&#039;s reliability function has been determined, other calculations can then be performed to obtain metrics of interest for the system. Such calculations include:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of warranty periods.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s failure rate.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s MTTF.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, optimization and reliability allocation techniques can be used to aid engineers in their design improvement efforts.  Another advantage in using analytical techniques is the ability to perform static calculations and analyze systems with a mixture of static and time-dependent components.  Finally, the reliability importance of components over time can be calculated with this methodology.&lt;br /&gt;
===Disadvantages of the Analytical Method===&lt;br /&gt;
The biggest disadvantage of the analytical method is that formulations can become very complicated.  The more complicated a system is, the larger and more difficult it will be to analytically formulate an expression for the system&#039;s reliability.  For particularly detailed systems this process can be quite time-consuming, even with the use of computers.  Furthermore, when the maintainability of the system or some of its components must be taken into consideration, analytical solutions become intractable.  In these situations, the use of simulation methods may be more advantageous than attempting to develop a solution analytically.  Simulation methods are presented in later chapters.&lt;br /&gt;
===Looking at a Simple &#039;&#039;Complex&#039;&#039; System Analytically===&lt;br /&gt;
&lt;br /&gt;
The complexity involved in an analytical solution can be best illustrated by looking at the simple &#039;&#039;complex&#039;&#039; system with 15 components, as shown in Figure 5.1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1.png|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.1 An RBD of a complex system.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system reliability for this system (computed using BlockSim) is shown next.  The first solution is provided using BlockSim&#039;s symbolic solution.  In symbolic mode, BlockSim breaks the equation into segments, identified by tokens, that need to be substituted into the final system equation for a complete solution.  This creates algebraic solutions that are more compact than if the substitutions were made.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; D2\cdot D3\cdot {{R}_{L}} \\ &lt;br /&gt;
D3= &amp;amp; +{{R}_{K}}\cdot IK \\ &lt;br /&gt;
IK= &amp;amp; +{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}} \\ &lt;br /&gt;
D2 = &amp;amp; +{{R}_{A}}\cdot {{R}_{E}}\cdot IE \\ &lt;br /&gt;
IE = &amp;amp; -D1\cdot {{R}_{M}}\cdot {{R}_{N}}+{{R}_{M}}\cdot {{R}_{N}}+D1 \\ &lt;br /&gt;
D1 = &amp;amp; +{{R}_{D}}\cdot ID \\ &lt;br /&gt;
ID = &amp;amp; -{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting the terms yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; {{R}_{A}}\cdot {{R}_{E}}\cdot {{R}_{L}}\cdot {{R}_{K}} \\ &lt;br /&gt;
&amp;amp; \cdot \{({{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})\cdot {{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}}-{{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}\} \\ &lt;br /&gt;
&amp;amp; \cdot \{{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}}\}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
BlockSim&#039;s automatic algebraic simplification would yield the following format for the above solution: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; (({{R}_{A}}\cdot {{R}_{E}}(-({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})){{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})))) \\ &lt;br /&gt;
&amp;amp; ({{R}_{K}}({{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; +RI\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}})){{R}_{L}})  \ (eqn 2)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this equation, each  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  represents the reliability function of a block.  For example, if  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  has a Weibull distribution, then each  &amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;  and so forth.  Substitution of each component&#039;s reliability function in Eqn.2 will result in an analytical expression for the system reliability as a function of time, or  &amp;lt;math&amp;gt;{{R}_{s}}(t)&amp;lt;/math&amp;gt; , which is the same as  &amp;lt;math&amp;gt;(1-cd{{f}_{System}}).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Obtaining Other Functions of Interest===&lt;br /&gt;
Once the system reliability equation (or the cumulative density function,  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt; ) has been determined, other functions and metrics of interest can be derived.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Consider the following simple system:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-2.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.2 Simple two-component system. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Furthermore, assume that component 1 follows an exponential distribution with a mean of 10,000 (&amp;lt;math&amp;gt;\mu =10,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;\lambda =1/10,000)&amp;lt;/math&amp;gt;  and component 2 follows a Weibull distribution with  &amp;lt;math&amp;gt;\beta =6&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt; .  The reliability equation of this system is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}  \ (eqn 3) &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{F}_{S}}(t)= &amp;amp; 1-({{R}_{1}}(t)\cdot {{R}_{2}}(t)) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \right) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
{{analytical system pdf}}&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability====&lt;br /&gt;
Conditional reliability is the probability of a system successfully completing another mission following the successful completion of a previous mission.  The time of the previous mission and the time for the mission to be undertaken must be taken into account for conditional reliability calculations.  The system&#039;s conditional reliability function is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(T,t)=\frac{R(T+t)}{R(T)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Eqn.6 gives the reliability for a new mission of duration  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  having already accumulated  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  hours of operation up to the start of this new mission. The system is evaluated to assure that it will start the next mission successfully.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2, the reliability for mission of  &amp;lt;math&amp;gt;t=1,000&amp;lt;/math&amp;gt;  hours, having an age of  &amp;lt;math&amp;gt;T=500&amp;lt;/math&amp;gt;  hours, is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(T=500,t=1000)= &amp;amp; \frac{R(T+t)}{R(T)} \\ &lt;br /&gt;
= &amp;amp; \frac{R(1500)}{R(500)} \\ &lt;br /&gt;
= &amp;amp; \frac{{{e}^{-\tfrac{1500}{10,000}}}\cdot {{e}^{-{{\left( \tfrac{1500}{10,000} \right)}^{6}}}}}{{{e}^{-\tfrac{500}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{500}{10,000} \right)}^{6}}}}} \\ &lt;br /&gt;
= &amp;amp; 0.9048=90.48%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.3.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.4.png|thumb|center|400px|]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability for Components====&lt;br /&gt;
&lt;br /&gt;
Now in this formulation, it was assumed that the accumulated age was equivalent for both units. That is, both started life at zero and aged to 500.  It is possible to consider an individual component that has already accumulated some age (used component) in the same formulation.  To illustrate this, assume that component 2 started life with an age of T=100.  Then the reliability equation of the system, as given in Eqn.3, would need to be modified to include a conditional term for 2, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{R}_{1}}(t)\cdot \frac{{{R}_{2}}({{T}_{2}}+t)}{{{R}_{2}}({{T}_{2}})} \ (eqn 7)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In BlockSim, the start age input box may be used to specify a starting age greater than zero.&lt;br /&gt;
{{system failure rate analytical}}&lt;br /&gt;
&lt;br /&gt;
====System Mean Life (Mean Time To Failure)====&lt;br /&gt;
The mean life (or mean time to failure, MTTF) can be obtained by integrating the system reliability function from zero to infinity: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;MTTF=\int_{0}^{\infty }{{R}_{s}}\left( t \right)dt   \ (eqn 10)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mean time is a performance index and does not provide any information about the behavior of the failure distribution of the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)dt \\ &lt;br /&gt;
= &amp;amp; 5978.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Warranty Period and BX Life====&lt;br /&gt;
Sometimes it is desirable to know the time value associated with a certain reliability.  Warranty periods are often calculated by determining what percentage of the failure population can be covered financially and estimating the time at which this portion of the population will fail.  Similarly, engineering specifications may call for a certain BX life, which also represents a time period during which a certain proportion of the population will fail.  For example, the B10 life is the time in which 10% of the population will fail.  &lt;br /&gt;
This is obtained by setting  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  to the desired value and solving for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}\left( t \right)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To compute the time by which reliability would be equal to 90%, Eqn.11 is recast as follows and solved for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this case,  &amp;lt;math&amp;gt;t=1053.59&amp;lt;/math&amp;gt; .  Equivalently, the B10 life for this system is also  &amp;lt;math&amp;gt;1053.59&amp;lt;/math&amp;gt; .&lt;br /&gt;
Except for some trivial cases, a closed form solution for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  cannot be obtained.   Thus, it is necessary to solve for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  using numerical methods.  BlockSim uses numerical methods.&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
Consider a system consisting of three exponential units in series with the following failure rates (in failures per hour):  &amp;lt;math&amp;gt;{{\lambda }_{1}}&amp;lt;/math&amp;gt;  = 0.0002,  &amp;lt;math&amp;gt;{{\lambda }_{2}}&amp;lt;/math&amp;gt;  = 0.0005 and  &amp;lt;math&amp;gt;{{\lambda }_{3}}&amp;lt;/math&amp;gt;  = 0.0001.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the reliability equation for the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the reliability of the system after 150 hours of operation?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s failure rate equation.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the MTTF for the system?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What should the warranty period be for a 90% reliability?&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The analytical expression for the reliability of the system is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\lambda }_{1}}t}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot {{e}^{-{{\lambda }_{1}}t}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
At 150 hours of operation, the reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{e}^{-(0.0002+0.0005+0.0001)150}} \\ &lt;br /&gt;
= &amp;amp; 0.8869\text{ or }88.69%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.12 is taken with respect to time, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; -\frac{d[{{R}_{s}}(t)]}{dt} \\ &lt;br /&gt;
= &amp;amp; -\frac{d\left[ {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}} \right]}{dt} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
:•	The system&#039;s failure rate can now be obtained simply by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn. 13 by the system&#039;s reliability function given in Eqn.12, and:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}\left( t \right)= &amp;amp; \frac{{{f}_{s}}\left( t \right)}{{{R}_{s}}\left( t \right)} \\ &lt;br /&gt;
= &amp;amp; \frac{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}}{{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}) \\ &lt;br /&gt;
= &amp;amp; 0.0008\text{ }fr/hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Combining Eqn.10 and Eqn. 12, the system&#039;s MTTF can be obtained:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }{{R}_{s}}\left( t \right)dt \\ &lt;br /&gt;
= &amp;amp; \int_{0}^{\infty }{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}dt \\ &lt;br /&gt;
= &amp;amp; \frac{1}{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})} \\ &lt;br /&gt;
= &amp;amp; 1250\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Solving Eqn. 12 with respect to time will yield the corresponding warranty period for a 90% reliability.  In this case, the system reliability equation is simple and a closed form solution exists.  The warranty time can now be found by solving:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
t= &amp;amp; -\frac{\ln (R)}{{{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}} \\ &lt;br /&gt;
= &amp;amp; -\frac{\ln (0.9)}{0.0008} \\ &lt;br /&gt;
= &amp;amp; 131.7\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the warranty period should be 132 hours.&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
Consider the system shown in Figure 5.5.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.5.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.5 Complex bridge system in Example 2. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Components  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  through  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  are Weibull distributed with  &amp;lt;math&amp;gt;\beta =1.2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1230&amp;lt;/math&amp;gt;  hours.  The starting and ending blocks cannot fail.  &amp;lt;br&amp;gt;&lt;br /&gt;
Determine the following:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability equation for the system and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s failure rate equation and the corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The warranty time for a 90% reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability for a 200-hour mission, if it is known that the system has already successfully operated for 200 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The first step is to obtain the reliability function for the system.  The methods described in the previous chapter can be employed, such as the event space or path-tracing methods.  Using BlockSim, the following reliability equation is obtained:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; ({{R}_{Start}}\cdot {{R}_{End}}(2{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}))  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that since the starting and ending blocks cannot fail,  &amp;lt;math&amp;gt;{{R}_{Start}}=1&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{End}}=1,&amp;lt;/math&amp;gt;  Eqn.14 can be reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; 2\cdot {{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  is the reliability equation for Component A, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{B}}&amp;lt;/math&amp;gt;  is the reliability equation for Component  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; , etc.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the components in this example are identical, the system reliability equation can be further reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2R{{(t)}^{2}}+2R{{(t)}^{3}}-5R{{(t)}^{4}}+2R{{(t)}^{5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Or, in terms of the failure distribution:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.6.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.6.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.6 Reliability plot for the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.18 is taken with respect to time, resulting in: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; 4\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
&amp;amp; -20\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  can now be plotted for different time values,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , as shown in Figure 5.7.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system&#039;s failure rate can now be obtained by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn.19 by the system&#039;s reliability function given in Eqn.18, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}(t)= &amp;amp; \frac{4\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}} \\ &lt;br /&gt;
&amp;amp; +\frac{-20\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.8.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.7.gif|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.7 &#039;&#039;pdf&#039;&#039; plot for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of the system is obtained by integrating the system&#039;s reliability function given by Eqn. 18 from time zero to infinity, as given by Eqn. 10.  Using BlockSim&#039;s Analytical QCP, an  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of 1007.8 hours is calculated, as shown in Figure 5.9.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The warranty time can be obtained by solving Eqn. 18 with respect to time for a system reliability  &amp;lt;math&amp;gt;{{R}_{s}}=0.9&amp;lt;/math&amp;gt; .  Using the Analytical QCP and selecting the &amp;lt;br&amp;gt;&lt;br /&gt;
Warranty Time option, a time of 372.72 hours is obtained, as shown in Figure 5.10.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.8.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.8 Failure rate for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.9.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.9 MTTF of the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.10.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.10 Time at which &#039;&#039;R&#039;&#039;=0.9 or 90% for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.11.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.11 Conditional reliability calculation for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, the conditional reliability can be obtained using Eqn.6 and Eqn.18, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(200,200)= &amp;amp; \frac{R(400)}{R(200)} \\ &lt;br /&gt;
= &amp;amp; \frac{0.883825}{0.975321} \\ &lt;br /&gt;
= &amp;amp; 0.906189  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated using BlockSim&#039;s Analytical QCP, as shown in Figure 5.11.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Approximating the System CDF=&lt;br /&gt;
&lt;br /&gt;
In many cases, it is valuable to fit a distribution that represents the system&#039;s times-to-failure.  This can be useful when the system is part of a larger assembly and may be used for repeated calculations or in calculations for other systems.  In cases such as this, it can be useful to characterize the system&#039;s behavior by fitting a distribution to the overall system and calculating parameters for this distribution.   This is equivalent to fitting a single distribution to describe  &amp;lt;math&amp;gt;{{R}_{S}}(t&amp;lt;/math&amp;gt; ).  In essence, it is like reducing the entire system to a component in order to simplify calculations.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
To compute an approximate reliability function for this system,  &amp;lt;math&amp;gt;{{R}_{A}}(t)\simeq {{R}_{S}}(t)&amp;lt;/math&amp;gt; , one would compute  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  pairs of reliability and time values and then fit a single distribution to the data, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 10,396.7)=10% \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 9,361.9)=20% \\ &lt;br /&gt;
&amp;amp; ... \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 1,053.6)=90%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A single distribution,  &amp;lt;math&amp;gt;{{R}_{A}}(t)&amp;lt;/math&amp;gt; , that approximates  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  can now be computed from these pairs using life data analysis methods.  If using the Weibull++ software, one would enter the values as free form data.&lt;br /&gt;
&lt;br /&gt;
===Example 3===&lt;br /&gt;
Compute a single Weibull distribution approximation for the system in Example 2.&lt;br /&gt;
====Solution to Example 3====&lt;br /&gt;
The system in the previous example, shown in Figure 5.5, can be approximated by use of a 2-parameter Weibull distribution with  &amp;lt;math&amp;gt;\beta =2.02109&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1123.51&amp;lt;/math&amp;gt; .  In BlockSim, this is accomplished by representing the entire system as one distribution by going to the Distribution Fit window Figure 5.12. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.13.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.12 Representing a system with a distribution.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.12.PNG|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.13 Distribution Fitting window.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
by clicking the Distribution Fit Window, the Distribution Estimator window will appear (Figure 5.13).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this window you can select a distribution to represent the data. BlockSim will then generate a number of system failure times based on the system&#039;s reliability function. The system&#039;s reliability function can be used to solve for a time value associated with that unreliability value. The distribution of the generated time values can then be fitted to a probability distribution function.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a value of  &amp;lt;math&amp;gt;F(t)=0.11&amp;lt;/math&amp;gt; .  Using the system&#039;s reliability equation and solving for time, the corresponding time-to-failure for a 0.11 unreliability can be calculated.  &amp;lt;br&amp;gt;&lt;br /&gt;
For the system of Example 2, the time for a 0.11 unreliability is 389.786 hours.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
When enough points have been generated, the selected distribution will be fitted to this data set and the distribution&#039;s parameters will be returned.  In addition, if ReliaSoft&#039;s Weibull++ is installed, the generated data can be viewed/analyzed using a Weibull++ instance, as shown in Figure 5.14.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.14.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.14 Using Weibull++ to calculate distribution parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
It is recommended that the analyst examine the fit to ascertain the applicability of the approximation.&lt;br /&gt;
&lt;br /&gt;
=Duty Cycle=&lt;br /&gt;
&lt;br /&gt;
Components of a system may not operate continuously during a system&#039;s mission, or may be subjected to loads greater or lesser than the rated loads during system operation.  To model this, a factor called the Duty Cycle ( &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt; ) is used.  The duty cycle may also be used to account for changes in environmental stress, such as temperature changes, that may effect the operation of a component.  The duty cycle is a positive value, with a default value of 1 representing continuous operation at rated load, and any values other than 1 representing other load values with respect to the rated load value (or total operating time).   A duty cycle value higher than 1 indicates a load in excess of the rated value.  A duty cycle value lower than 1 indicates that the component is operating at a load lower than the rated load or not operating continuously during the system&#039;s mission.  For instance, a duty cycle of 0.5 may be used for a component that operates only half of the time during the system&#039;s mission.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The reliability metrics for a component with a duty cycle are calculated as follows. Let  &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt;  represent the duty cycle during a particular mission of the component,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  represent the mission time and  &amp;lt;math&amp;gt;{t}&#039;&amp;lt;/math&amp;gt;  represent the accumulated age. Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{t}&#039;={{d}_{c}}\times t&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability equation for the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R({t}&#039;)=R({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The component &#039;&#039;pdf&#039;&#039; is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;f({t}&#039;)=-\frac{d(R({t}&#039;))}{dt}=-\frac{d(R({{d}_{c}}\times t))}{dt}={{d}_{c}}f({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The failure rate of the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\lambda ({t}&#039;)=\frac{f({t}&#039;)}{R({t}&#039;)}=\frac{{{d}_{c}}f({{d}_{c}}\times t)}{R({{d}_{c}}\times t)}={{d}_{c}}\lambda ({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Example 4===&lt;br /&gt;
Consider a computer system with three components: a processor, a hard drive and a CD drive in series as shown next.  Assume that all three components follow a Weibull failure distribution with the parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}=1.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=5000&amp;lt;/math&amp;gt;  for the processor,  &amp;lt;math&amp;gt;{{\beta }_{2}}=2.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{2}}=3000&amp;lt;/math&amp;gt;  for the hard drive, and  &amp;lt;math&amp;gt;{{\beta }_{3}}=2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=4000&amp;lt;/math&amp;gt;  for the CD drive.  Determine the reliability of the computer system after one year (365 days) of operation, assuming that the CD drive is used only 30% of the time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5ex4.png|thumb|center|300px|]]&lt;br /&gt;
&lt;br /&gt;
====Solution to Example 4====&lt;br /&gt;
The reliability of the processor after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  {{R}_{processor}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{5000} \right)}^{1.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\text{ or }98.05%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the hard drive after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{harddrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{3000} \right)}^{2.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9948\text{ or }99.48%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the CD drive after 365 days of operation (taking into account the 30% operation using a duty cycle of 0.3) is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{CDdrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{{{d}_{c}}\times 365}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{0.3\times 365}{4000} \right)}^{2}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9993\text{ or }99.93%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.15.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.15 Result for the computer system reliability.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
Thus the reliability of the computer system after 365 days of operation is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{s}}(365)= &amp;amp; {{R}_{processor}}(365)\cdot {{R}_{harddrive}}(365)\cdot {{R}_{CDdrive}}(365) \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\cdot 0.9948\cdot 0.9993 \\ &lt;br /&gt;
	  = &amp;amp; 0.9747\text{ or }97.47%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;	&lt;br /&gt;
&amp;lt;br&amp;gt;	&lt;br /&gt;
This result can be obtained in BlockSim as shown in Figure DutyCycleExResults.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Load Sharing=&lt;br /&gt;
As presented in earlier chapters, a reliability block diagram (RBD) allows you to graphically represent how the components within a system are reliability-wise connected.  In most cases, independence is assumed across the components within the system.  For example, the failure of component A does not affect the failure of component B.  However, if a system consists of components that are sharing a load, then the assumption of independence no longer holds true.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If one component fails, then the component(s) that are still operating will have to assume the failed unit&#039;s portion of the load.  Therefore, the reliabilities of the surviving unit(s) will change.  Calculating the system reliability is no longer an easy proposition.  In the case of load sharing components, the change of the failure distributions of the surviving components must be known in order to determine the system&#039;s reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate this, consider the a system of two units connected reliability-wise in parallel (Figure 5.16).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-16.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.16 Two units connected reliability-wise in parallel.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Assume that the units must supply an output of 8 volts and that if both units are operational, each unit is to supply 50% of the total output.  If one of the units fails, then the surviving unit supplies 100%.  Furthermore, assume that having to supply the entire load has a negative impact on the reliability characteristics of the surviving unit.  Since the reliability characteristics of the unit change based on whether both or only one is operating, a life distribution along with a life-stress relationship (as discussed in Chapter 3) will be needed to model each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the steps needed, we will create the model starting from raw data.  Assume that a total of 20 units were tested to failure at 7, 10 and 15 volts.  The test data set is presented in the next table.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-17.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
For this example, Units 1 and 2 are the same component.  Therefore, only one set of data was collected.  However, it is possible that the load sharing components in a system may not be the same.  If that were the case, data would need to be collected for each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The data set in Table 1 was analyzed using ReliaSoft&#039;s ALTA software (as shown in Figure 5.17) with the Inverse Power Law as the underlying life-stress relationship and Weibull as the life distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated model parameters,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , are shown next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
\beta = &amp;amp; 1.9239 \\ &lt;br /&gt;
K= &amp;amp; 3.2387\times {{10}^{-7}} \\ &lt;br /&gt;
n= &amp;amp; 3.4226  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:Or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( 3.2387\times {{10}^{-7}}S_{1}^{3.4226}t \right)}^{1.9239}}}}  \ (eqn 20)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1}}(t,{{S}_{1}})=\beta KS_{1}^{n}{{\left( KS_{1}^{n}t \right)}^{\beta -1}}{{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}}  \ (eqn 21 )&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And for this case:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{R}_{2}}(t,{{S}_{2}})  \\ &lt;br /&gt;
{{f}_{1}}(t,{{S}_{1}})= &amp;amp; {{f}_{2}}(t,{{S}_{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.17.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.17 Using ALTA to calculate component parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 5.18 shows a plot of Eqn.20.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Now that the failure properties have been determined using the test data, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t,S)= &amp;amp; {{R}_{1}}(t,{{S}_{1}})\cdot {{R}_{2}}(t,{{S}_{2}}) \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{1}}\left( x,{{S}_{1}} \right)\cdot {{R}_{2}}(x,{{S}_{2}})\cdot \left( \frac{{{R}_{2}}({{t}_{1e}}+(t-x),S)}{{{R}_{2}}({{t}_{1e}},S)} \right)dx \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{2}}\left( x,{{S}_{2}} \right)\cdot {{R}_{1}}(x,{{S}_{1}})\cdot \left( \frac{{{R}_{1}}({{t}_{2e}}+(t-x),S)}{{{R}_{1}}({{t}_{2e}},S)} \right)dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:Where: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{S}_{1}}= &amp;amp; {{P}_{1}}S \\ &lt;br /&gt;
{{S}_{2}}= &amp;amp; {{P}_{2}}S  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-18.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig. 5.18 Reliability curves for different voltage output conditions. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.png|thumb|center|395px]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  is the total load (or required output).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{P}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{P}_{2}}&amp;lt;/math&amp;gt;  are the portion of the total load that each unit supports when both units are operational.  In this case,  &amp;lt;math&amp;gt;{{P}_{1}}={{P}_{2}}=0.5=50%.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{S}_{2}}&amp;lt;/math&amp;gt;  represent the portions of the load that Unit 1 and Unit 2 must support when both units are operational.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{{{1}_{e}}}}&amp;lt;/math&amp;gt;  is the equivalent operating time for Unit 1 if it had been operating at  &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  instead of  &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt; .  A graphical representation of the equivalent time is shown in Figure 5.19, where the curve marked by L represents the low stress (load) and the curve marked by H represents the high stress (load).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  can be calculated by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t)= &amp;amp; {{R}_{1}}({{t}_{1e}}) \\ &lt;br /&gt;
{{e}^{-{{(tKS_{1}^{n})}^{\beta }}}}= &amp;amp; {{e}^{-{{({{t}_{1e}}K{{S}^{n}})}^{\beta }}}} \\ &lt;br /&gt;
tS_{1}^{n}= &amp;amp; {{t}_{1e}}{{S}^{n}} \\ &lt;br /&gt;
{{t}_{1e}}= &amp;amp; t{{\left( \frac{{{S}_{1}}}{S} \right)}^{n}},\text{     }{{S}_{1}}={{P}_{1}}S \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{1e}}=tP_{1}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  can be calculated the same way, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{2}}(t)= &amp;amp; {{R}_{2}}({{t}_{2e}}) \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{2e}}=tP_{2}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this example, the reliability equations for Unit 1 and Unit 2 are the same since they are the same type of component and demonstrate the same failure properties.  In addition, the total output is divided equally between the two units (when both units are operating), so  &amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  will also be the same.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The next step is to determine the reliability of the system after 8,760 hours,  &amp;lt;math&amp;gt;R(t=8,760)&amp;lt;/math&amp;gt; .  Using Eqn. 22 the system reliability is found to be:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t=8760)= &amp;amp; 0.8567 \\ &lt;br /&gt;
= &amp;amp; 85.67%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Load Sharing in BlockSim===&lt;br /&gt;
BlockSim uses this formulation when computing reliabilities of units in a load sharing configuration.  When using the System Reliability Equation window, BlockSim returns a single token for the reliability of units in a load sharing configuration (as well as in the case of standby redundancy, discussed in the next section).  As an example, consider the following RBD with Unit 1 in series with a container that includes two load sharing units.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.2.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
BlockSim will return the system equation as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{System}}=+{{R}_{LS}}\cdot {{R}_{1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{LS}}&amp;lt;/math&amp;gt;  implies a form similar to Eqn. 22.  BlockSim allows for  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -out-of- &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  units in a load sharing configuration.&lt;br /&gt;
&lt;br /&gt;
===Example 5===&lt;br /&gt;
A component has five possible failure modes,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , and the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are interdependent.  The system will fail if mode  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  occurs, mode  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  occurs or two out of the three  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes occur.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  each have a Weibull distribution, with a  &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt;  and 15,000 respectively.  Events  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  each have an exponential distribution with a mean of 10,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If any  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  event occurs (i.e.  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt; ), the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are more likely to occur.  Specifically, the mean times of the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are halved.  Determine the reliability at 1000 hours for this component.&lt;br /&gt;
====Solution to Example 5====&lt;br /&gt;
The first step is to create the RBD.  Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and a load sharing container with the  &amp;lt;math&amp;gt;{{B}_{i}}&amp;lt;/math&amp;gt;  modes must be drawn in series, as illustrated next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.3.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The next step is to define the properties for each block, including those for the container.  Setting the failure distributions for modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is simple.  The more difficult part is setting the properties for the container and the contained blocks.  Based on the problem statement, the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are in a 2-out-of-3 load sharing redundancy.  When all three are working (i.e. when no  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode has occurred), each block has an exponential distribution with &amp;lt;math&amp;gt;\mu=10,000&amp;lt;/math&amp;gt;.  If one  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode occurs, then the two surviving units have an exponential distribution with  &amp;lt;math&amp;gt;\mu =5,000.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assume a Power Life-Stress relationship for the components.  Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\mu }_{1}}= &amp;amp; \frac{1}{KV_{1}^{n}}  \ (eqn 23)\\ &lt;br /&gt;
{{\mu }_{2}}= &amp;amp; \frac{1}{KV_{2}^{n}}  \ (eqn 24)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{1}}=10,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{V}_{1}}=1&amp;lt;/math&amp;gt;  in Eqn. 23 and casting it in terms of  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
10,000= &amp;amp; \frac{1}{K}  \ (eqn 25) \\ &lt;br /&gt;
K = &amp;amp; \frac{1}{10,000}=0.0001  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{2}}=5,000&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{V}_{2}}=1.5&amp;lt;/math&amp;gt;  (because if one fails, then each survivor takes on an additional 0.5 units of load) and Eqn. 25 for  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  in Eqn.24 yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
5,000= &amp;amp; \frac{1}{0.0001\cdot {{(1.5)}^{n}}} \\ &lt;br /&gt;
0.5= &amp;amp; {{(1.5)}^{-n}} \\ &lt;br /&gt;
\ln (0.5)= &amp;amp; -n\ln (1.5) \\ &lt;br /&gt;
n = &amp;amp; 1.7095  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This also could have been computed in ALTA, as shown in Figure 5.20, or with the Load &amp;amp; Life Parameter Experimenter in BlockSim, as shown in Figure 5.21.&lt;br /&gt;
 &lt;br /&gt;
At this point, the parameters for the load sharing units have been computed and can be set, as shown in Figure 5.22.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The next step is to set the weight proportionality factor.  This factor defines the portion of the load that the particular item carries while operating, as well as the load that shifts to the remaining units upon failure of the item.  To illustrate, assume three units (1, 2 and 3) are in a load sharing container with weight proportionality factors of 1, 2 and 3 respectively (and a 1-out-of-3 requirement).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Unit 1 carries  &amp;lt;math&amp;gt;\left( \tfrac{1}{1+2+3} \right)=0.166&amp;lt;/math&amp;gt;  or 16.6% of the total load.&lt;br /&gt;
:•	Unit 2 carries  &amp;lt;math&amp;gt;\left( \tfrac{2}{1+2+3} \right)=0.333&amp;lt;/math&amp;gt;  or 33.3% of the total load.&lt;br /&gt;
:•	Unit 3 carries  &amp;lt;math&amp;gt;\left( \tfrac{3}{1+2+3} \right)=0.50&amp;lt;/math&amp;gt;  or 50% of the total load.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The actual load on each unit then becomes the product of the entire load defined for the container times the portion carried by that unit.  For example, if the container load is 100 lbs, then the portion assigned to Unit 1 will be  &amp;lt;math&amp;gt;100\cdot 0.166=16.6&amp;lt;/math&amp;gt;  lbs.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the current example, all units share the same load and thus have equal weight proportionality factors.  Because these factors are relative, if the same number is used for all three items then the results will be the same.  Thus, weight proportional factor is set equal to 1 for each item.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5.20.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.20 Calculation performed in ALTA.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.21_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.21 Quick Parameter Estimator &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.21_3.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.22 Quick Parameter Estimator results &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.23_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.23 Defining Weight Proportional Factor. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The last properties that need to be defined are the total load and the 2-out-of-3 redundancy.  The total load is dependent on how the parameters were computed.  In this case, total load was assumed to be 3 when the parameters were computed (i.e. the load per item was 1 when all worked and 1.5 when two worked).  This is defined at the container level, set No. of Paths required = 3.&lt;br /&gt;
When all of the parameters have been specified in BlockSim, the reliability at 1,000 hours can be determined.  From the Analytical QCP, this is found to be 98.57%.&lt;br /&gt;
&lt;br /&gt;
=Standby Components=&lt;br /&gt;
&lt;br /&gt;
In the previous section, the case of a system with load sharing components was presented.  This is a form of redundancy with dependent components. That is, the failure of one component affects the failure of the other(s).  This section presents another form of redundancy: standby redundancy.  In standby redundancy the redundant components are set to be under a lighter load condition (or no load) while not needed and under the operating load when they are activated.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In standby redundancy the components are set to have two states: an active state and a standby state.  Components in standby redundancy have two failure distributions, one for each state.  When in the standby state, they have a quiescent (or dormant) failure distribution and when operating, they have an active failure distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case that both quiescent and active failure distributions are the same, the units are in a simple parallel configuration (also called a hot standby configuration).  When the rate of failure of the standby component is lower in quiescent mode than in active mode, that is called a warm standby configuration.  When the rate of failure of the standby component is zero in quiescent mode (i.e. the component cannot fail when in standby), that is called a cold standby configuration.  &lt;br /&gt;
&lt;br /&gt;
===Simple Standby Configuration===&lt;br /&gt;
&lt;br /&gt;
Consider two components in a standby configuration.  Component 1 is the active component with a Weibull failure distribution with parameters  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Component 2 is the standby component.  When Component 2 is operating, it also has a Weibull failure distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Furthermore, assume the following cases for the quiescent distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1:  The quiescent distribution is the same as the active distribution (hot standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2:  The quiescent distribution is a Weibull  distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 2000 (warm standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3: The component cannot fail in quiescent mode (cold standby).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this case, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be obtained using the following equation:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(t)={{R}_{1}}(t)+\underset{0}{\overset{t}{\mathop \int }}\,{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x)\cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}dx   \ (eqn 26)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode, such that: &amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
&amp;lt;br&amp;gt;  &lt;br /&gt;
&amp;lt;math&amp;gt;{{R}_{2;SB}}(x)={{R}_{2;A}}({{t}_{e}})  \ (eqn 27)&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Eqn. 27 can be solved for  &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  and substituted into Eqn.26.&lt;br /&gt;
Figure 5.24 illustrates the example as entered in BlockSim using a standby container.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5.24.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.24 Standby container.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The active and standby blocks are within a container, which is used to specify standby redundancy.  Since the standby component has two distributions (active and quiescent), the Block Properties window of the standby block has two pages for specifying each one.  Figures 5.24 and 5.26 illustrate these pages.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system reliability results for 1000 hours are given in the following table:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-24.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Note that even though the  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  value for the quiescent distribution is the same as in the active distribution, it is possible that the two can be different. That is, the failure modes present during the quiescent mode could be different from the modes present during the active mode.  In that sense, the two distribution types can be different as well (e.g. lognormal when quiescent and Weibull when active).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In many cases when considering standby systems, a switching device may also be present that switches from the failed active component to the standby component.  The reliability of the switch can also be incorporated into Eqn.26, as presented in the next section.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim&#039;s System Reliability Equation window returns a single token for the reliability of units in a standby configuration.  This is the same as the load sharing case presented in the previous section. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image: Fig 5.25.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.25 Defining the active failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.26.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.26 Defining the quiescent failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Reliability of Standby Systems with a Switching Device===&lt;br /&gt;
&lt;br /&gt;
In many cases when dealing with standby systems, a switching device is present that will switch to the standby component when the active component fails.  Therefore, the failure properties of the switch must also be included in the analysis.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.26.2.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In most cases when the reliability of a switch is to be included in the analysis, two probabilities can be considered.  The first and most common one is the probability of the switch performing the action (i.e. switching) when requested to do so.  This is called Switch Probability per Request in BlockSim and is expressed as a static probability (e.g. 90%).  The second probability is the quiescent reliability of the switch.  This is the reliability of the switch as it ages (e.g. the switch might wear out with age due to corrosion, material degradation, etc.). Thus it is possible for the switch to fail before the active component fails.  However, a switch failure does not cause the system to fail, but rather causes the system to fail only if the switch is needed and the switch has failed.  For example, if the active component does not fail until the mission end time and the switch fails, then the system does not fail.  However, if the active component fails and the switch has also failed, then the system cannot be switched to the standby component and it therefore fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In analyzing standby components with a switching device, either or both failure probabilities (during the switching or while waiting to switch) can be considered for the switch, since each probability can represent different failure modes.  For example, the switch probability per request may represent software-related issues or the probability of detecting the failure of an active component, and the quiescent probability may represent wear-out type failures of the switch.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the formulation, consider the previous example that assumes perfect switching.  To examine the effects of including an imperfect switch, assume that when the active component fails there is a 90% probability that the switch will switch from the active component to the standby component.  In addition, assume that the switch can also fail due to a wear-out failure mode described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.7 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 5000.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , is given by the following equation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t)= &amp;amp; {{R}_{1}}(t) \\ &lt;br /&gt;
&amp;amp; +\underset{0}{\overset{t}{\mathop \int }}\,\{{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x) \\ &lt;br /&gt;
&amp;amp; \cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}\cdot {{R}_{SW;Q}}(x)\cdot {{R}_{SW;REQ}}(x)\}dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;Q}}&amp;lt;/math&amp;gt;  is the quiescent reliability of the switch.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;REQ}}&amp;lt;/math&amp;gt;  is the switch probability per request.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
This problem can be solved in BlockSim by including these probabilities in the container&#039;s properties, as shown in Figures 5.27 and 5.28.  In BlockSim, the standby container is acting as the switch.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.28_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.27 Standby container (switch) failure probabilities while attempting to switch &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.27.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.28 Standby container (switch) failure distribution while waiting to switch &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that there are additional properties that can be specified in BlockSim for a switch, such as Switch Restart Probability, No. of Restarts and Switch Delay Time.  In many applications, the switch is re-tested (or re-cycled) if it fails to switch the first time.  In these cases, it might be possible that it switches in the second or third, or  &amp;lt;math&amp;gt;{{n}^{th}}&amp;lt;/math&amp;gt; attempt.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The Switch Restart Probability specifies each additional attempt&#039;s probability of successfully switching and the Finite Restarts specifies the total number of attempts.  Note that the Switch Restart Probability specifies the probability of success of each trial (or attempt).  The probability of success of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  consecutive trials is calculated by BlockSim using the binomial distribution and this probability is then incorporated into Eqn. (stb2a).  The Switch Delay Time property is related to repairable systems and is considered in BlockSim only when using simulation.  When using the analytical solution (i.e. for a non-repairable system), this property is ignored.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving the analytical solution (as given by Eqn. stb2a), the following results are obtained.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-30.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From the table above, it can be seen that the presence of a switching device has a significant effect on the reliability of a standby system.  It is therefore important when modeling standby redundancy to incorporate the switching device reliability properties.  It should be noted that this methodology is not the same as treating the switching device as another series component with the standby subsystem.  This would be valid only if the failure of the switch resulted in the failure of system (e.g. switch failing open).  In Eqn. (stb2a), the Switch Probability per Request and quiescent probability are present only in the second term of the equation.  Treating these two failure modes as a series configuration with the standby subsystem would imply that they are also present when the active component is functioning (i.e. first term of Eqn. stb2a).  This is invalid and would result in the underestimation of the reliability of the system.  In other words, these two failure modes become significant only when the active component fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As an example, and if we consider the warm standby case, the reliability of the system without the switch is 70.57% at 1000 hours.  If the system was modeled so that the switching device was in series with the warm standby subsystem, the result would have been:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(1000)= &amp;amp; {{R}_{Standby}}(1000)\cdot {{R}_{sw,Q(1000)}}\cdot {{R}_{sw,req}} \\ &lt;br /&gt;
= &amp;amp; 0.7057\cdot 0.9372\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.5952  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case where a switch failure mode causes the standby subsystem to fail, then this mode can be modeled as an individual block in series with the standby subsystem.&lt;br /&gt;
&lt;br /&gt;
===Example 6===&lt;br /&gt;
Consider a car with four new tires and a full-size spare.  Assume the following failure characteristics:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires follow a Weibull distribution with a  ..  and an  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  40,000 miles while on the car due to wear.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires also have a probability of failing due to puncture or other causes.  For this, assume a constant rate for this occurrence with a probability of 1 every 50,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	When not on the car (i.e. is a spare), a tire&#039;s probability of failing also has a Weibull distribution with a  &amp;lt;math&amp;gt;\beta =&amp;lt;/math&amp;gt;  2 and  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  120,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Assume a mission of 1,000 miles.  If a tire fails during this trip, it will be replaced with the spare.  However, the spare will not be repaired during the trip.  In other words, the trip will continue with the spare on the car and if the spare fails the system will fail.  Determine the probability of system failure.&lt;br /&gt;
====Solution to Example 6====&lt;br /&gt;
Active failure distribution for tires:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to wear-out, Weibull  &amp;lt;math&amp;gt;\beta =4&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =40,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to random puncture, exponential  &amp;lt;math&amp;gt;\mu =50,000.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	The quiescent failure distribution is a Weibull distribution with &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =120,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The block diagram for each tire has two blocks in series, one block representing the wear-out mode and the other the random puncture mode, as shown next:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:small5.gif|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
There are five tires, four active and one standby (represented in the diagram by a standby container with a 4-out-of-5 requirement), as shown next: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BStirecontainer.png|thumb|center|400px|]]&lt;br /&gt;
 &lt;br /&gt;
For the standby Wear block, set the active failure and the quiescent distributions, but for the Puncture block, set only the active puncture distribution (because the tire cannot fail due to puncture while stored).  Using BlockSim, the probability of system failure is found to be 0.003 or 0.3%. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Note Regarding Numerical Integration Solutions=&lt;br /&gt;
&lt;br /&gt;
Load sharing and standby solutions in BlockSim are performed using numerical integration routines.  As with any numerical analysis routine, the solution error depends on the number of iterations performed, the step size chosen and related factors, plus the behavior of the underlying function.  By default, BlockSim uses a certain set of preset factors.  In general, these defaults are sufficient for most problems.  If a higher precision or verification of the precision for a specific problem is required, BlockSim&#039;s preset options can be modified and/or the integration error can be assessed using the Integration Parameters... option for each container.  For more details, you can refer to the documentation on the Algorithm Setup window in the BlockSim 7 User&#039;s Guide.&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15564</id>
		<title>Time-Dependent System Reliability (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15564"/>
		<updated>2012-02-13T22:13:47Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Reliability of Standby Systems with a Switching Device */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|5}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the previous chapter, different system configuration types were examined, as well as different methods for obtaining the system&#039;s reliability function analytically.  Because the reliabilities in the problems presented were treated as probabilities (e.g.  &amp;lt;math&amp;gt;P(A)&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; ), the reliability values and equations presented were referred to as static (not time-dependent).  Thus, in the prior chapter, the life distributions of the components were not incorporated in the process of calculating the system reliability.  In this chapter, time dependency in the reliability function will be introduced.  We will develop the models necessary to observe the reliability over the life of the system, instead of at just one point in time.  In addition, performance measures such as failure rate, MTTF and warranty time will be estimated for the entire system.  The methods of obtaining the reliability function analytically remain identical to the ones presented in the previous chapter, with the exception that the reliabilities will be functions of time.  In other words, instead of dealing with  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , we will use  &amp;lt;math&amp;gt;{{R}_{i}}(t)&amp;lt;/math&amp;gt; .  All examples in this chapter assume that no repairs are performed on the components.  &lt;br /&gt;
&lt;br /&gt;
=Analytical Life Predictions=&lt;br /&gt;
The analytical approach presented in the prior chapter involved the determination of a mathematical expression that describes the reliability of the system, expressed in terms of the reliabilities of its components.  So far we have estimated only static system reliability (at a fixed time).  For example, in the case of a system with three components in series, the system&#039;s reliability equation was given by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}  \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The values of  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  were given for a common time and the reliability of the system was estimated for that time.  However, since the component failure characteristics can be described by distributions, the system reliability is actually time-dependent.  In this case, Eqn. (1) can be rewritten as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability of the system for any mission time can now be estimated.  Assuming a Weibull life distribution for each component, Eqn. (1) can now be expressed in terms of each component&#039;s reliability function, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the same manner, any life distribution can be substituted into the system reliability equation.  Suppose that the times-to-failure of the first component are described with a Weibull distribution, the times-to-failure of the second component with an exponential distribution and the times-to-failure of the third component with a normal distribution.  Then Eqn. (1) can be written as:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot \left[ 1-\Phi \left( \frac{t-{{\mu }_{3}}}{{{\sigma }_{3}}} \right) \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It can be seen that the biggest challenge is in obtaining the system&#039;s reliability function in terms of component reliabilities, which has already been discussed in depth.  Once this has been achieved, calculating the reliability of the system for any mission duration is just a matter of substituting the corresponding component reliability functions into the system reliability equation.&lt;br /&gt;
===Advantages of the Analytical Method===&lt;br /&gt;
The primary advantage of the analytical solution is that it produces a mathematical expression that describes the reliability of the system.  Once the system&#039;s reliability function has been determined, other calculations can then be performed to obtain metrics of interest for the system. Such calculations include:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of warranty periods.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s failure rate.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s MTTF.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, optimization and reliability allocation techniques can be used to aid engineers in their design improvement efforts.  Another advantage in using analytical techniques is the ability to perform static calculations and analyze systems with a mixture of static and time-dependent components.  Finally, the reliability importance of components over time can be calculated with this methodology.&lt;br /&gt;
===Disadvantages of the Analytical Method===&lt;br /&gt;
The biggest disadvantage of the analytical method is that formulations can become very complicated.  The more complicated a system is, the larger and more difficult it will be to analytically formulate an expression for the system&#039;s reliability.  For particularly detailed systems this process can be quite time-consuming, even with the use of computers.  Furthermore, when the maintainability of the system or some of its components must be taken into consideration, analytical solutions become intractable.  In these situations, the use of simulation methods may be more advantageous than attempting to develop a solution analytically.  Simulation methods are presented in later chapters.&lt;br /&gt;
===Looking at a Simple &#039;&#039;Complex&#039;&#039; System Analytically===&lt;br /&gt;
&lt;br /&gt;
The complexity involved in an analytical solution can be best illustrated by looking at the simple &#039;&#039;complex&#039;&#039; system with 15 components, as shown in Figure 5.1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1.png|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.1 An RBD of a complex system.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system reliability for this system (computed using BlockSim) is shown next.  The first solution is provided using BlockSim&#039;s symbolic solution.  In symbolic mode, BlockSim breaks the equation into segments, identified by tokens, that need to be substituted into the final system equation for a complete solution.  This creates algebraic solutions that are more compact than if the substitutions were made.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; D2\cdot D3\cdot {{R}_{L}} \\ &lt;br /&gt;
D3= &amp;amp; +{{R}_{K}}\cdot IK \\ &lt;br /&gt;
IK= &amp;amp; +{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}} \\ &lt;br /&gt;
D2 = &amp;amp; +{{R}_{A}}\cdot {{R}_{E}}\cdot IE \\ &lt;br /&gt;
IE = &amp;amp; -D1\cdot {{R}_{M}}\cdot {{R}_{N}}+{{R}_{M}}\cdot {{R}_{N}}+D1 \\ &lt;br /&gt;
D1 = &amp;amp; +{{R}_{D}}\cdot ID \\ &lt;br /&gt;
ID = &amp;amp; -{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting the terms yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; {{R}_{A}}\cdot {{R}_{E}}\cdot {{R}_{L}}\cdot {{R}_{K}} \\ &lt;br /&gt;
&amp;amp; \cdot \{({{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})\cdot {{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}}-{{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}\} \\ &lt;br /&gt;
&amp;amp; \cdot \{{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}}\}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
BlockSim&#039;s automatic algebraic simplification would yield the following format for the above solution: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; (({{R}_{A}}\cdot {{R}_{E}}(-({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})){{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})))) \\ &lt;br /&gt;
&amp;amp; ({{R}_{K}}({{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; +RI\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}})){{R}_{L}})  \ (eqn 2)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this equation, each  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  represents the reliability function of a block.  For example, if  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  has a Weibull distribution, then each  &amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;  and so forth.  Substitution of each component&#039;s reliability function in Eqn.2 will result in an analytical expression for the system reliability as a function of time, or  &amp;lt;math&amp;gt;{{R}_{s}}(t)&amp;lt;/math&amp;gt; , which is the same as  &amp;lt;math&amp;gt;(1-cd{{f}_{System}}).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Obtaining Other Functions of Interest===&lt;br /&gt;
Once the system reliability equation (or the cumulative density function,  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt; ) has been determined, other functions and metrics of interest can be derived.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Consider the following simple system:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-2.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.2 Simple two-component system. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Furthermore, assume that component 1 follows an exponential distribution with a mean of 10,000 (&amp;lt;math&amp;gt;\mu =10,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;\lambda =1/10,000)&amp;lt;/math&amp;gt;  and component 2 follows a Weibull distribution with  &amp;lt;math&amp;gt;\beta =6&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt; .  The reliability equation of this system is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}  \ (eqn 3) &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{F}_{S}}(t)= &amp;amp; 1-({{R}_{1}}(t)\cdot {{R}_{2}}(t)) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \right) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
{{analytical system pdf}}&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability====&lt;br /&gt;
Conditional reliability is the probability of a system successfully completing another mission following the successful completion of a previous mission.  The time of the previous mission and the time for the mission to be undertaken must be taken into account for conditional reliability calculations.  The system&#039;s conditional reliability function is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(T,t)=\frac{R(T+t)}{R(T)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Eqn.6 gives the reliability for a new mission of duration  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  having already accumulated  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  hours of operation up to the start of this new mission. The system is evaluated to assure that it will start the next mission successfully.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2, the reliability for mission of  &amp;lt;math&amp;gt;t=1,000&amp;lt;/math&amp;gt;  hours, having an age of  &amp;lt;math&amp;gt;T=500&amp;lt;/math&amp;gt;  hours, is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(T=500,t=1000)= &amp;amp; \frac{R(T+t)}{R(T)} \\ &lt;br /&gt;
= &amp;amp; \frac{R(1500)}{R(500)} \\ &lt;br /&gt;
= &amp;amp; \frac{{{e}^{-\tfrac{1500}{10,000}}}\cdot {{e}^{-{{\left( \tfrac{1500}{10,000} \right)}^{6}}}}}{{{e}^{-\tfrac{500}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{500}{10,000} \right)}^{6}}}}} \\ &lt;br /&gt;
= &amp;amp; 0.9048=90.48%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.3.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.4.png|thumb|center|400px|]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability for Components====&lt;br /&gt;
&lt;br /&gt;
Now in this formulation, it was assumed that the accumulated age was equivalent for both units. That is, both started life at zero and aged to 500.  It is possible to consider an individual component that has already accumulated some age (used component) in the same formulation.  To illustrate this, assume that component 2 started life with an age of T=100.  Then the reliability equation of the system, as given in Eqn.3, would need to be modified to include a conditional term for 2, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{R}_{1}}(t)\cdot \frac{{{R}_{2}}({{T}_{2}}+t)}{{{R}_{2}}({{T}_{2}})} \ (eqn 7)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In BlockSim, the start age input box may be used to specify a starting age greater than zero.&lt;br /&gt;
{{system failure rate analytical}}&lt;br /&gt;
&lt;br /&gt;
====System Mean Life (Mean Time To Failure)====&lt;br /&gt;
The mean life (or mean time to failure, MTTF) can be obtained by integrating the system reliability function from zero to infinity: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;MTTF=\int_{0}^{\infty }{{R}_{s}}\left( t \right)dt   \ (eqn 10)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mean time is a performance index and does not provide any information about the behavior of the failure distribution of the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)dt \\ &lt;br /&gt;
= &amp;amp; 5978.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Warranty Period and BX Life====&lt;br /&gt;
Sometimes it is desirable to know the time value associated with a certain reliability.  Warranty periods are often calculated by determining what percentage of the failure population can be covered financially and estimating the time at which this portion of the population will fail.  Similarly, engineering specifications may call for a certain BX life, which also represents a time period during which a certain proportion of the population will fail.  For example, the B10 life is the time in which 10% of the population will fail.  &lt;br /&gt;
This is obtained by setting  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  to the desired value and solving for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}\left( t \right)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To compute the time by which reliability would be equal to 90%, Eqn.11 is recast as follows and solved for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this case,  &amp;lt;math&amp;gt;t=1053.59&amp;lt;/math&amp;gt; .  Equivalently, the B10 life for this system is also  &amp;lt;math&amp;gt;1053.59&amp;lt;/math&amp;gt; .&lt;br /&gt;
Except for some trivial cases, a closed form solution for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  cannot be obtained.   Thus, it is necessary to solve for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  using numerical methods.  BlockSim uses numerical methods.&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
Consider a system consisting of three exponential units in series with the following failure rates (in failures per hour):  &amp;lt;math&amp;gt;{{\lambda }_{1}}&amp;lt;/math&amp;gt;  = 0.0002,  &amp;lt;math&amp;gt;{{\lambda }_{2}}&amp;lt;/math&amp;gt;  = 0.0005 and  &amp;lt;math&amp;gt;{{\lambda }_{3}}&amp;lt;/math&amp;gt;  = 0.0001.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the reliability equation for the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the reliability of the system after 150 hours of operation?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s failure rate equation.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the MTTF for the system?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What should the warranty period be for a 90% reliability?&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The analytical expression for the reliability of the system is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\lambda }_{1}}t}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot {{e}^{-{{\lambda }_{1}}t}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
At 150 hours of operation, the reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{e}^{-(0.0002+0.0005+0.0001)150}} \\ &lt;br /&gt;
= &amp;amp; 0.8869\text{ or }88.69%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.12 is taken with respect to time, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; -\frac{d[{{R}_{s}}(t)]}{dt} \\ &lt;br /&gt;
= &amp;amp; -\frac{d\left[ {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}} \right]}{dt} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
:•	The system&#039;s failure rate can now be obtained simply by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn. 13 by the system&#039;s reliability function given in Eqn.12, and:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}\left( t \right)= &amp;amp; \frac{{{f}_{s}}\left( t \right)}{{{R}_{s}}\left( t \right)} \\ &lt;br /&gt;
= &amp;amp; \frac{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}}{{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}) \\ &lt;br /&gt;
= &amp;amp; 0.0008\text{ }fr/hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Combining Eqn.10 and Eqn. 12, the system&#039;s MTTF can be obtained:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }{{R}_{s}}\left( t \right)dt \\ &lt;br /&gt;
= &amp;amp; \int_{0}^{\infty }{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}dt \\ &lt;br /&gt;
= &amp;amp; \frac{1}{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})} \\ &lt;br /&gt;
= &amp;amp; 1250\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Solving Eqn. 12 with respect to time will yield the corresponding warranty period for a 90% reliability.  In this case, the system reliability equation is simple and a closed form solution exists.  The warranty time can now be found by solving:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
t= &amp;amp; -\frac{\ln (R)}{{{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}} \\ &lt;br /&gt;
= &amp;amp; -\frac{\ln (0.9)}{0.0008} \\ &lt;br /&gt;
= &amp;amp; 131.7\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the warranty period should be 132 hours.&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
Consider the system shown in Figure 5.5.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.5.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.5 Complex bridge system in Example 2. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Components  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  through  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  are Weibull distributed with  &amp;lt;math&amp;gt;\beta =1.2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1230&amp;lt;/math&amp;gt;  hours.  The starting and ending blocks cannot fail.  &amp;lt;br&amp;gt;&lt;br /&gt;
Determine the following:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability equation for the system and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s failure rate equation and the corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The warranty time for a 90% reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability for a 200-hour mission, if it is known that the system has already successfully operated for 200 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The first step is to obtain the reliability function for the system.  The methods described in the previous chapter can be employed, such as the event space or path-tracing methods.  Using BlockSim, the following reliability equation is obtained:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; ({{R}_{Start}}\cdot {{R}_{End}}(2{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}))  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that since the starting and ending blocks cannot fail,  &amp;lt;math&amp;gt;{{R}_{Start}}=1&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{End}}=1,&amp;lt;/math&amp;gt;  Eqn.14 can be reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; 2\cdot {{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  is the reliability equation for Component A, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{B}}&amp;lt;/math&amp;gt;  is the reliability equation for Component  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; , etc.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the components in this example are identical, the system reliability equation can be further reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2R{{(t)}^{2}}+2R{{(t)}^{3}}-5R{{(t)}^{4}}+2R{{(t)}^{5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Or, in terms of the failure distribution:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.6.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.6.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.6 Reliability plot for the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.18 is taken with respect to time, resulting in: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; 4\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
&amp;amp; -20\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  can now be plotted for different time values,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , as shown in Figure 5.7.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system&#039;s failure rate can now be obtained by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn.19 by the system&#039;s reliability function given in Eqn.18, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}(t)= &amp;amp; \frac{4\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}} \\ &lt;br /&gt;
&amp;amp; +\frac{-20\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.8.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.7.gif|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.7 &#039;&#039;pdf&#039;&#039; plot for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of the system is obtained by integrating the system&#039;s reliability function given by Eqn. 18 from time zero to infinity, as given by Eqn. 10.  Using BlockSim&#039;s Analytical QCP, an  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of 1007.8 hours is calculated, as shown in Figure 5.9.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The warranty time can be obtained by solving Eqn. 18 with respect to time for a system reliability  &amp;lt;math&amp;gt;{{R}_{s}}=0.9&amp;lt;/math&amp;gt; .  Using the Analytical QCP and selecting the &amp;lt;br&amp;gt;&lt;br /&gt;
Warranty Time option, a time of 372.72 hours is obtained, as shown in Figure 5.10.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.8.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.8 Failure rate for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.9.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.9 MTTF of the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.10.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.10 Time at which &#039;&#039;R&#039;&#039;=0.9 or 90% for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.11.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.11 Conditional reliability calculation for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, the conditional reliability can be obtained using Eqn.6 and Eqn.18, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(200,200)= &amp;amp; \frac{R(400)}{R(200)} \\ &lt;br /&gt;
= &amp;amp; \frac{0.883825}{0.975321} \\ &lt;br /&gt;
= &amp;amp; 0.906189  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated using BlockSim&#039;s Analytical QCP, as shown in Figure 5.11.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Approximating the System CDF=&lt;br /&gt;
&lt;br /&gt;
In many cases, it is valuable to fit a distribution that represents the system&#039;s times-to-failure.  This can be useful when the system is part of a larger assembly and may be used for repeated calculations or in calculations for other systems.  In cases such as this, it can be useful to characterize the system&#039;s behavior by fitting a distribution to the overall system and calculating parameters for this distribution.   This is equivalent to fitting a single distribution to describe  &amp;lt;math&amp;gt;{{R}_{S}}(t&amp;lt;/math&amp;gt; ).  In essence, it is like reducing the entire system to a component in order to simplify calculations.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
To compute an approximate reliability function for this system,  &amp;lt;math&amp;gt;{{R}_{A}}(t)\simeq {{R}_{S}}(t)&amp;lt;/math&amp;gt; , one would compute  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  pairs of reliability and time values and then fit a single distribution to the data, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 10,396.7)=10% \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 9,361.9)=20% \\ &lt;br /&gt;
&amp;amp; ... \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 1,053.6)=90%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A single distribution,  &amp;lt;math&amp;gt;{{R}_{A}}(t)&amp;lt;/math&amp;gt; , that approximates  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  can now be computed from these pairs using life data analysis methods.  If using the Weibull++ software, one would enter the values as free form data.&lt;br /&gt;
&lt;br /&gt;
===Example 3===&lt;br /&gt;
Compute a single Weibull distribution approximation for the system in Example 2.&lt;br /&gt;
====Solution to Example 3====&lt;br /&gt;
The system in the previous example, shown in Figure 5.5, can be approximated by use of a 2-parameter Weibull distribution with  &amp;lt;math&amp;gt;\beta =2.02109&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1123.51&amp;lt;/math&amp;gt; .  In BlockSim, this is accomplished by representing the entire system as one distribution by going to the Distribution Fit window Figure 5.12. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.13.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.12 Representing a system with a distribution.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.12.PNG|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.13 Distribution Fitting window.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
by clicking the Distribution Fit Window, the Distribution Estimator window will appear (Figure 5.13).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this window you can select a distribution to represent the data. BlockSim will then generate a number of system failure times based on the system&#039;s reliability function. The system&#039;s reliability function can be used to solve for a time value associated with that unreliability value. The distribution of the generated time values can then be fitted to a probability distribution function.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a value of  &amp;lt;math&amp;gt;F(t)=0.11&amp;lt;/math&amp;gt; .  Using the system&#039;s reliability equation and solving for time, the corresponding time-to-failure for a 0.11 unreliability can be calculated.  &amp;lt;br&amp;gt;&lt;br /&gt;
For the system of Example 2, the time for a 0.11 unreliability is 389.786 hours.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
When enough points have been generated, the selected distribution will be fitted to this data set and the distribution&#039;s parameters will be returned.  In addition, if ReliaSoft&#039;s Weibull++ is installed, the generated data can be viewed/analyzed using a Weibull++ instance, as shown in Figure 5.14.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.14.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.14 Using Weibull++ to calculate distribution parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
It is recommended that the analyst examine the fit to ascertain the applicability of the approximation.&lt;br /&gt;
&lt;br /&gt;
=Duty Cycle=&lt;br /&gt;
&lt;br /&gt;
Components of a system may not operate continuously during a system&#039;s mission, or may be subjected to loads greater or lesser than the rated loads during system operation.  To model this, a factor called the Duty Cycle ( &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt; ) is used.  The duty cycle may also be used to account for changes in environmental stress, such as temperature changes, that may effect the operation of a component.  The duty cycle is a positive value, with a default value of 1 representing continuous operation at rated load, and any values other than 1 representing other load values with respect to the rated load value (or total operating time).   A duty cycle value higher than 1 indicates a load in excess of the rated value.  A duty cycle value lower than 1 indicates that the component is operating at a load lower than the rated load or not operating continuously during the system&#039;s mission.  For instance, a duty cycle of 0.5 may be used for a component that operates only half of the time during the system&#039;s mission.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The reliability metrics for a component with a duty cycle are calculated as follows. Let  &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt;  represent the duty cycle during a particular mission of the component,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  represent the mission time and  &amp;lt;math&amp;gt;{t}&#039;&amp;lt;/math&amp;gt;  represent the accumulated age. Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{t}&#039;={{d}_{c}}\times t&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability equation for the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R({t}&#039;)=R({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The component &#039;&#039;pdf&#039;&#039; is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;f({t}&#039;)=-\frac{d(R({t}&#039;))}{dt}=-\frac{d(R({{d}_{c}}\times t))}{dt}={{d}_{c}}f({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The failure rate of the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\lambda ({t}&#039;)=\frac{f({t}&#039;)}{R({t}&#039;)}=\frac{{{d}_{c}}f({{d}_{c}}\times t)}{R({{d}_{c}}\times t)}={{d}_{c}}\lambda ({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Example 4===&lt;br /&gt;
Consider a computer system with three components: a processor, a hard drive and a CD drive in series as shown next.  Assume that all three components follow a Weibull failure distribution with the parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}=1.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=5000&amp;lt;/math&amp;gt;  for the processor,  &amp;lt;math&amp;gt;{{\beta }_{2}}=2.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{2}}=3000&amp;lt;/math&amp;gt;  for the hard drive, and  &amp;lt;math&amp;gt;{{\beta }_{3}}=2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=4000&amp;lt;/math&amp;gt;  for the CD drive.  Determine the reliability of the computer system after one year (365 days) of operation, assuming that the CD drive is used only 30% of the time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5ex4.png|thumb|center|300px|]]&lt;br /&gt;
&lt;br /&gt;
====Solution to Example 4====&lt;br /&gt;
The reliability of the processor after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  {{R}_{processor}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{5000} \right)}^{1.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\text{ or }98.05%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the hard drive after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{harddrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{3000} \right)}^{2.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9948\text{ or }99.48%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the CD drive after 365 days of operation (taking into account the 30% operation using a duty cycle of 0.3) is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{CDdrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{{{d}_{c}}\times 365}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{0.3\times 365}{4000} \right)}^{2}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9993\text{ or }99.93%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.15.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.15 Result for the computer system reliability.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
Thus the reliability of the computer system after 365 days of operation is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{s}}(365)= &amp;amp; {{R}_{processor}}(365)\cdot {{R}_{harddrive}}(365)\cdot {{R}_{CDdrive}}(365) \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\cdot 0.9948\cdot 0.9993 \\ &lt;br /&gt;
	  = &amp;amp; 0.9747\text{ or }97.47%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;	&lt;br /&gt;
&amp;lt;br&amp;gt;	&lt;br /&gt;
This result can be obtained in BlockSim as shown in Figure DutyCycleExResults.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Load Sharing=&lt;br /&gt;
As presented in earlier chapters, a reliability block diagram (RBD) allows you to graphically represent how the components within a system are reliability-wise connected.  In most cases, independence is assumed across the components within the system.  For example, the failure of component A does not affect the failure of component B.  However, if a system consists of components that are sharing a load, then the assumption of independence no longer holds true.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If one component fails, then the component(s) that are still operating will have to assume the failed unit&#039;s portion of the load.  Therefore, the reliabilities of the surviving unit(s) will change.  Calculating the system reliability is no longer an easy proposition.  In the case of load sharing components, the change of the failure distributions of the surviving components must be known in order to determine the system&#039;s reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate this, consider the a system of two units connected reliability-wise in parallel (Figure 5.16).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-16.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.16 Two units connected reliability-wise in parallel.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Assume that the units must supply an output of 8 volts and that if both units are operational, each unit is to supply 50% of the total output.  If one of the units fails, then the surviving unit supplies 100%.  Furthermore, assume that having to supply the entire load has a negative impact on the reliability characteristics of the surviving unit.  Since the reliability characteristics of the unit change based on whether both or only one is operating, a life distribution along with a life-stress relationship (as discussed in Chapter 3) will be needed to model each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the steps needed, we will create the model starting from raw data.  Assume that a total of 20 units were tested to failure at 7, 10 and 15 volts.  The test data set is presented in the next table.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-17.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
For this example, Units 1 and 2 are the same component.  Therefore, only one set of data was collected.  However, it is possible that the load sharing components in a system may not be the same.  If that were the case, data would need to be collected for each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The data set in Table 1 was analyzed using ReliaSoft&#039;s ALTA software (as shown in Figure 5.17) with the Inverse Power Law as the underlying life-stress relationship and Weibull as the life distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated model parameters,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , are shown next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
\beta = &amp;amp; 1.9239 \\ &lt;br /&gt;
K= &amp;amp; 3.2387\times {{10}^{-7}} \\ &lt;br /&gt;
n= &amp;amp; 3.4226  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:Or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( 3.2387\times {{10}^{-7}}S_{1}^{3.4226}t \right)}^{1.9239}}}}  \ (eqn 20)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1}}(t,{{S}_{1}})=\beta KS_{1}^{n}{{\left( KS_{1}^{n}t \right)}^{\beta -1}}{{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}}  \ (eqn 21 )&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And for this case:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{R}_{2}}(t,{{S}_{2}})  \\ &lt;br /&gt;
{{f}_{1}}(t,{{S}_{1}})= &amp;amp; {{f}_{2}}(t,{{S}_{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.17.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.17 Using ALTA to calculate component parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 5.18 shows a plot of Eqn.20.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Now that the failure properties have been determined using the test data, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t,S)= &amp;amp; {{R}_{1}}(t,{{S}_{1}})\cdot {{R}_{2}}(t,{{S}_{2}}) \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{1}}\left( x,{{S}_{1}} \right)\cdot {{R}_{2}}(x,{{S}_{2}})\cdot \left( \frac{{{R}_{2}}({{t}_{1e}}+(t-x),S)}{{{R}_{2}}({{t}_{1e}},S)} \right)dx \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{2}}\left( x,{{S}_{2}} \right)\cdot {{R}_{1}}(x,{{S}_{1}})\cdot \left( \frac{{{R}_{1}}({{t}_{2e}}+(t-x),S)}{{{R}_{1}}({{t}_{2e}},S)} \right)dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:Where: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{S}_{1}}= &amp;amp; {{P}_{1}}S \\ &lt;br /&gt;
{{S}_{2}}= &amp;amp; {{P}_{2}}S  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-18.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig. 5.18 Reliability curves for different voltage output conditions. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.png|thumb|center|395px]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  is the total load (or required output).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{P}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{P}_{2}}&amp;lt;/math&amp;gt;  are the portion of the total load that each unit supports when both units are operational.  In this case,  &amp;lt;math&amp;gt;{{P}_{1}}={{P}_{2}}=0.5=50%.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{S}_{2}}&amp;lt;/math&amp;gt;  represent the portions of the load that Unit 1 and Unit 2 must support when both units are operational.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{{{1}_{e}}}}&amp;lt;/math&amp;gt;  is the equivalent operating time for Unit 1 if it had been operating at  &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  instead of  &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt; .  A graphical representation of the equivalent time is shown in Figure 5.19, where the curve marked by L represents the low stress (load) and the curve marked by H represents the high stress (load).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  can be calculated by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t)= &amp;amp; {{R}_{1}}({{t}_{1e}}) \\ &lt;br /&gt;
{{e}^{-{{(tKS_{1}^{n})}^{\beta }}}}= &amp;amp; {{e}^{-{{({{t}_{1e}}K{{S}^{n}})}^{\beta }}}} \\ &lt;br /&gt;
tS_{1}^{n}= &amp;amp; {{t}_{1e}}{{S}^{n}} \\ &lt;br /&gt;
{{t}_{1e}}= &amp;amp; t{{\left( \frac{{{S}_{1}}}{S} \right)}^{n}},\text{     }{{S}_{1}}={{P}_{1}}S \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{1e}}=tP_{1}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  can be calculated the same way, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{2}}(t)= &amp;amp; {{R}_{2}}({{t}_{2e}}) \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{2e}}=tP_{2}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this example, the reliability equations for Unit 1 and Unit 2 are the same since they are the same type of component and demonstrate the same failure properties.  In addition, the total output is divided equally between the two units (when both units are operating), so  &amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  will also be the same.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The next step is to determine the reliability of the system after 8,760 hours,  &amp;lt;math&amp;gt;R(t=8,760)&amp;lt;/math&amp;gt; .  Using Eqn. 22 the system reliability is found to be:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t=8760)= &amp;amp; 0.8567 \\ &lt;br /&gt;
= &amp;amp; 85.67%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Load Sharing in BlockSim===&lt;br /&gt;
BlockSim uses this formulation when computing reliabilities of units in a load sharing configuration.  When using the System Reliability Equation window, BlockSim returns a single token for the reliability of units in a load sharing configuration (as well as in the case of standby redundancy, discussed in the next section).  As an example, consider the following RBD with Unit 1 in series with a container that includes two load sharing units.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.2.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
BlockSim will return the system equation as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{System}}=+{{R}_{LS}}\cdot {{R}_{1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{LS}}&amp;lt;/math&amp;gt;  implies a form similar to Eqn. 22.  BlockSim allows for  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -out-of- &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  units in a load sharing configuration.&lt;br /&gt;
&lt;br /&gt;
===Example 5===&lt;br /&gt;
A component has five possible failure modes,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , and the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are interdependent.  The system will fail if mode  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  occurs, mode  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  occurs or two out of the three  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes occur.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  each have a Weibull distribution, with a  &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt;  and 15,000 respectively.  Events  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  each have an exponential distribution with a mean of 10,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If any  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  event occurs (i.e.  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt; ), the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are more likely to occur.  Specifically, the mean times of the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are halved.  Determine the reliability at 1000 hours for this component.&lt;br /&gt;
====Solution to Example 5====&lt;br /&gt;
The first step is to create the RBD.  Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and a load sharing container with the  &amp;lt;math&amp;gt;{{B}_{i}}&amp;lt;/math&amp;gt;  modes must be drawn in series, as illustrated next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.3.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The next step is to define the properties for each block, including those for the container.  Setting the failure distributions for modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is simple.  The more difficult part is setting the properties for the container and the contained blocks.  Based on the problem statement, the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are in a 2-out-of-3 load sharing redundancy.  When all three are working (i.e. when no  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode has occurred), each block has an exponential distribution with &amp;lt;math&amp;gt;\mu=10,000&amp;lt;/math&amp;gt;.  If one  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode occurs, then the two surviving units have an exponential distribution with  &amp;lt;math&amp;gt;\mu =5,000.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assume a Power Life-Stress relationship for the components.  Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\mu }_{1}}= &amp;amp; \frac{1}{KV_{1}^{n}}  \ (eqn 23)\\ &lt;br /&gt;
{{\mu }_{2}}= &amp;amp; \frac{1}{KV_{2}^{n}}  \ (eqn 24)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{1}}=10,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{V}_{1}}=1&amp;lt;/math&amp;gt;  in Eqn. 23 and casting it in terms of  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
10,000= &amp;amp; \frac{1}{K}  \ (eqn 25) \\ &lt;br /&gt;
K = &amp;amp; \frac{1}{10,000}=0.0001  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{2}}=5,000&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{V}_{2}}=1.5&amp;lt;/math&amp;gt;  (because if one fails, then each survivor takes on an additional 0.5 units of load) and Eqn. 25 for  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  in Eqn.24 yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
5,000= &amp;amp; \frac{1}{0.0001\cdot {{(1.5)}^{n}}} \\ &lt;br /&gt;
0.5= &amp;amp; {{(1.5)}^{-n}} \\ &lt;br /&gt;
\ln (0.5)= &amp;amp; -n\ln (1.5) \\ &lt;br /&gt;
n = &amp;amp; 1.7095  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This also could have been computed in ALTA, as shown in Figure 5.20, or with the Load &amp;amp; Life Parameter Experimenter in BlockSim, as shown in Figure 5.21.&lt;br /&gt;
 &lt;br /&gt;
At this point, the parameters for the load sharing units have been computed and can be set, as shown in Figure 5.22.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The next step is to set the weight proportionality factor.  This factor defines the portion of the load that the particular item carries while operating, as well as the load that shifts to the remaining units upon failure of the item.  To illustrate, assume three units (1, 2 and 3) are in a load sharing container with weight proportionality factors of 1, 2 and 3 respectively (and a 1-out-of-3 requirement).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Unit 1 carries  &amp;lt;math&amp;gt;\left( \tfrac{1}{1+2+3} \right)=0.166&amp;lt;/math&amp;gt;  or 16.6% of the total load.&lt;br /&gt;
:•	Unit 2 carries  &amp;lt;math&amp;gt;\left( \tfrac{2}{1+2+3} \right)=0.333&amp;lt;/math&amp;gt;  or 33.3% of the total load.&lt;br /&gt;
:•	Unit 3 carries  &amp;lt;math&amp;gt;\left( \tfrac{3}{1+2+3} \right)=0.50&amp;lt;/math&amp;gt;  or 50% of the total load.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The actual load on each unit then becomes the product of the entire load defined for the container times the portion carried by that unit.  For example, if the container load is 100 lbs, then the portion assigned to Unit 1 will be  &amp;lt;math&amp;gt;100\cdot 0.166=16.6&amp;lt;/math&amp;gt;  lbs.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the current example, all units share the same load and thus have equal weight proportionality factors.  Because these factors are relative, if the same number is used for all three items then the results will be the same.  Thus, weight proportional factor is set equal to 1 for each item.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5.20.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.20 Calculation performed in ALTA.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.21_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.21 Quick Parameter Estimator &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.21_3.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.22 Quick Parameter Estimator results &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.23_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.23 Defining Weight Proportional Factor. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The last properties that need to be defined are the total load and the 2-out-of-3 redundancy.  The total load is dependent on how the parameters were computed.  In this case, total load was assumed to be 3 when the parameters were computed (i.e. the load per item was 1 when all worked and 1.5 when two worked).  This is defined at the container level, set No. of Paths required = 3.&lt;br /&gt;
When all of the parameters have been specified in BlockSim, the reliability at 1,000 hours can be determined.  From the Analytical QCP, this is found to be 98.57%.&lt;br /&gt;
&lt;br /&gt;
=Standby Components=&lt;br /&gt;
&lt;br /&gt;
In the previous section, the case of a system with load sharing components was presented.  This is a form of redundancy with dependent components. That is, the failure of one component affects the failure of the other(s).  This section presents another form of redundancy: standby redundancy.  In standby redundancy the redundant components are set to be under a lighter load condition (or no load) while not needed and under the operating load when they are activated.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In standby redundancy the components are set to have two states: an active state and a standby state.  Components in standby redundancy have two failure distributions, one for each state.  When in the standby state, they have a quiescent (or dormant) failure distribution and when operating, they have an active failure distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case that both quiescent and active failure distributions are the same, the units are in a simple parallel configuration (also called a hot standby configuration).  When the rate of failure of the standby component is lower in quiescent mode than in active mode, that is called a warm standby configuration.  When the rate of failure of the standby component is zero in quiescent mode (i.e. the component cannot fail when in standby), that is called a cold standby configuration.  &lt;br /&gt;
&lt;br /&gt;
===Simple Standby Configuration===&lt;br /&gt;
&lt;br /&gt;
Consider two components in a standby configuration.  Component 1 is the active component with a Weibull failure distribution with parameters  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Component 2 is the standby component.  When Component 2 is operating, it also has a Weibull failure distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Furthermore, assume the following cases for the quiescent distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1:  The quiescent distribution is the same as the active distribution (hot standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2:  The quiescent distribution is a Weibull  distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 2000 (warm standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3: The component cannot fail in quiescent mode (cold standby).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this case, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be obtained using the following equation:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(t)={{R}_{1}}(t)+\underset{0}{\overset{t}{\mathop \int }}\,{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x)\cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}dx   \ (eqn 26)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode, such that: &amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
&amp;lt;br&amp;gt;  &lt;br /&gt;
&amp;lt;math&amp;gt;{{R}_{2;SB}}(x)={{R}_{2;A}}({{t}_{e}})  \ (eqn 27)&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Eqn. 27 can be solved for  &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  and substituted into Eqn.26.&lt;br /&gt;
Figure 5.24 illustrates the example as entered in BlockSim using a standby container.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5.24.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.24 Standby container.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The active and standby blocks are within a container, which is used to specify standby redundancy.  Since the standby component has two distributions (active and quiescent), the Block Properties window of the standby block has two pages for specifying each one.  Figures 5.24 and 5.26 illustrate these pages.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system reliability results for 1000 hours are given in the following table:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-24.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Note that even though the  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  value for the quiescent distribution is the same as in the active distribution, it is possible that the two can be different. That is, the failure modes present during the quiescent mode could be different from the modes present during the active mode.  In that sense, the two distribution types can be different as well (e.g. lognormal when quiescent and Weibull when active).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In many cases when considering standby systems, a switching device may also be present that switches from the failed active component to the standby component.  The reliability of the switch can also be incorporated into Eqn.26, as presented in the next section.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim&#039;s System Reliability Equation window returns a single token for the reliability of units in a standby configuration.  This is the same as the load sharing case presented in the previous section. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image: Fig 5.25.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.25 Defining the active failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.26.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.26 Defining the quiescent failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Reliability of Standby Systems with a Switching Device===&lt;br /&gt;
&lt;br /&gt;
In many cases when dealing with standby systems, a switching device is present that will switch to the standby component when the active component fails.  Therefore, the failure properties of the switch must also be included in the analysis.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.26.2.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In most cases when the reliability of a switch is to be included in the analysis, two probabilities can be considered.  The first and most common one is the probability of the switch performing the action (i.e. switching) when requested to do so.  This is called Switch Probability per Request in BlockSim and is expressed as a static probability (e.g. 90%).  The second probability is the quiescent reliability of the switch.  This is the reliability of the switch as it ages (e.g. the switch might wear out with age due to corrosion, material degradation, etc.). Thus it is possible for the switch to fail before the active component fails.  However, a switch failure does not cause the system to fail, but rather causes the system to fail only if the switch is needed and the switch has failed.  For example, if the active component does not fail until the mission end time and the switch fails, then the system does not fail.  However, if the active component fails and the switch has also failed, then the system cannot be switched to the standby component and it therefore fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In analyzing standby components with a switching device, either or both failure probabilities (during the switching or while waiting to switch) can be considered for the switch, since each probability can represent different failure modes.  For example, the switch probability per request may represent software-related issues or the probability of detecting the failure of an active component, and the quiescent probability may represent wear-out type failures of the switch.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the formulation, consider the previous example that assumes perfect switching.  To examine the effects of including an imperfect switch, assume that when the active component fails there is a 90% probability that the switch will switch from the active component to the standby component.  In addition, assume that the switch can also fail due to a wear-out failure mode described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.7 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 5000.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , is given by the following equation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t)= &amp;amp; {{R}_{1}}(t) \\ &lt;br /&gt;
&amp;amp; +\underset{0}{\overset{t}{\mathop \int }}\,\{{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x) \\ &lt;br /&gt;
&amp;amp; \cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}\cdot {{R}_{SW;Q}}(x)\cdot {{R}_{SW;REQ}}(x)\}dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;Q}}&amp;lt;/math&amp;gt;  is the quiescent reliability of the switch.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;REQ}}&amp;lt;/math&amp;gt;  is the switch probability per request.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
This problem can be solved in BlockSim by including these probabilities in the container&#039;s properties, as shown in Figures 5.27 and 5.28.  In BlockSim, the standby container is acting as the switch.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.28_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.27 Standby container (switch) failure probabilities while attempting to switch &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.27.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.28 Standby container (switch) failure distribution while waiting to switch &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that there are additional properties that can be specified in BlockSim for a switch, such as Switch Restart Probability, Finite Restarts and Switch Delay Time.  In many applications, the switch is re-tested (or re-cycled) if it fails to switch the first time.  In these cases, it might be possible that it switches in the second or third, or  &amp;lt;math&amp;gt;{{n}^{th}}&amp;lt;/math&amp;gt; attempt.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The Switch Restart Probability specifies each additional attempt&#039;s probability of successfully switching and the Finite Restarts specifies the total number of attempts.  Note that the Switch Restart Probability specifies the probability of success of each trial (or attempt).  The probability of success of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  consecutive trials is calculated by BlockSim using the binomial distribution and this probability is then incorporated into Eqn. (stb2a).  The Switch Delay Time property is related to repairable systems and is considered in BlockSim only when using simulation.  When using the analytical solution (i.e. for a non-repairable system), this property is ignored.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving the analytical solution (as given by Eqn. stb2a), the following results are obtained.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-30.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From the table above, it can be seen that the presence of a switching device has a significant effect on the reliability of a standby system.  It is therefore important when modeling standby redundancy to incorporate the switching device reliability properties.  It should be noted that this methodology is not the same as treating the switching device as another series component with the standby subsystem.  This would be valid only if the failure of the switch resulted in the failure of system (e.g. switch failing open).  In Eqn. (stb2a), the Switch Probability per Request and quiescent probability are present only in the second term of the equation.  Treating these two failure modes as a series configuration with the standby subsystem would imply that they are also present when the active component is functioning (i.e. first term of Eqn. stb2a).  This is invalid and would result in the underestimation of the reliability of the system.  In other words, these two failure modes become significant only when the active component fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As an example, and if we consider the warm standby case, the reliability of the system without the switch is 70.57% at 1000 hours.  If the system was modeled so that the switching device was in series with the warm standby subsystem, the result would have been:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(1000)= &amp;amp; {{R}_{Standby}}(1000)\cdot {{R}_{sw,Q(1000)}}\cdot {{R}_{sw,req}} \\ &lt;br /&gt;
= &amp;amp; 0.7057\cdot 0.9372\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.5952  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case where a switch failure mode causes the standby subsystem to fail, then this mode can be modeled as an individual block in series with the standby subsystem.&lt;br /&gt;
&lt;br /&gt;
===Example 6===&lt;br /&gt;
Consider a car with four new tires and a full-size spare.  Assume the following failure characteristics:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires follow a Weibull distribution with a  ..  and an  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  40,000 miles while on the car due to wear.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires also have a probability of failing due to puncture or other causes.  For this, assume a constant rate for this occurrence with a probability of 1 every 50,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	When not on the car (i.e. is a spare), a tire&#039;s probability of failing also has a Weibull distribution with a  &amp;lt;math&amp;gt;\beta =&amp;lt;/math&amp;gt;  2 and  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  120,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Assume a mission of 1,000 miles.  If a tire fails during this trip, it will be replaced with the spare.  However, the spare will not be repaired during the trip.  In other words, the trip will continue with the spare on the car and if the spare fails the system will fail.  Determine the probability of system failure.&lt;br /&gt;
====Solution to Example 6====&lt;br /&gt;
Active failure distribution for tires:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to wear-out, Weibull  &amp;lt;math&amp;gt;\beta =4&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =40,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to random puncture, exponential  &amp;lt;math&amp;gt;\mu =50,000.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	The quiescent failure distribution is a Weibull distribution with &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =120,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The block diagram for each tire has two blocks in series, one block representing the wear-out mode and the other the random puncture mode, as shown next:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:small5.gif|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
There are five tires, four active and one standby (represented in the diagram by a standby container with a 4-out-of-5 requirement), as shown next: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BStirecontainer.png|thumb|center|400px|]]&lt;br /&gt;
 &lt;br /&gt;
For the standby Wear block, set the active failure and the quiescent distributions, but for the Puncture block, set only the active puncture distribution (because the tire cannot fail due to puncture while stored).  Using BlockSim, the probability of system failure is found to be 0.003 or 0.3%. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Note Regarding Numerical Integration Solutions=&lt;br /&gt;
&lt;br /&gt;
Load sharing and standby solutions in BlockSim are performed using numerical integration routines.  As with any numerical analysis routine, the solution error depends on the number of iterations performed, the step size chosen and related factors, plus the behavior of the underlying function.  By default, BlockSim uses a certain set of preset factors.  In general, these defaults are sufficient for most problems.  If a higher precision or verification of the precision for a specific problem is required, BlockSim&#039;s preset options can be modified and/or the integration error can be assessed using the Integration Parameters... option for each container.  For more details, you can refer to the documentation on the Algorithm Setup window in the BlockSim 7 User&#039;s Guide.&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15563</id>
		<title>Time-Dependent System Reliability (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15563"/>
		<updated>2012-02-13T22:12:11Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Standby Components */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|5}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the previous chapter, different system configuration types were examined, as well as different methods for obtaining the system&#039;s reliability function analytically.  Because the reliabilities in the problems presented were treated as probabilities (e.g.  &amp;lt;math&amp;gt;P(A)&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; ), the reliability values and equations presented were referred to as static (not time-dependent).  Thus, in the prior chapter, the life distributions of the components were not incorporated in the process of calculating the system reliability.  In this chapter, time dependency in the reliability function will be introduced.  We will develop the models necessary to observe the reliability over the life of the system, instead of at just one point in time.  In addition, performance measures such as failure rate, MTTF and warranty time will be estimated for the entire system.  The methods of obtaining the reliability function analytically remain identical to the ones presented in the previous chapter, with the exception that the reliabilities will be functions of time.  In other words, instead of dealing with  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , we will use  &amp;lt;math&amp;gt;{{R}_{i}}(t)&amp;lt;/math&amp;gt; .  All examples in this chapter assume that no repairs are performed on the components.  &lt;br /&gt;
&lt;br /&gt;
=Analytical Life Predictions=&lt;br /&gt;
The analytical approach presented in the prior chapter involved the determination of a mathematical expression that describes the reliability of the system, expressed in terms of the reliabilities of its components.  So far we have estimated only static system reliability (at a fixed time).  For example, in the case of a system with three components in series, the system&#039;s reliability equation was given by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}  \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The values of  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  were given for a common time and the reliability of the system was estimated for that time.  However, since the component failure characteristics can be described by distributions, the system reliability is actually time-dependent.  In this case, Eqn. (1) can be rewritten as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability of the system for any mission time can now be estimated.  Assuming a Weibull life distribution for each component, Eqn. (1) can now be expressed in terms of each component&#039;s reliability function, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the same manner, any life distribution can be substituted into the system reliability equation.  Suppose that the times-to-failure of the first component are described with a Weibull distribution, the times-to-failure of the second component with an exponential distribution and the times-to-failure of the third component with a normal distribution.  Then Eqn. (1) can be written as:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot \left[ 1-\Phi \left( \frac{t-{{\mu }_{3}}}{{{\sigma }_{3}}} \right) \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It can be seen that the biggest challenge is in obtaining the system&#039;s reliability function in terms of component reliabilities, which has already been discussed in depth.  Once this has been achieved, calculating the reliability of the system for any mission duration is just a matter of substituting the corresponding component reliability functions into the system reliability equation.&lt;br /&gt;
===Advantages of the Analytical Method===&lt;br /&gt;
The primary advantage of the analytical solution is that it produces a mathematical expression that describes the reliability of the system.  Once the system&#039;s reliability function has been determined, other calculations can then be performed to obtain metrics of interest for the system. Such calculations include:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of warranty periods.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s failure rate.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s MTTF.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, optimization and reliability allocation techniques can be used to aid engineers in their design improvement efforts.  Another advantage in using analytical techniques is the ability to perform static calculations and analyze systems with a mixture of static and time-dependent components.  Finally, the reliability importance of components over time can be calculated with this methodology.&lt;br /&gt;
===Disadvantages of the Analytical Method===&lt;br /&gt;
The biggest disadvantage of the analytical method is that formulations can become very complicated.  The more complicated a system is, the larger and more difficult it will be to analytically formulate an expression for the system&#039;s reliability.  For particularly detailed systems this process can be quite time-consuming, even with the use of computers.  Furthermore, when the maintainability of the system or some of its components must be taken into consideration, analytical solutions become intractable.  In these situations, the use of simulation methods may be more advantageous than attempting to develop a solution analytically.  Simulation methods are presented in later chapters.&lt;br /&gt;
===Looking at a Simple &#039;&#039;Complex&#039;&#039; System Analytically===&lt;br /&gt;
&lt;br /&gt;
The complexity involved in an analytical solution can be best illustrated by looking at the simple &#039;&#039;complex&#039;&#039; system with 15 components, as shown in Figure 5.1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1.png|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.1 An RBD of a complex system.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system reliability for this system (computed using BlockSim) is shown next.  The first solution is provided using BlockSim&#039;s symbolic solution.  In symbolic mode, BlockSim breaks the equation into segments, identified by tokens, that need to be substituted into the final system equation for a complete solution.  This creates algebraic solutions that are more compact than if the substitutions were made.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; D2\cdot D3\cdot {{R}_{L}} \\ &lt;br /&gt;
D3= &amp;amp; +{{R}_{K}}\cdot IK \\ &lt;br /&gt;
IK= &amp;amp; +{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}} \\ &lt;br /&gt;
D2 = &amp;amp; +{{R}_{A}}\cdot {{R}_{E}}\cdot IE \\ &lt;br /&gt;
IE = &amp;amp; -D1\cdot {{R}_{M}}\cdot {{R}_{N}}+{{R}_{M}}\cdot {{R}_{N}}+D1 \\ &lt;br /&gt;
D1 = &amp;amp; +{{R}_{D}}\cdot ID \\ &lt;br /&gt;
ID = &amp;amp; -{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting the terms yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; {{R}_{A}}\cdot {{R}_{E}}\cdot {{R}_{L}}\cdot {{R}_{K}} \\ &lt;br /&gt;
&amp;amp; \cdot \{({{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})\cdot {{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}}-{{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}\} \\ &lt;br /&gt;
&amp;amp; \cdot \{{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}}\}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
BlockSim&#039;s automatic algebraic simplification would yield the following format for the above solution: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; (({{R}_{A}}\cdot {{R}_{E}}(-({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})){{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})))) \\ &lt;br /&gt;
&amp;amp; ({{R}_{K}}({{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; +RI\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}})){{R}_{L}})  \ (eqn 2)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this equation, each  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  represents the reliability function of a block.  For example, if  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  has a Weibull distribution, then each  &amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;  and so forth.  Substitution of each component&#039;s reliability function in Eqn.2 will result in an analytical expression for the system reliability as a function of time, or  &amp;lt;math&amp;gt;{{R}_{s}}(t)&amp;lt;/math&amp;gt; , which is the same as  &amp;lt;math&amp;gt;(1-cd{{f}_{System}}).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Obtaining Other Functions of Interest===&lt;br /&gt;
Once the system reliability equation (or the cumulative density function,  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt; ) has been determined, other functions and metrics of interest can be derived.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Consider the following simple system:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-2.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.2 Simple two-component system. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Furthermore, assume that component 1 follows an exponential distribution with a mean of 10,000 (&amp;lt;math&amp;gt;\mu =10,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;\lambda =1/10,000)&amp;lt;/math&amp;gt;  and component 2 follows a Weibull distribution with  &amp;lt;math&amp;gt;\beta =6&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt; .  The reliability equation of this system is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}  \ (eqn 3) &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{F}_{S}}(t)= &amp;amp; 1-({{R}_{1}}(t)\cdot {{R}_{2}}(t)) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \right) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
{{analytical system pdf}}&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability====&lt;br /&gt;
Conditional reliability is the probability of a system successfully completing another mission following the successful completion of a previous mission.  The time of the previous mission and the time for the mission to be undertaken must be taken into account for conditional reliability calculations.  The system&#039;s conditional reliability function is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(T,t)=\frac{R(T+t)}{R(T)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Eqn.6 gives the reliability for a new mission of duration  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  having already accumulated  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  hours of operation up to the start of this new mission. The system is evaluated to assure that it will start the next mission successfully.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2, the reliability for mission of  &amp;lt;math&amp;gt;t=1,000&amp;lt;/math&amp;gt;  hours, having an age of  &amp;lt;math&amp;gt;T=500&amp;lt;/math&amp;gt;  hours, is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(T=500,t=1000)= &amp;amp; \frac{R(T+t)}{R(T)} \\ &lt;br /&gt;
= &amp;amp; \frac{R(1500)}{R(500)} \\ &lt;br /&gt;
= &amp;amp; \frac{{{e}^{-\tfrac{1500}{10,000}}}\cdot {{e}^{-{{\left( \tfrac{1500}{10,000} \right)}^{6}}}}}{{{e}^{-\tfrac{500}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{500}{10,000} \right)}^{6}}}}} \\ &lt;br /&gt;
= &amp;amp; 0.9048=90.48%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.3.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.4.png|thumb|center|400px|]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability for Components====&lt;br /&gt;
&lt;br /&gt;
Now in this formulation, it was assumed that the accumulated age was equivalent for both units. That is, both started life at zero and aged to 500.  It is possible to consider an individual component that has already accumulated some age (used component) in the same formulation.  To illustrate this, assume that component 2 started life with an age of T=100.  Then the reliability equation of the system, as given in Eqn.3, would need to be modified to include a conditional term for 2, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{R}_{1}}(t)\cdot \frac{{{R}_{2}}({{T}_{2}}+t)}{{{R}_{2}}({{T}_{2}})} \ (eqn 7)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In BlockSim, the start age input box may be used to specify a starting age greater than zero.&lt;br /&gt;
{{system failure rate analytical}}&lt;br /&gt;
&lt;br /&gt;
====System Mean Life (Mean Time To Failure)====&lt;br /&gt;
The mean life (or mean time to failure, MTTF) can be obtained by integrating the system reliability function from zero to infinity: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;MTTF=\int_{0}^{\infty }{{R}_{s}}\left( t \right)dt   \ (eqn 10)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mean time is a performance index and does not provide any information about the behavior of the failure distribution of the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)dt \\ &lt;br /&gt;
= &amp;amp; 5978.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Warranty Period and BX Life====&lt;br /&gt;
Sometimes it is desirable to know the time value associated with a certain reliability.  Warranty periods are often calculated by determining what percentage of the failure population can be covered financially and estimating the time at which this portion of the population will fail.  Similarly, engineering specifications may call for a certain BX life, which also represents a time period during which a certain proportion of the population will fail.  For example, the B10 life is the time in which 10% of the population will fail.  &lt;br /&gt;
This is obtained by setting  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  to the desired value and solving for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}\left( t \right)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To compute the time by which reliability would be equal to 90%, Eqn.11 is recast as follows and solved for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this case,  &amp;lt;math&amp;gt;t=1053.59&amp;lt;/math&amp;gt; .  Equivalently, the B10 life for this system is also  &amp;lt;math&amp;gt;1053.59&amp;lt;/math&amp;gt; .&lt;br /&gt;
Except for some trivial cases, a closed form solution for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  cannot be obtained.   Thus, it is necessary to solve for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  using numerical methods.  BlockSim uses numerical methods.&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
Consider a system consisting of three exponential units in series with the following failure rates (in failures per hour):  &amp;lt;math&amp;gt;{{\lambda }_{1}}&amp;lt;/math&amp;gt;  = 0.0002,  &amp;lt;math&amp;gt;{{\lambda }_{2}}&amp;lt;/math&amp;gt;  = 0.0005 and  &amp;lt;math&amp;gt;{{\lambda }_{3}}&amp;lt;/math&amp;gt;  = 0.0001.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the reliability equation for the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the reliability of the system after 150 hours of operation?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s failure rate equation.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the MTTF for the system?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What should the warranty period be for a 90% reliability?&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The analytical expression for the reliability of the system is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\lambda }_{1}}t}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot {{e}^{-{{\lambda }_{1}}t}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
At 150 hours of operation, the reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{e}^{-(0.0002+0.0005+0.0001)150}} \\ &lt;br /&gt;
= &amp;amp; 0.8869\text{ or }88.69%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.12 is taken with respect to time, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; -\frac{d[{{R}_{s}}(t)]}{dt} \\ &lt;br /&gt;
= &amp;amp; -\frac{d\left[ {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}} \right]}{dt} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
:•	The system&#039;s failure rate can now be obtained simply by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn. 13 by the system&#039;s reliability function given in Eqn.12, and:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}\left( t \right)= &amp;amp; \frac{{{f}_{s}}\left( t \right)}{{{R}_{s}}\left( t \right)} \\ &lt;br /&gt;
= &amp;amp; \frac{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}}{{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}) \\ &lt;br /&gt;
= &amp;amp; 0.0008\text{ }fr/hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Combining Eqn.10 and Eqn. 12, the system&#039;s MTTF can be obtained:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }{{R}_{s}}\left( t \right)dt \\ &lt;br /&gt;
= &amp;amp; \int_{0}^{\infty }{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}dt \\ &lt;br /&gt;
= &amp;amp; \frac{1}{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})} \\ &lt;br /&gt;
= &amp;amp; 1250\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Solving Eqn. 12 with respect to time will yield the corresponding warranty period for a 90% reliability.  In this case, the system reliability equation is simple and a closed form solution exists.  The warranty time can now be found by solving:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
t= &amp;amp; -\frac{\ln (R)}{{{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}} \\ &lt;br /&gt;
= &amp;amp; -\frac{\ln (0.9)}{0.0008} \\ &lt;br /&gt;
= &amp;amp; 131.7\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the warranty period should be 132 hours.&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
Consider the system shown in Figure 5.5.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.5.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.5 Complex bridge system in Example 2. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Components  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  through  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  are Weibull distributed with  &amp;lt;math&amp;gt;\beta =1.2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1230&amp;lt;/math&amp;gt;  hours.  The starting and ending blocks cannot fail.  &amp;lt;br&amp;gt;&lt;br /&gt;
Determine the following:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability equation for the system and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s failure rate equation and the corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The warranty time for a 90% reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability for a 200-hour mission, if it is known that the system has already successfully operated for 200 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The first step is to obtain the reliability function for the system.  The methods described in the previous chapter can be employed, such as the event space or path-tracing methods.  Using BlockSim, the following reliability equation is obtained:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; ({{R}_{Start}}\cdot {{R}_{End}}(2{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}))  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that since the starting and ending blocks cannot fail,  &amp;lt;math&amp;gt;{{R}_{Start}}=1&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{End}}=1,&amp;lt;/math&amp;gt;  Eqn.14 can be reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; 2\cdot {{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  is the reliability equation for Component A, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{B}}&amp;lt;/math&amp;gt;  is the reliability equation for Component  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; , etc.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the components in this example are identical, the system reliability equation can be further reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2R{{(t)}^{2}}+2R{{(t)}^{3}}-5R{{(t)}^{4}}+2R{{(t)}^{5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Or, in terms of the failure distribution:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.6.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.6.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.6 Reliability plot for the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.18 is taken with respect to time, resulting in: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; 4\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
&amp;amp; -20\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  can now be plotted for different time values,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , as shown in Figure 5.7.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system&#039;s failure rate can now be obtained by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn.19 by the system&#039;s reliability function given in Eqn.18, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}(t)= &amp;amp; \frac{4\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}} \\ &lt;br /&gt;
&amp;amp; +\frac{-20\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.8.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.7.gif|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.7 &#039;&#039;pdf&#039;&#039; plot for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of the system is obtained by integrating the system&#039;s reliability function given by Eqn. 18 from time zero to infinity, as given by Eqn. 10.  Using BlockSim&#039;s Analytical QCP, an  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of 1007.8 hours is calculated, as shown in Figure 5.9.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The warranty time can be obtained by solving Eqn. 18 with respect to time for a system reliability  &amp;lt;math&amp;gt;{{R}_{s}}=0.9&amp;lt;/math&amp;gt; .  Using the Analytical QCP and selecting the &amp;lt;br&amp;gt;&lt;br /&gt;
Warranty Time option, a time of 372.72 hours is obtained, as shown in Figure 5.10.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.8.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.8 Failure rate for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.9.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.9 MTTF of the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.10.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.10 Time at which &#039;&#039;R&#039;&#039;=0.9 or 90% for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.11.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.11 Conditional reliability calculation for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, the conditional reliability can be obtained using Eqn.6 and Eqn.18, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(200,200)= &amp;amp; \frac{R(400)}{R(200)} \\ &lt;br /&gt;
= &amp;amp; \frac{0.883825}{0.975321} \\ &lt;br /&gt;
= &amp;amp; 0.906189  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated using BlockSim&#039;s Analytical QCP, as shown in Figure 5.11.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Approximating the System CDF=&lt;br /&gt;
&lt;br /&gt;
In many cases, it is valuable to fit a distribution that represents the system&#039;s times-to-failure.  This can be useful when the system is part of a larger assembly and may be used for repeated calculations or in calculations for other systems.  In cases such as this, it can be useful to characterize the system&#039;s behavior by fitting a distribution to the overall system and calculating parameters for this distribution.   This is equivalent to fitting a single distribution to describe  &amp;lt;math&amp;gt;{{R}_{S}}(t&amp;lt;/math&amp;gt; ).  In essence, it is like reducing the entire system to a component in order to simplify calculations.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
To compute an approximate reliability function for this system,  &amp;lt;math&amp;gt;{{R}_{A}}(t)\simeq {{R}_{S}}(t)&amp;lt;/math&amp;gt; , one would compute  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  pairs of reliability and time values and then fit a single distribution to the data, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 10,396.7)=10% \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 9,361.9)=20% \\ &lt;br /&gt;
&amp;amp; ... \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 1,053.6)=90%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A single distribution,  &amp;lt;math&amp;gt;{{R}_{A}}(t)&amp;lt;/math&amp;gt; , that approximates  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  can now be computed from these pairs using life data analysis methods.  If using the Weibull++ software, one would enter the values as free form data.&lt;br /&gt;
&lt;br /&gt;
===Example 3===&lt;br /&gt;
Compute a single Weibull distribution approximation for the system in Example 2.&lt;br /&gt;
====Solution to Example 3====&lt;br /&gt;
The system in the previous example, shown in Figure 5.5, can be approximated by use of a 2-parameter Weibull distribution with  &amp;lt;math&amp;gt;\beta =2.02109&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1123.51&amp;lt;/math&amp;gt; .  In BlockSim, this is accomplished by representing the entire system as one distribution by going to the Distribution Fit window Figure 5.12. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.13.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.12 Representing a system with a distribution.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.12.PNG|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.13 Distribution Fitting window.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
by clicking the Distribution Fit Window, the Distribution Estimator window will appear (Figure 5.13).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this window you can select a distribution to represent the data. BlockSim will then generate a number of system failure times based on the system&#039;s reliability function. The system&#039;s reliability function can be used to solve for a time value associated with that unreliability value. The distribution of the generated time values can then be fitted to a probability distribution function.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a value of  &amp;lt;math&amp;gt;F(t)=0.11&amp;lt;/math&amp;gt; .  Using the system&#039;s reliability equation and solving for time, the corresponding time-to-failure for a 0.11 unreliability can be calculated.  &amp;lt;br&amp;gt;&lt;br /&gt;
For the system of Example 2, the time for a 0.11 unreliability is 389.786 hours.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
When enough points have been generated, the selected distribution will be fitted to this data set and the distribution&#039;s parameters will be returned.  In addition, if ReliaSoft&#039;s Weibull++ is installed, the generated data can be viewed/analyzed using a Weibull++ instance, as shown in Figure 5.14.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.14.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.14 Using Weibull++ to calculate distribution parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
It is recommended that the analyst examine the fit to ascertain the applicability of the approximation.&lt;br /&gt;
&lt;br /&gt;
=Duty Cycle=&lt;br /&gt;
&lt;br /&gt;
Components of a system may not operate continuously during a system&#039;s mission, or may be subjected to loads greater or lesser than the rated loads during system operation.  To model this, a factor called the Duty Cycle ( &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt; ) is used.  The duty cycle may also be used to account for changes in environmental stress, such as temperature changes, that may effect the operation of a component.  The duty cycle is a positive value, with a default value of 1 representing continuous operation at rated load, and any values other than 1 representing other load values with respect to the rated load value (or total operating time).   A duty cycle value higher than 1 indicates a load in excess of the rated value.  A duty cycle value lower than 1 indicates that the component is operating at a load lower than the rated load or not operating continuously during the system&#039;s mission.  For instance, a duty cycle of 0.5 may be used for a component that operates only half of the time during the system&#039;s mission.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The reliability metrics for a component with a duty cycle are calculated as follows. Let  &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt;  represent the duty cycle during a particular mission of the component,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  represent the mission time and  &amp;lt;math&amp;gt;{t}&#039;&amp;lt;/math&amp;gt;  represent the accumulated age. Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{t}&#039;={{d}_{c}}\times t&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability equation for the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R({t}&#039;)=R({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The component &#039;&#039;pdf&#039;&#039; is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;f({t}&#039;)=-\frac{d(R({t}&#039;))}{dt}=-\frac{d(R({{d}_{c}}\times t))}{dt}={{d}_{c}}f({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The failure rate of the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\lambda ({t}&#039;)=\frac{f({t}&#039;)}{R({t}&#039;)}=\frac{{{d}_{c}}f({{d}_{c}}\times t)}{R({{d}_{c}}\times t)}={{d}_{c}}\lambda ({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Example 4===&lt;br /&gt;
Consider a computer system with three components: a processor, a hard drive and a CD drive in series as shown next.  Assume that all three components follow a Weibull failure distribution with the parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}=1.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=5000&amp;lt;/math&amp;gt;  for the processor,  &amp;lt;math&amp;gt;{{\beta }_{2}}=2.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{2}}=3000&amp;lt;/math&amp;gt;  for the hard drive, and  &amp;lt;math&amp;gt;{{\beta }_{3}}=2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=4000&amp;lt;/math&amp;gt;  for the CD drive.  Determine the reliability of the computer system after one year (365 days) of operation, assuming that the CD drive is used only 30% of the time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5ex4.png|thumb|center|300px|]]&lt;br /&gt;
&lt;br /&gt;
====Solution to Example 4====&lt;br /&gt;
The reliability of the processor after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  {{R}_{processor}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{5000} \right)}^{1.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\text{ or }98.05%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the hard drive after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{harddrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{3000} \right)}^{2.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9948\text{ or }99.48%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the CD drive after 365 days of operation (taking into account the 30% operation using a duty cycle of 0.3) is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{CDdrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{{{d}_{c}}\times 365}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{0.3\times 365}{4000} \right)}^{2}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9993\text{ or }99.93%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.15.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.15 Result for the computer system reliability.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
Thus the reliability of the computer system after 365 days of operation is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{s}}(365)= &amp;amp; {{R}_{processor}}(365)\cdot {{R}_{harddrive}}(365)\cdot {{R}_{CDdrive}}(365) \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\cdot 0.9948\cdot 0.9993 \\ &lt;br /&gt;
	  = &amp;amp; 0.9747\text{ or }97.47%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;	&lt;br /&gt;
&amp;lt;br&amp;gt;	&lt;br /&gt;
This result can be obtained in BlockSim as shown in Figure DutyCycleExResults.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Load Sharing=&lt;br /&gt;
As presented in earlier chapters, a reliability block diagram (RBD) allows you to graphically represent how the components within a system are reliability-wise connected.  In most cases, independence is assumed across the components within the system.  For example, the failure of component A does not affect the failure of component B.  However, if a system consists of components that are sharing a load, then the assumption of independence no longer holds true.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If one component fails, then the component(s) that are still operating will have to assume the failed unit&#039;s portion of the load.  Therefore, the reliabilities of the surviving unit(s) will change.  Calculating the system reliability is no longer an easy proposition.  In the case of load sharing components, the change of the failure distributions of the surviving components must be known in order to determine the system&#039;s reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate this, consider the a system of two units connected reliability-wise in parallel (Figure 5.16).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-16.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.16 Two units connected reliability-wise in parallel.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Assume that the units must supply an output of 8 volts and that if both units are operational, each unit is to supply 50% of the total output.  If one of the units fails, then the surviving unit supplies 100%.  Furthermore, assume that having to supply the entire load has a negative impact on the reliability characteristics of the surviving unit.  Since the reliability characteristics of the unit change based on whether both or only one is operating, a life distribution along with a life-stress relationship (as discussed in Chapter 3) will be needed to model each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the steps needed, we will create the model starting from raw data.  Assume that a total of 20 units were tested to failure at 7, 10 and 15 volts.  The test data set is presented in the next table.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-17.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
For this example, Units 1 and 2 are the same component.  Therefore, only one set of data was collected.  However, it is possible that the load sharing components in a system may not be the same.  If that were the case, data would need to be collected for each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The data set in Table 1 was analyzed using ReliaSoft&#039;s ALTA software (as shown in Figure 5.17) with the Inverse Power Law as the underlying life-stress relationship and Weibull as the life distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated model parameters,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , are shown next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
\beta = &amp;amp; 1.9239 \\ &lt;br /&gt;
K= &amp;amp; 3.2387\times {{10}^{-7}} \\ &lt;br /&gt;
n= &amp;amp; 3.4226  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:Or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( 3.2387\times {{10}^{-7}}S_{1}^{3.4226}t \right)}^{1.9239}}}}  \ (eqn 20)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1}}(t,{{S}_{1}})=\beta KS_{1}^{n}{{\left( KS_{1}^{n}t \right)}^{\beta -1}}{{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}}  \ (eqn 21 )&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And for this case:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{R}_{2}}(t,{{S}_{2}})  \\ &lt;br /&gt;
{{f}_{1}}(t,{{S}_{1}})= &amp;amp; {{f}_{2}}(t,{{S}_{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.17.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.17 Using ALTA to calculate component parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 5.18 shows a plot of Eqn.20.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Now that the failure properties have been determined using the test data, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t,S)= &amp;amp; {{R}_{1}}(t,{{S}_{1}})\cdot {{R}_{2}}(t,{{S}_{2}}) \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{1}}\left( x,{{S}_{1}} \right)\cdot {{R}_{2}}(x,{{S}_{2}})\cdot \left( \frac{{{R}_{2}}({{t}_{1e}}+(t-x),S)}{{{R}_{2}}({{t}_{1e}},S)} \right)dx \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{2}}\left( x,{{S}_{2}} \right)\cdot {{R}_{1}}(x,{{S}_{1}})\cdot \left( \frac{{{R}_{1}}({{t}_{2e}}+(t-x),S)}{{{R}_{1}}({{t}_{2e}},S)} \right)dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:Where: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{S}_{1}}= &amp;amp; {{P}_{1}}S \\ &lt;br /&gt;
{{S}_{2}}= &amp;amp; {{P}_{2}}S  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-18.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig. 5.18 Reliability curves for different voltage output conditions. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.png|thumb|center|395px]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  is the total load (or required output).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{P}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{P}_{2}}&amp;lt;/math&amp;gt;  are the portion of the total load that each unit supports when both units are operational.  In this case,  &amp;lt;math&amp;gt;{{P}_{1}}={{P}_{2}}=0.5=50%.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{S}_{2}}&amp;lt;/math&amp;gt;  represent the portions of the load that Unit 1 and Unit 2 must support when both units are operational.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{{{1}_{e}}}}&amp;lt;/math&amp;gt;  is the equivalent operating time for Unit 1 if it had been operating at  &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  instead of  &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt; .  A graphical representation of the equivalent time is shown in Figure 5.19, where the curve marked by L represents the low stress (load) and the curve marked by H represents the high stress (load).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  can be calculated by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t)= &amp;amp; {{R}_{1}}({{t}_{1e}}) \\ &lt;br /&gt;
{{e}^{-{{(tKS_{1}^{n})}^{\beta }}}}= &amp;amp; {{e}^{-{{({{t}_{1e}}K{{S}^{n}})}^{\beta }}}} \\ &lt;br /&gt;
tS_{1}^{n}= &amp;amp; {{t}_{1e}}{{S}^{n}} \\ &lt;br /&gt;
{{t}_{1e}}= &amp;amp; t{{\left( \frac{{{S}_{1}}}{S} \right)}^{n}},\text{     }{{S}_{1}}={{P}_{1}}S \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{1e}}=tP_{1}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  can be calculated the same way, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{2}}(t)= &amp;amp; {{R}_{2}}({{t}_{2e}}) \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{2e}}=tP_{2}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this example, the reliability equations for Unit 1 and Unit 2 are the same since they are the same type of component and demonstrate the same failure properties.  In addition, the total output is divided equally between the two units (when both units are operating), so  &amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  will also be the same.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The next step is to determine the reliability of the system after 8,760 hours,  &amp;lt;math&amp;gt;R(t=8,760)&amp;lt;/math&amp;gt; .  Using Eqn. 22 the system reliability is found to be:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t=8760)= &amp;amp; 0.8567 \\ &lt;br /&gt;
= &amp;amp; 85.67%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Load Sharing in BlockSim===&lt;br /&gt;
BlockSim uses this formulation when computing reliabilities of units in a load sharing configuration.  When using the System Reliability Equation window, BlockSim returns a single token for the reliability of units in a load sharing configuration (as well as in the case of standby redundancy, discussed in the next section).  As an example, consider the following RBD with Unit 1 in series with a container that includes two load sharing units.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.2.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
BlockSim will return the system equation as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{System}}=+{{R}_{LS}}\cdot {{R}_{1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{LS}}&amp;lt;/math&amp;gt;  implies a form similar to Eqn. 22.  BlockSim allows for  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -out-of- &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  units in a load sharing configuration.&lt;br /&gt;
&lt;br /&gt;
===Example 5===&lt;br /&gt;
A component has five possible failure modes,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , and the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are interdependent.  The system will fail if mode  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  occurs, mode  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  occurs or two out of the three  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes occur.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  each have a Weibull distribution, with a  &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt;  and 15,000 respectively.  Events  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  each have an exponential distribution with a mean of 10,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If any  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  event occurs (i.e.  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt; ), the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are more likely to occur.  Specifically, the mean times of the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are halved.  Determine the reliability at 1000 hours for this component.&lt;br /&gt;
====Solution to Example 5====&lt;br /&gt;
The first step is to create the RBD.  Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and a load sharing container with the  &amp;lt;math&amp;gt;{{B}_{i}}&amp;lt;/math&amp;gt;  modes must be drawn in series, as illustrated next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.3.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The next step is to define the properties for each block, including those for the container.  Setting the failure distributions for modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is simple.  The more difficult part is setting the properties for the container and the contained blocks.  Based on the problem statement, the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are in a 2-out-of-3 load sharing redundancy.  When all three are working (i.e. when no  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode has occurred), each block has an exponential distribution with &amp;lt;math&amp;gt;\mu=10,000&amp;lt;/math&amp;gt;.  If one  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode occurs, then the two surviving units have an exponential distribution with  &amp;lt;math&amp;gt;\mu =5,000.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assume a Power Life-Stress relationship for the components.  Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\mu }_{1}}= &amp;amp; \frac{1}{KV_{1}^{n}}  \ (eqn 23)\\ &lt;br /&gt;
{{\mu }_{2}}= &amp;amp; \frac{1}{KV_{2}^{n}}  \ (eqn 24)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{1}}=10,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{V}_{1}}=1&amp;lt;/math&amp;gt;  in Eqn. 23 and casting it in terms of  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
10,000= &amp;amp; \frac{1}{K}  \ (eqn 25) \\ &lt;br /&gt;
K = &amp;amp; \frac{1}{10,000}=0.0001  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{2}}=5,000&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{V}_{2}}=1.5&amp;lt;/math&amp;gt;  (because if one fails, then each survivor takes on an additional 0.5 units of load) and Eqn. 25 for  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  in Eqn.24 yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
5,000= &amp;amp; \frac{1}{0.0001\cdot {{(1.5)}^{n}}} \\ &lt;br /&gt;
0.5= &amp;amp; {{(1.5)}^{-n}} \\ &lt;br /&gt;
\ln (0.5)= &amp;amp; -n\ln (1.5) \\ &lt;br /&gt;
n = &amp;amp; 1.7095  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This also could have been computed in ALTA, as shown in Figure 5.20, or with the Load &amp;amp; Life Parameter Experimenter in BlockSim, as shown in Figure 5.21.&lt;br /&gt;
 &lt;br /&gt;
At this point, the parameters for the load sharing units have been computed and can be set, as shown in Figure 5.22.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The next step is to set the weight proportionality factor.  This factor defines the portion of the load that the particular item carries while operating, as well as the load that shifts to the remaining units upon failure of the item.  To illustrate, assume three units (1, 2 and 3) are in a load sharing container with weight proportionality factors of 1, 2 and 3 respectively (and a 1-out-of-3 requirement).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Unit 1 carries  &amp;lt;math&amp;gt;\left( \tfrac{1}{1+2+3} \right)=0.166&amp;lt;/math&amp;gt;  or 16.6% of the total load.&lt;br /&gt;
:•	Unit 2 carries  &amp;lt;math&amp;gt;\left( \tfrac{2}{1+2+3} \right)=0.333&amp;lt;/math&amp;gt;  or 33.3% of the total load.&lt;br /&gt;
:•	Unit 3 carries  &amp;lt;math&amp;gt;\left( \tfrac{3}{1+2+3} \right)=0.50&amp;lt;/math&amp;gt;  or 50% of the total load.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The actual load on each unit then becomes the product of the entire load defined for the container times the portion carried by that unit.  For example, if the container load is 100 lbs, then the portion assigned to Unit 1 will be  &amp;lt;math&amp;gt;100\cdot 0.166=16.6&amp;lt;/math&amp;gt;  lbs.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the current example, all units share the same load and thus have equal weight proportionality factors.  Because these factors are relative, if the same number is used for all three items then the results will be the same.  Thus, weight proportional factor is set equal to 1 for each item.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5.20.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.20 Calculation performed in ALTA.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.21_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.21 Quick Parameter Estimator &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.21_3.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.22 Quick Parameter Estimator results &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.23_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.23 Defining Weight Proportional Factor. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The last properties that need to be defined are the total load and the 2-out-of-3 redundancy.  The total load is dependent on how the parameters were computed.  In this case, total load was assumed to be 3 when the parameters were computed (i.e. the load per item was 1 when all worked and 1.5 when two worked).  This is defined at the container level, set No. of Paths required = 3.&lt;br /&gt;
When all of the parameters have been specified in BlockSim, the reliability at 1,000 hours can be determined.  From the Analytical QCP, this is found to be 98.57%.&lt;br /&gt;
&lt;br /&gt;
=Standby Components=&lt;br /&gt;
&lt;br /&gt;
In the previous section, the case of a system with load sharing components was presented.  This is a form of redundancy with dependent components. That is, the failure of one component affects the failure of the other(s).  This section presents another form of redundancy: standby redundancy.  In standby redundancy the redundant components are set to be under a lighter load condition (or no load) while not needed and under the operating load when they are activated.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In standby redundancy the components are set to have two states: an active state and a standby state.  Components in standby redundancy have two failure distributions, one for each state.  When in the standby state, they have a quiescent (or dormant) failure distribution and when operating, they have an active failure distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case that both quiescent and active failure distributions are the same, the units are in a simple parallel configuration (also called a hot standby configuration).  When the rate of failure of the standby component is lower in quiescent mode than in active mode, that is called a warm standby configuration.  When the rate of failure of the standby component is zero in quiescent mode (i.e. the component cannot fail when in standby), that is called a cold standby configuration.  &lt;br /&gt;
&lt;br /&gt;
===Simple Standby Configuration===&lt;br /&gt;
&lt;br /&gt;
Consider two components in a standby configuration.  Component 1 is the active component with a Weibull failure distribution with parameters  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Component 2 is the standby component.  When Component 2 is operating, it also has a Weibull failure distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Furthermore, assume the following cases for the quiescent distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1:  The quiescent distribution is the same as the active distribution (hot standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2:  The quiescent distribution is a Weibull  distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 2000 (warm standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3: The component cannot fail in quiescent mode (cold standby).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this case, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be obtained using the following equation:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(t)={{R}_{1}}(t)+\underset{0}{\overset{t}{\mathop \int }}\,{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x)\cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}dx   \ (eqn 26)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode, such that: &amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
&amp;lt;br&amp;gt;  &lt;br /&gt;
&amp;lt;math&amp;gt;{{R}_{2;SB}}(x)={{R}_{2;A}}({{t}_{e}})  \ (eqn 27)&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Eqn. 27 can be solved for  &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  and substituted into Eqn.26.&lt;br /&gt;
Figure 5.24 illustrates the example as entered in BlockSim using a standby container.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5.24.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.24 Standby container.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The active and standby blocks are within a container, which is used to specify standby redundancy.  Since the standby component has two distributions (active and quiescent), the Block Properties window of the standby block has two pages for specifying each one.  Figures 5.24 and 5.26 illustrate these pages.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system reliability results for 1000 hours are given in the following table:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-24.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Note that even though the  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  value for the quiescent distribution is the same as in the active distribution, it is possible that the two can be different. That is, the failure modes present during the quiescent mode could be different from the modes present during the active mode.  In that sense, the two distribution types can be different as well (e.g. lognormal when quiescent and Weibull when active).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In many cases when considering standby systems, a switching device may also be present that switches from the failed active component to the standby component.  The reliability of the switch can also be incorporated into Eqn.26, as presented in the next section.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim&#039;s System Reliability Equation window returns a single token for the reliability of units in a standby configuration.  This is the same as the load sharing case presented in the previous section. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image: Fig 5.25.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.25 Defining the active failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.26.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.26 Defining the quiescent failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Reliability of Standby Systems with a Switching Device===&lt;br /&gt;
&lt;br /&gt;
In many cases when dealing with standby systems, a switching device is present that will switch to the standby component when the active component fails.  Therefore, the failure properties of the switch must also be included in the analysis.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.26.2.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In most cases when the reliability of a switch is to be included in the analysis, two probabilities can be considered.  The first and most common one is the probability of the switch performing the action (i.e. switching) when requested to do so.  This is called Switch Probability per Request in BlockSim and is expressed as a static probability (e.g. 90%).  The second probability is the quiescent reliability of the switch.  This is the reliability of the switch as it ages (e.g. the switch might wear out with age due to corrosion, material degradation, etc.). Thus it is possible for the switch to fail before the active component fails.  However, a switch failure does not cause the system to fail, but rather causes the system to fail only if the switch is needed and the switch has failed.  For example, if the active component does not fail until the mission end time and the switch fails, then the system does not fail.  However, if the active component fails and the switch has also failed, then the system cannot be switched to the standby component and it therefore fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In analyzing standby components with a switching device, either or both failure probabilities (during the switching or while waiting to switch) can be considered for the switch, since each probability can represent different failure modes.  For example, the switch probability per request may represent software-related issues or the probability of detecting the failure of an active component, and the quiescent probability may represent wear-out type failures of the switch.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the formulation, consider the previous example that assumes perfect switching.  To examine the effects of including an imperfect switch, assume that when the active component fails there is a 90% probability that the switch will switch from the active component to the standby component.  In addition, assume that the switch can also fail due to a wear-out failure mode described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.7 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 5000.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , is given by the following equation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t)= &amp;amp; {{R}_{1}}(t) \\ &lt;br /&gt;
&amp;amp; +\underset{0}{\overset{t}{\mathop \int }}\,\{{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x) \\ &lt;br /&gt;
&amp;amp; \cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}\cdot {{R}_{SW;Q}}(x)\cdot {{R}_{SW;REQ}}(x)\}dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;Q}}&amp;lt;/math&amp;gt;  is the quiescent reliability of the switch.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;REQ}}&amp;lt;/math&amp;gt;  is the switch probability per request.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
This problem can be solved in BlockSim by including these probabilities in the container&#039;s properties, as shown in Figures fig23 and fig24.  In BlockSim, the standby container is acting as the switch.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.28_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.27 Standby container (switch) failure probabilities while attempting to switch &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.27.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.28 Standby container (switch) failure distribution while waiting to switch &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that there are additional properties that can be specified in BlockSim for a switch, such as Switch Restart Probability, Finite Restarts and Switch Delay Time.  In many applications, the switch is re-tested (or re-cycled) if it fails to switch the first time.  In these cases, it might be possible that it switches in the second or third, or  &amp;lt;math&amp;gt;{{n}^{th}}&amp;lt;/math&amp;gt; attempt.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The Switch Restart Probability specifies each additional attempt&#039;s probability of successfully switching and the Finite Restarts specifies the total number of attempts.  Note that the Switch Restart Probability specifies the probability of success of each trial (or attempt).  The probability of success of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  consecutive trials is calculated by BlockSim using the binomial distribution and this probability is then incorporated into Eqn. (stb2a).  The Switch Delay Time property is related to repairable systems and is considered in BlockSim only when using simulation.  When using the analytical solution (i.e. for a non-repairable system), this property is ignored.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving the analytical solution (as given by Eqn. stb2a), the following results are obtained.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-30.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From the table above, it can be seen that the presence of a switching device has a significant effect on the reliability of a standby system.  It is therefore important when modeling standby redundancy to incorporate the switching device reliability properties.  It should be noted that this methodology is not the same as treating the switching device as another series component with the standby subsystem.  This would be valid only if the failure of the switch resulted in the failure of system (e.g. switch failing open).  In Eqn. (stb2a), the Switch Probability per Request and quiescent probability are present only in the second term of the equation.  Treating these two failure modes as a series configuration with the standby subsystem would imply that they are also present when the active component is functioning (i.e. first term of Eqn. stb2a).  This is invalid and would result in the underestimation of the reliability of the system.  In other words, these two failure modes become significant only when the active component fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As an example, and if we consider the warm standby case, the reliability of the system without the switch is 70.57% at 1000 hours.  If the system was modeled so that the switching device was in series with the warm standby subsystem, the result would have been:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(1000)= &amp;amp; {{R}_{Standby}}(1000)\cdot {{R}_{sw,Q(1000)}}\cdot {{R}_{sw,req}} \\ &lt;br /&gt;
= &amp;amp; 0.7057\cdot 0.9372\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.5952  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case where a switch failure mode causes the standby subsystem to fail, then this mode can be modeled as an individual block in series with the standby subsystem.&lt;br /&gt;
&lt;br /&gt;
===Example 6===&lt;br /&gt;
Consider a car with four new tires and a full-size spare.  Assume the following failure characteristics:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires follow a Weibull distribution with a  ..  and an  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  40,000 miles while on the car due to wear.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires also have a probability of failing due to puncture or other causes.  For this, assume a constant rate for this occurrence with a probability of 1 every 50,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	When not on the car (i.e. is a spare), a tire&#039;s probability of failing also has a Weibull distribution with a  &amp;lt;math&amp;gt;\beta =&amp;lt;/math&amp;gt;  2 and  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  120,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Assume a mission of 1,000 miles.  If a tire fails during this trip, it will be replaced with the spare.  However, the spare will not be repaired during the trip.  In other words, the trip will continue with the spare on the car and if the spare fails the system will fail.  Determine the probability of system failure.&lt;br /&gt;
====Solution to Example 6====&lt;br /&gt;
Active failure distribution for tires:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to wear-out, Weibull  &amp;lt;math&amp;gt;\beta =4&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =40,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to random puncture, exponential  &amp;lt;math&amp;gt;\mu =50,000.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	The quiescent failure distribution is a Weibull distribution with &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =120,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The block diagram for each tire has two blocks in series, one block representing the wear-out mode and the other the random puncture mode, as shown next:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:small5.gif|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
There are five tires, four active and one standby (represented in the diagram by a standby container with a 4-out-of-5 requirement), as shown next: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BStirecontainer.png|thumb|center|400px|]]&lt;br /&gt;
 &lt;br /&gt;
For the standby Wear block, set the active failure and the quiescent distributions, but for the Puncture block, set only the active puncture distribution (because the tire cannot fail due to puncture while stored).  Using BlockSim, the probability of system failure is found to be 0.003 or 0.3%. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Note Regarding Numerical Integration Solutions=&lt;br /&gt;
&lt;br /&gt;
Load sharing and standby solutions in BlockSim are performed using numerical integration routines.  As with any numerical analysis routine, the solution error depends on the number of iterations performed, the step size chosen and related factors, plus the behavior of the underlying function.  By default, BlockSim uses a certain set of preset factors.  In general, these defaults are sufficient for most problems.  If a higher precision or verification of the precision for a specific problem is required, BlockSim&#039;s preset options can be modified and/or the integration error can be assessed using the Integration Parameters... option for each container.  For more details, you can refer to the documentation on the Algorithm Setup window in the BlockSim 7 User&#039;s Guide.&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15559</id>
		<title>Time-Dependent System Reliability (Analytical)</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=Time-Dependent_System_Reliability_(Analytical)&amp;diff=15559"/>
		<updated>2012-02-13T22:08:42Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: /* Reliability of Standby Systems with a Switching Device */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Template:bsbook|5}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the previous chapter, different system configuration types were examined, as well as different methods for obtaining the system&#039;s reliability function analytically.  Because the reliabilities in the problems presented were treated as probabilities (e.g.  &amp;lt;math&amp;gt;P(A)&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; ), the reliability values and equations presented were referred to as static (not time-dependent).  Thus, in the prior chapter, the life distributions of the components were not incorporated in the process of calculating the system reliability.  In this chapter, time dependency in the reliability function will be introduced.  We will develop the models necessary to observe the reliability over the life of the system, instead of at just one point in time.  In addition, performance measures such as failure rate, MTTF and warranty time will be estimated for the entire system.  The methods of obtaining the reliability function analytically remain identical to the ones presented in the previous chapter, with the exception that the reliabilities will be functions of time.  In other words, instead of dealing with  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt; , we will use  &amp;lt;math&amp;gt;{{R}_{i}}(t)&amp;lt;/math&amp;gt; .  All examples in this chapter assume that no repairs are performed on the components.  &lt;br /&gt;
&lt;br /&gt;
=Analytical Life Predictions=&lt;br /&gt;
The analytical approach presented in the prior chapter involved the determination of a mathematical expression that describes the reliability of the system, expressed in terms of the reliabilities of its components.  So far we have estimated only static system reliability (at a fixed time).  For example, in the case of a system with three components in series, the system&#039;s reliability equation was given by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}={{R}_{1}}\cdot {{R}_{2}}\cdot {{R}_{3}}  \ (eqn 1)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The values of  &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{R}_{2}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{3}}&amp;lt;/math&amp;gt;  were given for a common time and the reliability of the system was estimated for that time.  However, since the component failure characteristics can be described by distributions, the system reliability is actually time-dependent.  In this case, Eqn. (1) can be rewritten as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability of the system for any mission time can now be estimated.  Assuming a Weibull life distribution for each component, Eqn. (1) can now be expressed in terms of each component&#039;s reliability function, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}}\cdot {{e}^{-{{\left( \tfrac{t}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the same manner, any life distribution can be substituted into the system reliability equation.  Suppose that the times-to-failure of the first component are described with a Weibull distribution, the times-to-failure of the second component with an exponential distribution and the times-to-failure of the third component with a normal distribution.  Then Eqn. (1) can be written as:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot \left[ 1-\Phi \left( \frac{t-{{\mu }_{3}}}{{{\sigma }_{3}}} \right) \right]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
It can be seen that the biggest challenge is in obtaining the system&#039;s reliability function in terms of component reliabilities, which has already been discussed in depth.  Once this has been achieved, calculating the reliability of the system for any mission duration is just a matter of substituting the corresponding component reliability functions into the system reliability equation.&lt;br /&gt;
===Advantages of the Analytical Method===&lt;br /&gt;
The primary advantage of the analytical solution is that it produces a mathematical expression that describes the reliability of the system.  Once the system&#039;s reliability function has been determined, other calculations can then be performed to obtain metrics of interest for the system. Such calculations include:  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of warranty periods.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s failure rate.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Determination of the system&#039;s MTTF.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In addition, optimization and reliability allocation techniques can be used to aid engineers in their design improvement efforts.  Another advantage in using analytical techniques is the ability to perform static calculations and analyze systems with a mixture of static and time-dependent components.  Finally, the reliability importance of components over time can be calculated with this methodology.&lt;br /&gt;
===Disadvantages of the Analytical Method===&lt;br /&gt;
The biggest disadvantage of the analytical method is that formulations can become very complicated.  The more complicated a system is, the larger and more difficult it will be to analytically formulate an expression for the system&#039;s reliability.  For particularly detailed systems this process can be quite time-consuming, even with the use of computers.  Furthermore, when the maintainability of the system or some of its components must be taken into consideration, analytical solutions become intractable.  In these situations, the use of simulation methods may be more advantageous than attempting to develop a solution analytically.  Simulation methods are presented in later chapters.&lt;br /&gt;
===Looking at a Simple &#039;&#039;Complex&#039;&#039; System Analytically===&lt;br /&gt;
&lt;br /&gt;
The complexity involved in an analytical solution can be best illustrated by looking at the simple &#039;&#039;complex&#039;&#039; system with 15 components, as shown in Figure 5.1.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1.png|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.1 An RBD of a complex system.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system reliability for this system (computed using BlockSim) is shown next.  The first solution is provided using BlockSim&#039;s symbolic solution.  In symbolic mode, BlockSim breaks the equation into segments, identified by tokens, that need to be substituted into the final system equation for a complete solution.  This creates algebraic solutions that are more compact than if the substitutions were made.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; D2\cdot D3\cdot {{R}_{L}} \\ &lt;br /&gt;
D3= &amp;amp; +{{R}_{K}}\cdot IK \\ &lt;br /&gt;
IK= &amp;amp; +{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}} \\ &lt;br /&gt;
D2 = &amp;amp; +{{R}_{A}}\cdot {{R}_{E}}\cdot IE \\ &lt;br /&gt;
IE = &amp;amp; -D1\cdot {{R}_{M}}\cdot {{R}_{N}}+{{R}_{M}}\cdot {{R}_{N}}+D1 \\ &lt;br /&gt;
D1 = &amp;amp; +{{R}_{D}}\cdot ID \\ &lt;br /&gt;
ID = &amp;amp; -{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting the terms yields: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; {{R}_{A}}\cdot {{R}_{E}}\cdot {{R}_{L}}\cdot {{R}_{K}} \\ &lt;br /&gt;
&amp;amp; \cdot \{({{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})\cdot {{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}}-{{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}}\} \\ &lt;br /&gt;
&amp;amp; \cdot \{{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}}\}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
BlockSim&#039;s automatic algebraic simplification would yield the following format for the above solution: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{System}}= &amp;amp; (({{R}_{A}}\cdot {{R}_{E}}(-({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})){{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{M}}\cdot {{R}_{N}} \\ &lt;br /&gt;
&amp;amp; +({{R}_{D}}(-{{R}_{B}}\cdot {{R}_{C}}+{{R}_{B}}+{{R}_{C}})))) \\ &lt;br /&gt;
&amp;amp; ({{R}_{K}}({{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{O}}\cdot {{R}_{G}}\cdot {{R}_{F}}-{{R}_{I}}\cdot {{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{I}}\cdot {{R}_{O}}\cdot {{R}_{F}}\cdot {{R}_{H}}-{{R}_{J}}\cdot {{R}_{G}}\cdot {{R}_{F}}\cdot {{R}_{H}} \\ &lt;br /&gt;
&amp;amp; +RI\cdot {{R}_{O}}\cdot {{R}_{F}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{I}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{F}}\cdot {{R}_{H}}+{{R}_{J}}\cdot {{R}_{G}})){{R}_{L}})  \ (eqn 2)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this equation, each  &amp;lt;math&amp;gt;{{R}_{i}}&amp;lt;/math&amp;gt;  represents the reliability function of a block.  For example, if  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  has a Weibull distribution, then each  &amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;  and so forth.  Substitution of each component&#039;s reliability function in Eqn.2 will result in an analytical expression for the system reliability as a function of time, or  &amp;lt;math&amp;gt;{{R}_{s}}(t)&amp;lt;/math&amp;gt; , which is the same as  &amp;lt;math&amp;gt;(1-cd{{f}_{System}}).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Obtaining Other Functions of Interest===&lt;br /&gt;
Once the system reliability equation (or the cumulative density function,  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt; ) has been determined, other functions and metrics of interest can be derived.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Consider the following simple system:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-2.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.2 Simple two-component system. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Furthermore, assume that component 1 follows an exponential distribution with a mean of 10,000 (&amp;lt;math&amp;gt;\mu =10,000,&amp;lt;/math&amp;gt;   &amp;lt;math&amp;gt;\lambda =1/10,000)&amp;lt;/math&amp;gt;  and component 2 follows a Weibull distribution with  &amp;lt;math&amp;gt;\beta =6&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt; .  The reliability equation of this system is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}  \ (eqn 3) &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The system  &amp;lt;math&amp;gt;cdf&amp;lt;/math&amp;gt;  is:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{F}_{S}}(t)= &amp;amp; 1-({{R}_{1}}(t)\cdot {{R}_{2}}(t)) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\lambda t}}\cdot {{e}^{-{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \right) \\ &lt;br /&gt;
= &amp;amp; 1-\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
{{analytical system pdf}}&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability====&lt;br /&gt;
Conditional reliability is the probability of a system successfully completing another mission following the successful completion of a previous mission.  The time of the previous mission and the time for the mission to be undertaken must be taken into account for conditional reliability calculations.  The system&#039;s conditional reliability function is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(T,t)=\frac{R(T+t)}{R(T)}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Eqn.6 gives the reliability for a new mission of duration  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  having already accumulated  &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt;  hours of operation up to the start of this new mission. The system is evaluated to assure that it will start the next mission successfully.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2, the reliability for mission of  &amp;lt;math&amp;gt;t=1,000&amp;lt;/math&amp;gt;  hours, having an age of  &amp;lt;math&amp;gt;T=500&amp;lt;/math&amp;gt;  hours, is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(T=500,t=1000)= &amp;amp; \frac{R(T+t)}{R(T)} \\ &lt;br /&gt;
= &amp;amp; \frac{R(1500)}{R(500)} \\ &lt;br /&gt;
= &amp;amp; \frac{{{e}^{-\tfrac{1500}{10,000}}}\cdot {{e}^{-{{\left( \tfrac{1500}{10,000} \right)}^{6}}}}}{{{e}^{-\tfrac{500}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{500}{10,000} \right)}^{6}}}}} \\ &lt;br /&gt;
= &amp;amp; 0.9048=90.48%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.3.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.4.png|thumb|center|400px|]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Conditional Reliability for Components====&lt;br /&gt;
&lt;br /&gt;
Now in this formulation, it was assumed that the accumulated age was equivalent for both units. That is, both started life at zero and aged to 500.  It is possible to consider an individual component that has already accumulated some age (used component) in the same formulation.  To illustrate this, assume that component 2 started life with an age of T=100.  Then the reliability equation of the system, as given in Eqn.3, would need to be modified to include a conditional term for 2, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{R}_{1}}(t)\cdot \frac{{{R}_{2}}({{T}_{2}}+t)}{{{R}_{2}}({{T}_{2}})} \ (eqn 7)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In BlockSim, the start age input box may be used to specify a starting age greater than zero.&lt;br /&gt;
{{system failure rate analytical}}&lt;br /&gt;
&lt;br /&gt;
====System Mean Life (Mean Time To Failure)====&lt;br /&gt;
The mean life (or mean time to failure, MTTF) can be obtained by integrating the system reliability function from zero to infinity: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;MTTF=\int_{0}^{\infty }{{R}_{s}}\left( t \right)dt   \ (eqn 10)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mean time is a performance index and does not provide any information about the behavior of the failure distribution of the system.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }\left( {{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}} \right)dt \\ &lt;br /&gt;
= &amp;amp; 5978.9  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
====Warranty Period and BX Life====&lt;br /&gt;
Sometimes it is desirable to know the time value associated with a certain reliability.  Warranty periods are often calculated by determining what percentage of the failure population can be covered financially and estimating the time at which this portion of the population will fail.  Similarly, engineering specifications may call for a certain BX life, which also represents a time period during which a certain proportion of the population will fail.  For example, the B10 life is the time in which 10% of the population will fail.  &lt;br /&gt;
This is obtained by setting  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  to the desired value and solving for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}\left( t \right)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To compute the time by which reliability would be equal to 90%, Eqn.11 is recast as follows and solved for  &amp;lt;math&amp;gt;t.&amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;0.90={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this case,  &amp;lt;math&amp;gt;t=1053.59&amp;lt;/math&amp;gt; .  Equivalently, the B10 life for this system is also  &amp;lt;math&amp;gt;1053.59&amp;lt;/math&amp;gt; .&lt;br /&gt;
Except for some trivial cases, a closed form solution for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  cannot be obtained.   Thus, it is necessary to solve for  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  using numerical methods.  BlockSim uses numerical methods.&lt;br /&gt;
&lt;br /&gt;
===Example 1===&lt;br /&gt;
Consider a system consisting of three exponential units in series with the following failure rates (in failures per hour):  &amp;lt;math&amp;gt;{{\lambda }_{1}}&amp;lt;/math&amp;gt;  = 0.0002,  &amp;lt;math&amp;gt;{{\lambda }_{2}}&amp;lt;/math&amp;gt;  = 0.0005 and  &amp;lt;math&amp;gt;{{\lambda }_{3}}&amp;lt;/math&amp;gt;  = 0.0001.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the reliability equation for the system.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the reliability of the system after 150 hours of operation?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s  &amp;lt;math&amp;gt;pdf.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	Obtain the system&#039;s failure rate equation.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What is the MTTF for the system?&amp;lt;br&amp;gt;&lt;br /&gt;
:•	What should the warranty period be for a 90% reliability?&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution to Example 1====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The analytical expression for the reliability of the system is given by:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{R}_{1}}(t)\cdot {{R}_{2}}(t)\cdot {{R}_{3}}(t) \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\lambda }_{1}}t}}\cdot {{e}^{-{{\lambda }_{2}}t}}\cdot {{e}^{-{{\lambda }_{1}}t}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
At 150 hours of operation, the reliability of the system is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; {{e}^{-(0.0002+0.0005+0.0001)150}} \\ &lt;br /&gt;
= &amp;amp; 0.8869\text{ or }88.69%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.12 is taken with respect to time, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; -\frac{d[{{R}_{s}}(t)]}{dt} \\ &lt;br /&gt;
= &amp;amp; -\frac{d\left[ {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}} \right]}{dt} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
:•	The system&#039;s failure rate can now be obtained simply by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn. 13 by the system&#039;s reliability function given in Eqn.12, and:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}\left( t \right)= &amp;amp; \frac{{{f}_{s}}\left( t \right)}{{{R}_{s}}\left( t \right)} \\ &lt;br /&gt;
= &amp;amp; \frac{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})\cdot {{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}}{{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}} \\ &lt;br /&gt;
= &amp;amp; ({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}) \\ &lt;br /&gt;
= &amp;amp; 0.0008\text{ }fr/hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Combining Eqn.10 and Eqn. 12, the system&#039;s MTTF can be obtained:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
MTTF= &amp;amp; \int_{0}^{\infty }{{R}_{s}}\left( t \right)dt \\ &lt;br /&gt;
= &amp;amp; \int_{0}^{\infty }{{e}^{-({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})t}}dt \\ &lt;br /&gt;
= &amp;amp; \frac{1}{({{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}})} \\ &lt;br /&gt;
= &amp;amp; 1250\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Solving Eqn. 12 with respect to time will yield the corresponding warranty period for a 90% reliability.  In this case, the system reliability equation is simple and a closed form solution exists.  The warranty time can now be found by solving:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
t= &amp;amp; -\frac{\ln (R)}{{{\lambda }_{1}}+{{\lambda }_{2}}+{{\lambda }_{3}}} \\ &lt;br /&gt;
= &amp;amp; -\frac{\ln (0.9)}{0.0008} \\ &lt;br /&gt;
= &amp;amp; 131.7\text{ }hr  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the warranty period should be 132 hours.&lt;br /&gt;
&lt;br /&gt;
===Example 2===&lt;br /&gt;
Consider the system shown in Figure 5.5.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.5.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.5 Complex bridge system in Example 2. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Components  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  through  &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;  are Weibull distributed with  &amp;lt;math&amp;gt;\beta =1.2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1230&amp;lt;/math&amp;gt;  hours.  The starting and ending blocks cannot fail.  &amp;lt;br&amp;gt;&lt;br /&gt;
Determine the following:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability equation for the system and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  and its corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The system&#039;s failure rate equation and the corresponding plot.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt; .&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The warranty time for a 90% reliability.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The reliability for a 200-hour mission, if it is known that the system has already successfully operated for 200 hours.&amp;lt;br&amp;gt;&lt;br /&gt;
====Solution====&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The first step is to obtain the reliability function for the system.  The methods described in the previous chapter can be employed, such as the event space or path-tracing methods.  Using BlockSim, the following reliability equation is obtained:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; ({{R}_{Start}}\cdot {{R}_{End}}(2{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}))  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that since the starting and ending blocks cannot fail,  &amp;lt;math&amp;gt;{{R}_{Start}}=1&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{R}_{End}}=1,&amp;lt;/math&amp;gt;  Eqn.14 can be reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{s}}(t)= &amp;amp; 2\cdot {{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}-{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{A}}\cdot {{R}_{D}}\cdot {{R}_{B}}\cdot {{R}_{E}}-{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; -{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}\cdot {{R}_{E}}+{{R}_{A}}\cdot {{R}_{C}}\cdot {{R}_{E}} \\ &lt;br /&gt;
&amp;amp; +{{R}_{D}}\cdot {{R}_{C}}\cdot {{R}_{B}}+{{R}_{A}}\cdot {{R}_{D}}+{{R}_{B}}\cdot {{R}_{E}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{A}}&amp;lt;/math&amp;gt;  is the reliability equation for Component A, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{A}}(t)={{e}^{-{{\left( \tfrac{t}{{{\eta }_{A}}} \right)}^{{{\beta }_{A}}}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{B}}&amp;lt;/math&amp;gt;  is the reliability equation for Component  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; , etc.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since the components in this example are identical, the system reliability equation can be further reduced to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2R{{(t)}^{2}}+2R{{(t)}^{3}}-5R{{(t)}^{4}}+2R{{(t)}^{5}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Or, in terms of the failure distribution:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{s}}(t)=2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.6.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.6.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.6 Reliability plot for the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
In order to obtain the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt; , the derivative of the reliability equation given in Eqn.18 is taken with respect to time, resulting in: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{f}_{s}}(t)= &amp;amp; 4\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}} \\ &lt;br /&gt;
&amp;amp; -20\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \frac{\beta }{\eta }{{\left( \frac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  can now be plotted for different time values,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , as shown in Figure 5.7.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system&#039;s failure rate can now be obtained by dividing the system&#039;s  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  given in Eqn.19 by the system&#039;s reliability function given in Eqn.18, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\lambda }_{s}}(t)= &amp;amp; \frac{4\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+6\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}} \\ &lt;br /&gt;
&amp;amp; +\frac{-20\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+10\cdot \tfrac{\beta }{\eta }{{\left( \tfrac{t}{\eta } \right)}^{\beta -1}}{{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}{2\cdot {{e}^{-2{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-3{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}-5\cdot {{e}^{-4{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}+2\cdot {{e}^{-5{{\left( \tfrac{t}{\eta } \right)}^{\beta }}}}}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The corresponding plot is given in Figure 5.8.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.7.gif|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.7 &#039;&#039;pdf&#039;&#039; plot for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of the system is obtained by integrating the system&#039;s reliability function given by Eqn. 18 from time zero to infinity, as given by Eqn. 10.  Using BlockSim&#039;s Analytical QCP, an  &amp;lt;math&amp;gt;MTTF&amp;lt;/math&amp;gt;  of 1007.8 hours is calculated, as shown in Figure 5.9.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The warranty time can be obtained by solving Eqn. 18 with respect to time for a system reliability  &amp;lt;math&amp;gt;{{R}_{s}}=0.9&amp;lt;/math&amp;gt; .  Using the Analytical QCP and selecting the &amp;lt;br&amp;gt;&lt;br /&gt;
Warranty Time option, a time of 372.72 hours is obtained, as shown in Figure 5.10.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.8.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.8 Failure rate for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.9.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.9 MTTF of the system in Figure 5.5. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.10.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.10 Time at which &#039;&#039;R&#039;&#039;=0.9 or 90% for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.11.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.11 Conditional reliability calculation for the system in Figure 5.5.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Lastly, the conditional reliability can be obtained using Eqn.6 and Eqn.18, or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(200,200)= &amp;amp; \frac{R(400)}{R(200)} \\ &lt;br /&gt;
= &amp;amp; \frac{0.883825}{0.975321} \\ &lt;br /&gt;
= &amp;amp; 0.906189  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated using BlockSim&#039;s Analytical QCP, as shown in Figure 5.11.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Approximating the System CDF=&lt;br /&gt;
&lt;br /&gt;
In many cases, it is valuable to fit a distribution that represents the system&#039;s times-to-failure.  This can be useful when the system is part of a larger assembly and may be used for repeated calculations or in calculations for other systems.  In cases such as this, it can be useful to characterize the system&#039;s behavior by fitting a distribution to the overall system and calculating parameters for this distribution.   This is equivalent to fitting a single distribution to describe  &amp;lt;math&amp;gt;{{R}_{S}}(t&amp;lt;/math&amp;gt; ).  In essence, it is like reducing the entire system to a component in order to simplify calculations.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
For the system in Figure 5.2: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{S}}(t)={{e}^{-\tfrac{1}{10,000}t}}\cdot {{e}^{-{{\left( \tfrac{t}{10,000} \right)}^{6}}}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
To compute an approximate reliability function for this system,  &amp;lt;math&amp;gt;{{R}_{A}}(t)\simeq {{R}_{S}}(t)&amp;lt;/math&amp;gt; , one would compute  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  pairs of reliability and time values and then fit a single distribution to the data, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 10,396.7)=10% \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 9,361.9)=20% \\ &lt;br /&gt;
&amp;amp; ... \\ &lt;br /&gt;
{{R}_{S}}(t= &amp;amp; 1,053.6)=90%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
A single distribution,  &amp;lt;math&amp;gt;{{R}_{A}}(t)&amp;lt;/math&amp;gt; , that approximates  &amp;lt;math&amp;gt;{{R}_{S}}(t)&amp;lt;/math&amp;gt;  can now be computed from these pairs using life data analysis methods.  If using the Weibull++ software, one would enter the values as free form data.&lt;br /&gt;
&lt;br /&gt;
===Example 3===&lt;br /&gt;
Compute a single Weibull distribution approximation for the system in Example 2.&lt;br /&gt;
====Solution to Example 3====&lt;br /&gt;
The system in the previous example, shown in Figure 5.5, can be approximated by use of a 2-parameter Weibull distribution with  &amp;lt;math&amp;gt;\beta =2.02109&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =1123.51&amp;lt;/math&amp;gt; .  In BlockSim, this is accomplished by representing the entire system as one distribution by going to the Distribution Fit window Figure 5.12. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.13.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.12 Representing a system with a distribution.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.12.PNG|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.13 Distribution Fitting window.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
by clicking the Distribution Fit Window, the Distribution Estimator window will appear (Figure 5.13).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this window you can select a distribution to represent the data. BlockSim will then generate a number of system failure times based on the system&#039;s reliability function. The system&#039;s reliability function can be used to solve for a time value associated with that unreliability value. The distribution of the generated time values can then be fitted to a probability distribution function.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consider a value of  &amp;lt;math&amp;gt;F(t)=0.11&amp;lt;/math&amp;gt; .  Using the system&#039;s reliability equation and solving for time, the corresponding time-to-failure for a 0.11 unreliability can be calculated.  &amp;lt;br&amp;gt;&lt;br /&gt;
For the system of Example 2, the time for a 0.11 unreliability is 389.786 hours.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
When enough points have been generated, the selected distribution will be fitted to this data set and the distribution&#039;s parameters will be returned.  In addition, if ReliaSoft&#039;s Weibull++ is installed, the generated data can be viewed/analyzed using a Weibull++ instance, as shown in Figure 5.14.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.14.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.14 Using Weibull++ to calculate distribution parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
It is recommended that the analyst examine the fit to ascertain the applicability of the approximation.&lt;br /&gt;
&lt;br /&gt;
=Duty Cycle=&lt;br /&gt;
&lt;br /&gt;
Components of a system may not operate continuously during a system&#039;s mission, or may be subjected to loads greater or lesser than the rated loads during system operation.  To model this, a factor called the Duty Cycle ( &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt; ) is used.  The duty cycle may also be used to account for changes in environmental stress, such as temperature changes, that may effect the operation of a component.  The duty cycle is a positive value, with a default value of 1 representing continuous operation at rated load, and any values other than 1 representing other load values with respect to the rated load value (or total operating time).   A duty cycle value higher than 1 indicates a load in excess of the rated value.  A duty cycle value lower than 1 indicates that the component is operating at a load lower than the rated load or not operating continuously during the system&#039;s mission.  For instance, a duty cycle of 0.5 may be used for a component that operates only half of the time during the system&#039;s mission.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The reliability metrics for a component with a duty cycle are calculated as follows. Let  &amp;lt;math&amp;gt;{{d}_{c}}&amp;lt;/math&amp;gt;  represent the duty cycle during a particular mission of the component,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt;  represent the mission time and  &amp;lt;math&amp;gt;{t}&#039;&amp;lt;/math&amp;gt;  represent the accumulated age. Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{t}&#039;={{d}_{c}}\times t&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The reliability equation for the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R({t}&#039;)=R({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The component &#039;&#039;pdf&#039;&#039; is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;f({t}&#039;)=-\frac{d(R({t}&#039;))}{dt}=-\frac{d(R({{d}_{c}}\times t))}{dt}={{d}_{c}}f({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The failure rate of the component is:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\lambda ({t}&#039;)=\frac{f({t}&#039;)}{R({t}&#039;)}=\frac{{{d}_{c}}f({{d}_{c}}\times t)}{R({{d}_{c}}\times t)}={{d}_{c}}\lambda ({{d}_{c}}\times t)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Example 4===&lt;br /&gt;
Consider a computer system with three components: a processor, a hard drive and a CD drive in series as shown next.  Assume that all three components follow a Weibull failure distribution with the parameters  &amp;lt;math&amp;gt;{{\beta }_{1}}=1.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{1}}=5000&amp;lt;/math&amp;gt;  for the processor,  &amp;lt;math&amp;gt;{{\beta }_{2}}=2.5&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{2}}=3000&amp;lt;/math&amp;gt;  for the hard drive, and  &amp;lt;math&amp;gt;{{\beta }_{3}}=2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{\eta }_{3}}=4000&amp;lt;/math&amp;gt;  for the CD drive.  Determine the reliability of the computer system after one year (365 days) of operation, assuming that the CD drive is used only 30% of the time.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5ex4.png|thumb|center|300px|]]&lt;br /&gt;
&lt;br /&gt;
====Solution to Example 4====&lt;br /&gt;
The reliability of the processor after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	  {{R}_{processor}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{1}}} \right)}^{{{\beta }_{1}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{5000} \right)}^{1.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\text{ or }98.05%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the hard drive after 365 days of operation is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{harddrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{365}{{{\eta }_{2}}} \right)}^{{{\beta }_{2}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{365}{3000} \right)}^{2.5}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9948\text{ or }99.48%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
The reliability of the CD drive after 365 days of operation (taking into account the 30% operation using a duty cycle of 0.3) is given by:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{CDdrive}}(365)= &amp;amp; {{e}^{-{{\left( \tfrac{{{d}_{c}}\times 365}{{{\eta }_{3}}} \right)}^{{{\beta }_{3}}}}}} \\ &lt;br /&gt;
	  = &amp;amp; {{e}^{-{{\left( \tfrac{0.3\times 365}{4000} \right)}^{2}}}} \\ &lt;br /&gt;
	  = &amp;amp; 0.9993\text{ or }99.93%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.15.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.15 Result for the computer system reliability.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
Thus the reliability of the computer system after 365 days of operation is:&lt;br /&gt;
	&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
	   {{R}_{s}}(365)= &amp;amp; {{R}_{processor}}(365)\cdot {{R}_{harddrive}}(365)\cdot {{R}_{CDdrive}}(365) \\ &lt;br /&gt;
	  = &amp;amp; 0.9805\cdot 0.9948\cdot 0.9993 \\ &lt;br /&gt;
	  = &amp;amp; 0.9747\text{ or }97.47%  &lt;br /&gt;
	\end{align}&amp;lt;/math&amp;gt;	&lt;br /&gt;
&amp;lt;br&amp;gt;	&lt;br /&gt;
This result can be obtained in BlockSim as shown in Figure DutyCycleExResults.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Load Sharing=&lt;br /&gt;
As presented in earlier chapters, a reliability block diagram (RBD) allows you to graphically represent how the components within a system are reliability-wise connected.  In most cases, independence is assumed across the components within the system.  For example, the failure of component A does not affect the failure of component B.  However, if a system consists of components that are sharing a load, then the assumption of independence no longer holds true.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If one component fails, then the component(s) that are still operating will have to assume the failed unit&#039;s portion of the load.  Therefore, the reliabilities of the surviving unit(s) will change.  Calculating the system reliability is no longer an easy proposition.  In the case of load sharing components, the change of the failure distributions of the surviving components must be known in order to determine the system&#039;s reliability.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate this, consider the a system of two units connected reliability-wise in parallel (Figure 5.16).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-16.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.16 Two units connected reliability-wise in parallel.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Assume that the units must supply an output of 8 volts and that if both units are operational, each unit is to supply 50% of the total output.  If one of the units fails, then the surviving unit supplies 100%.  Furthermore, assume that having to supply the entire load has a negative impact on the reliability characteristics of the surviving unit.  Since the reliability characteristics of the unit change based on whether both or only one is operating, a life distribution along with a life-stress relationship (as discussed in Chapter 3) will be needed to model each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the steps needed, we will create the model starting from raw data.  Assume that a total of 20 units were tested to failure at 7, 10 and 15 volts.  The test data set is presented in the next table.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-17.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
For this example, Units 1 and 2 are the same component.  Therefore, only one set of data was collected.  However, it is possible that the load sharing components in a system may not be the same.  If that were the case, data would need to be collected for each component.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The data set in Table 1 was analyzed using ReliaSoft&#039;s ALTA software (as shown in Figure 5.17) with the Inverse Power Law as the underlying life-stress relationship and Weibull as the life distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The estimated model parameters,  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; , are shown next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
\beta = &amp;amp; 1.9239 \\ &lt;br /&gt;
K= &amp;amp; 3.2387\times {{10}^{-7}} \\ &lt;br /&gt;
n= &amp;amp; 3.4226  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:Or: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}} \\ &lt;br /&gt;
= &amp;amp; {{e}^{-{{\left( 3.2387\times {{10}^{-7}}S_{1}^{3.4226}t \right)}^{1.9239}}}}  \ (eqn 20)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{f}_{1}}(t,{{S}_{1}})=\beta KS_{1}^{n}{{\left( KS_{1}^{n}t \right)}^{\beta -1}}{{e}^{-{{\left( KS_{1}^{n}t \right)}^{\beta }}}}  \ (eqn 21 )&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And for this case:&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t,{{S}_{1}})= &amp;amp; {{R}_{2}}(t,{{S}_{2}})  \\ &lt;br /&gt;
{{f}_{1}}(t,{{S}_{1}})= &amp;amp; {{f}_{2}}(t,{{S}_{2}})  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.17.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.17 Using ALTA to calculate component parameters.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 5.18 shows a plot of Eqn.20.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Now that the failure properties have been determined using the test data, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be calculated using the following equation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t,S)= &amp;amp; {{R}_{1}}(t,{{S}_{1}})\cdot {{R}_{2}}(t,{{S}_{2}}) \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{1}}\left( x,{{S}_{1}} \right)\cdot {{R}_{2}}(x,{{S}_{2}})\cdot \left( \frac{{{R}_{2}}({{t}_{1e}}+(t-x),S)}{{{R}_{2}}({{t}_{1e}},S)} \right)dx \\ &lt;br /&gt;
&amp;amp; +\underset{o}{\overset{t}{\mathop \int }}\,{{f}_{2}}\left( x,{{S}_{2}} \right)\cdot {{R}_{1}}(x,{{S}_{1}})\cdot \left( \frac{{{R}_{1}}({{t}_{2e}}+(t-x),S)}{{{R}_{1}}({{t}_{2e}},S)} \right)dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:Where: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{S}_{1}}= &amp;amp; {{P}_{1}}S \\ &lt;br /&gt;
{{S}_{2}}= &amp;amp; {{P}_{2}}S  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-18.png|thumb|center|300px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig. 5.18 Reliability curves for different voltage output conditions. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.png|thumb|center|395px]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  is the total load (or required output).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{P}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{P}_{2}}&amp;lt;/math&amp;gt;  are the portion of the total load that each unit supports when both units are operational.  In this case,  &amp;lt;math&amp;gt;{{P}_{1}}={{P}_{2}}=0.5=50%.&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{S}_{2}}&amp;lt;/math&amp;gt;  represent the portions of the load that Unit 1 and Unit 2 must support when both units are operational.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{{{1}_{e}}}}&amp;lt;/math&amp;gt;  is the equivalent operating time for Unit 1 if it had been operating at  &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;  instead of  &amp;lt;math&amp;gt;{{S}_{1}}&amp;lt;/math&amp;gt; .  A graphical representation of the equivalent time is shown in Figure 5.19, where the curve marked by L represents the low stress (load) and the curve marked by H represents the high stress (load).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  can be calculated by:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{1}}(t)= &amp;amp; {{R}_{1}}({{t}_{1e}}) \\ &lt;br /&gt;
{{e}^{-{{(tKS_{1}^{n})}^{\beta }}}}= &amp;amp; {{e}^{-{{({{t}_{1e}}K{{S}^{n}})}^{\beta }}}} \\ &lt;br /&gt;
tS_{1}^{n}= &amp;amp; {{t}_{1e}}{{S}^{n}} \\ &lt;br /&gt;
{{t}_{1e}}= &amp;amp; t{{\left( \frac{{{S}_{1}}}{S} \right)}^{n}},\text{     }{{S}_{1}}={{P}_{1}}S \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{1e}}=tP_{1}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  can be calculated the same way, or:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{2}}(t)= &amp;amp; {{R}_{2}}({{t}_{2e}}) \\ &lt;br /&gt;
\therefore  &amp;amp; {{t}_{2e}}=tP_{2}^{n}  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In this example, the reliability equations for Unit 1 and Unit 2 are the same since they are the same type of component and demonstrate the same failure properties.  In addition, the total output is divided equally between the two units (when both units are operating), so  &amp;lt;math&amp;gt;{{t}_{1e}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{t}_{2e}}&amp;lt;/math&amp;gt;  will also be the same.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The next step is to determine the reliability of the system after 8,760 hours,  &amp;lt;math&amp;gt;R(t=8,760)&amp;lt;/math&amp;gt; .  Using Eqn. 22 the system reliability is found to be:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t=8760)= &amp;amp; 0.8567 \\ &lt;br /&gt;
= &amp;amp; 85.67%  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Load Sharing in BlockSim===&lt;br /&gt;
BlockSim uses this formulation when computing reliabilities of units in a load sharing configuration.  When using the System Reliability Equation window, BlockSim returns a single token for the reliability of units in a load sharing configuration (as well as in the case of standby redundancy, discussed in the next section).  As an example, consider the following RBD with Unit 1 in series with a container that includes two load sharing units.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.2.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt; &lt;br /&gt;
BlockSim will return the system equation as: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;{{R}_{System}}=+{{R}_{LS}}\cdot {{R}_{1}}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where  &amp;lt;math&amp;gt;{{R}_{LS}}&amp;lt;/math&amp;gt;  implies a form similar to Eqn. 22.  BlockSim allows for  &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; -out-of- &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  units in a load sharing configuration.&lt;br /&gt;
&lt;br /&gt;
===Example 5===&lt;br /&gt;
A component has five possible failure modes,  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; , and the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are interdependent.  The system will fail if mode  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  occurs, mode  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  occurs or two out of the three  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes occur.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  each have a Weibull distribution, with a  &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =10,000&amp;lt;/math&amp;gt;  and 15,000 respectively.  Events  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt;  each have an exponential distribution with a mean of 10,000 hours.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If any  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  event occurs (i.e.  &amp;lt;math&amp;gt;{{B}_{A}}&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{B}_{B}}&amp;lt;/math&amp;gt;  or  &amp;lt;math&amp;gt;{{B}_{C}}&amp;lt;/math&amp;gt; ), the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are more likely to occur.  Specifically, the mean times of the remaining  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  events are halved.  Determine the reliability at 1000 hours for this component.&lt;br /&gt;
====Solution to Example 5====&lt;br /&gt;
The first step is to create the RBD.  Modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  and a load sharing container with the  &amp;lt;math&amp;gt;{{B}_{i}}&amp;lt;/math&amp;gt;  modes must be drawn in series, as illustrated next.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BS5.19.3.png|thumb|center|200px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The next step is to define the properties for each block, including those for the container.  Setting the failure distributions for modes  &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;  is simple.  The more difficult part is setting the properties for the container and the contained blocks.  Based on the problem statement, the  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  modes are in a 2-out-of-3 load sharing redundancy.  When all three are working (i.e. when no  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode has occurred), each block has an exponential distribution with &amp;lt;math&amp;gt;\mu=10,000&amp;lt;/math&amp;gt;.  If one  &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;  mode occurs, then the two surviving units have an exponential distribution with  &amp;lt;math&amp;gt;\mu =5,000.&amp;lt;/math&amp;gt; &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Assume a Power Life-Stress relationship for the components.  Then:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{\mu }_{1}}= &amp;amp; \frac{1}{KV_{1}^{n}}  \ (eqn 23)\\ &lt;br /&gt;
{{\mu }_{2}}= &amp;amp; \frac{1}{KV_{2}^{n}}  \ (eqn 24)&lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{1}}=10,000&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;{{V}_{1}}=1&amp;lt;/math&amp;gt;  in Eqn. 23 and casting it in terms of  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  yields:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
10,000= &amp;amp; \frac{1}{K}  \ (eqn 25) \\ &lt;br /&gt;
K = &amp;amp; \frac{1}{10,000}=0.0001  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Substituting  &amp;lt;math&amp;gt;{{\mu }_{2}}=5,000&amp;lt;/math&amp;gt; ,  &amp;lt;math&amp;gt;{{V}_{2}}=1.5&amp;lt;/math&amp;gt;  (because if one fails, then each survivor takes on an additional 0.5 units of load) and Eqn. 25 for  &amp;lt;math&amp;gt;K&amp;lt;/math&amp;gt;  in Eqn.24 yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
5,000= &amp;amp; \frac{1}{0.0001\cdot {{(1.5)}^{n}}} \\ &lt;br /&gt;
0.5= &amp;amp; {{(1.5)}^{-n}} \\ &lt;br /&gt;
\ln (0.5)= &amp;amp; -n\ln (1.5) \\ &lt;br /&gt;
n = &amp;amp; 1.7095  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This also could have been computed in ALTA, as shown in Figure 5.20, or with the Load &amp;amp; Life Parameter Experimenter in BlockSim, as shown in Figure 5.21.&lt;br /&gt;
 &lt;br /&gt;
At this point, the parameters for the load sharing units have been computed and can be set, as shown in Figure 5.22.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The next step is to set the weight proportionality factor.  This factor defines the portion of the load that the particular item carries while operating, as well as the load that shifts to the remaining units upon failure of the item.  To illustrate, assume three units (1, 2 and 3) are in a load sharing container with weight proportionality factors of 1, 2 and 3 respectively (and a 1-out-of-3 requirement).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:•	Unit 1 carries  &amp;lt;math&amp;gt;\left( \tfrac{1}{1+2+3} \right)=0.166&amp;lt;/math&amp;gt;  or 16.6% of the total load.&lt;br /&gt;
:•	Unit 2 carries  &amp;lt;math&amp;gt;\left( \tfrac{2}{1+2+3} \right)=0.333&amp;lt;/math&amp;gt;  or 33.3% of the total load.&lt;br /&gt;
:•	Unit 3 carries  &amp;lt;math&amp;gt;\left( \tfrac{3}{1+2+3} \right)=0.50&amp;lt;/math&amp;gt;  or 50% of the total load.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The actual load on each unit then becomes the product of the entire load defined for the container times the portion carried by that unit.  For example, if the container load is 100 lbs, then the portion assigned to Unit 1 will be  &amp;lt;math&amp;gt;100\cdot 0.166=16.6&amp;lt;/math&amp;gt;  lbs.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the current example, all units share the same load and thus have equal weight proportionality factors.  Because these factors are relative, if the same number is used for all three items then the results will be the same.  Thus, weight proportional factor is set equal to 1 for each item.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5.20.PNG|thumb|center|500px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.20 Calculation performed in ALTA.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.21_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.21 Quick Parameter Estimator &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.21_3.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.22 Quick Parameter Estimator results &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Fig 5.23_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.23 Defining Weight Proportional Factor. &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The last properties that need to be defined are the total load and the 2-out-of-3 redundancy.  The total load is dependent on how the parameters were computed.  In this case, total load was assumed to be 3 when the parameters were computed (i.e. the load per item was 1 when all worked and 1.5 when two worked).  This is defined at the container level, set No. of Paths required = 3.&lt;br /&gt;
When all of the parameters have been specified in BlockSim, the reliability at 1,000 hours can be determined.  From the Analytical QCP, this is found to be 98.57%.&lt;br /&gt;
&lt;br /&gt;
=Standby Components=&lt;br /&gt;
&lt;br /&gt;
In the previous section, the case of a system with load sharing components was presented.  This is a form of redundancy with dependent components. That is, the failure of one component affects the failure of the other(s).  This section presents another form of redundancy: standby redundancy.  In standby redundancy the redundant components are set to be under a lighter load condition (or no load) while not needed and under the operating load when they are activated.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In standby redundancy the components are set to have two states: an active state and a standby state.  Components in standby redundancy have two failure distributions, one for each state.  When in the standby state, they have a quiescent (or dormant) failure distribution and when operating, they have an active failure distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case that both quiescent and active failure distributions are the same, the units are in a simple parallel configuration (also called a hot standby configuration).  When the rate of failure of the standby component is lower in quiescent mode than in active mode, that is called a warm standby configuration.  When the rate of failure of the standby component is zero in quiescent mode (i.e. the component cannot fail when in standby), that is called a cold standby configuration.  &lt;br /&gt;
&lt;br /&gt;
===Simple Standby Configuration===&lt;br /&gt;
&lt;br /&gt;
Consider two components in a standby configuration.  Component 1 is the active component with a Weibull failure distribution with parameters  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Component 2 is the standby component.  When Component 2 is operating, it also has a Weibull failure distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 1,000.  Furthermore, assume the following cases for the quiescent distribution.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 1:  The quiescent distribution is the same as the active distribution (hot standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 2:  The quiescent distribution is a Weibull  distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.5 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 2000 (warm standby).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Case 3: The component cannot fail in quiescent mode (cold standby).&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
In this case, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , can be obtained using the following equation:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;R(t)={{R}_{1}}(t)+\underset{0}{\overset{t}{\mathop \int }}\,{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x)\cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}dx   \ (eqn 26)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode, such that: &amp;lt;br&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
&amp;lt;br&amp;gt;  &lt;br /&gt;
&amp;lt;math&amp;gt;{{R}_{2;SB}}(x)={{R}_{2;A}}({{t}_{e}})  \ (eqn 27)&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Eqn. 27 can be solved for  &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  and substituted into Eqn.26.&lt;br /&gt;
Figure 5.24 illustrates the example as entered in BlockSim using a standby container.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5.24.gif|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.24 Standby container.&amp;lt;/div&amp;gt;]]&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The active and standby blocks are within a container, which is used to specify standby redundancy.  Since the standby component has two distributions (active and quiescent), the Block Properties window of the standby block has two pages for specifying each one.  Figures 5.24 and 5.26 illustrate these pages.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
The system reliability results for 1000 hours are given in the following table:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-24.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
Note that even though the  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  value for the quiescent distribution is the same as in the active distribution, it is possible that the two can be different. That is, the failure modes present during the quiescent mode could be different from the modes present during the active mode.  In that sense, the two distribution types can be different as well (e.g. lognormal when quiescent and Weibull when active).&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In many cases when considering standby systems, a switching device may also be present that switches from the failed active component to the standby component.  The reliability of the switch can also be incorporated into Eqn.26, as presented in the next section.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
BlockSim&#039;s System Reliability Equation window returns a single token for the reliability of units in a standby configuration.  This is the same as the load sharing case presented in the previous section. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image: Fig 5.25.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.25 Defining the active failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.26.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.26 Defining the quiescent failure distribution &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Reliability of Standby Systems with a Switching Device===&lt;br /&gt;
&lt;br /&gt;
In many cases when dealing with standby systems, a switching device is present that will switch to the standby component when the active component fails.  Therefore, the failure properties of the switch must also be included in the analysis.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:BS5.26.2.png|thumb|center|300px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In most cases when the reliability of a switch is to be included in the analysis, two probabilities can be considered.  The first and most common one is the probability of the switch performing the action (i.e. switching) when requested to do so.  This is called Switch Probability per Request in BlockSim and is expressed as a static probability (e.g. 90%).  The second probability is the quiescent reliability of the switch.  This is the reliability of the switch as it ages (e.g. the switch might wear out with age due to corrosion, material degradation, etc.). Thus it is possible for the switch to fail before the active component fails.  However, a switch failure does not cause the system to fail, but rather causes the system to fail only if the switch is needed and the switch has failed.  For example, if the active component does not fail until the mission end time and the switch fails, then the system does not fail.  However, if the active component fails and the switch has also failed, then the system cannot be switched to the standby component and it therefore fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In analyzing standby components with a switching device, either or both failure probabilities (during the switching or while waiting to switch) can be considered for the switch, since each probability can represent different failure modes.  For example, the switch probability per request may represent software-related issues or the probability of detecting the failure of an active component, and the quiescent probability may represent wear-out type failures of the switch.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To illustrate the formulation, consider the previous example that assumes perfect switching.  To examine the effects of including an imperfect switch, assume that when the active component fails there is a 90% probability that the switch will switch from the active component to the standby component.  In addition, assume that the switch can also fail due to a wear-out failure mode described by a Weibull distribution with  &amp;lt;math&amp;gt;\beta &amp;lt;/math&amp;gt;  = 1.7 and  &amp;lt;math&amp;gt;\eta &amp;lt;/math&amp;gt;  = 5000.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, the reliability of the system at some time,  &amp;lt;math&amp;gt;t&amp;lt;/math&amp;gt; , is given by the following equation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
R(t)= &amp;amp; {{R}_{1}}(t) \\ &lt;br /&gt;
&amp;amp; +\underset{0}{\overset{t}{\mathop \int }}\,\{{{f}_{1}}(x)\cdot {{R}_{2;SB}}(x) \\ &lt;br /&gt;
&amp;amp; \cdot \frac{{{R}_{2;A}}({{t}_{e}}+t-x)}{{{R}_{2;A}}({{t}_{e}})}\cdot {{R}_{SW;Q}}(x)\cdot {{R}_{SW;REQ}}(x)\}dx  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{1}}&amp;lt;/math&amp;gt;  is the reliability of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{f}_{1}}&amp;lt;/math&amp;gt;  is the  &amp;lt;math&amp;gt;pdf&amp;lt;/math&amp;gt;  of the active component.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;SB}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in standby mode (quiescent reliability).&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{2;A}}&amp;lt;/math&amp;gt;  is the reliability of the standby component when in active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;Q}}&amp;lt;/math&amp;gt;  is the quiescent reliability of the switch.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{R}_{SW;REQ}}&amp;lt;/math&amp;gt;  is the switch probability per request.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	 &amp;lt;math&amp;gt;{{t}_{e}}&amp;lt;/math&amp;gt;  is the equivalent operating time for the standby unit if it had been operating at an active mode.&amp;lt;br&amp;gt;&lt;br /&gt;
This problem can be solved in BlockSim by including these probabilities in the container&#039;s properties, as shown in Figures fig23 and fig24.  In BlockSim, the standby container is acting as the switch.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.27.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.27 Standby container (switch) failure distribution while waiting to switch &amp;lt;/div&amp;gt;]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:Fig 5.28_2.PNG|thumb|center|400px|&amp;lt;div align=&amp;quot;center&amp;quot;&amp;gt; Fig 5.28 Standby container (switch) failure probabilities while attempting to switch &amp;lt;/div&amp;gt; ]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that there are additional properties that can be specified in BlockSim for a switch, such as Switch Restart Probability, Finite Restarts and Switch Delay Time.  In many applications, the switch is re-tested (or re-cycled) if it fails to switch the first time.  In these cases, it might be possible that it switches in the second or third, or  &amp;lt;math&amp;gt;{{n}^{th}}&amp;lt;/math&amp;gt; attempt.  &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The Switch Restart Probability specifies each additional attempt&#039;s probability of successfully switching and the Finite Restarts specifies the total number of attempts.  Note that the Switch Restart Probability specifies the probability of success of each trial (or attempt).  The probability of success of  &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;  consecutive trials is calculated by BlockSim using the binomial distribution and this probability is then incorporated into Eqn. (stb2a).  The Switch Delay Time property is related to repairable systems and is considered in BlockSim only when using simulation.  When using the analytical solution (i.e. for a non-repairable system), this property is ignored.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving the analytical solution (as given by Eqn. stb2a), the following results are obtained.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Image:5-30.png|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
From the table above, it can be seen that the presence of a switching device has a significant effect on the reliability of a standby system.  It is therefore important when modeling standby redundancy to incorporate the switching device reliability properties.  It should be noted that this methodology is not the same as treating the switching device as another series component with the standby subsystem.  This would be valid only if the failure of the switch resulted in the failure of system (e.g. switch failing open).  In Eqn. (stb2a), the Switch Probability per Request and quiescent probability are present only in the second term of the equation.  Treating these two failure modes as a series configuration with the standby subsystem would imply that they are also present when the active component is functioning (i.e. first term of Eqn. stb2a).  This is invalid and would result in the underestimation of the reliability of the system.  In other words, these two failure modes become significant only when the active component fails.&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
As an example, and if we consider the warm standby case, the reliability of the system without the switch is 70.57% at 1000 hours.  If the system was modeled so that the switching device was in series with the warm standby subsystem, the result would have been:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
::&amp;lt;math&amp;gt;\begin{align}&lt;br /&gt;
{{R}_{S}}(1000)= &amp;amp; {{R}_{Standby}}(1000)\cdot {{R}_{sw,Q(1000)}}\cdot {{R}_{sw,req}} \\ &lt;br /&gt;
= &amp;amp; 0.7057\cdot 0.9372\cdot 0.9 \\ &lt;br /&gt;
= &amp;amp; 0.5952  &lt;br /&gt;
\end{align}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the case where a switch failure mode causes the standby subsystem to fail, then this mode can be modeled as an individual block in series with the standby subsystem.&lt;br /&gt;
&lt;br /&gt;
===Example 6===&lt;br /&gt;
Consider a car with four new tires and a full-size spare.  Assume the following failure characteristics:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires follow a Weibull distribution with a  ..  and an  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  40,000 miles while on the car due to wear.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	The tires also have a probability of failing due to puncture or other causes.  For this, assume a constant rate for this occurrence with a probability of 1 every 50,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	When not on the car (i.e. is a spare), a tire&#039;s probability of failing also has a Weibull distribution with a  &amp;lt;math&amp;gt;\beta =&amp;lt;/math&amp;gt;  2 and  &amp;lt;math&amp;gt;\eta =&amp;lt;/math&amp;gt;  120,000 miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
Assume a mission of 1,000 miles.  If a tire fails during this trip, it will be replaced with the spare.  However, the spare will not be repaired during the trip.  In other words, the trip will continue with the spare on the car and if the spare fails the system will fail.  Determine the probability of system failure.&lt;br /&gt;
====Solution to Example 6====&lt;br /&gt;
Active failure distribution for tires:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to wear-out, Weibull  &amp;lt;math&amp;gt;\beta =4&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =40,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
:•	Due to random puncture, exponential  &amp;lt;math&amp;gt;\mu =50,000.&amp;lt;/math&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
:•	The quiescent failure distribution is a Weibull distribution with &amp;lt;math&amp;gt;\beta =2&amp;lt;/math&amp;gt;  and  &amp;lt;math&amp;gt;\eta =120,000&amp;lt;/math&amp;gt;  miles.&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
The block diagram for each tire has two blocks in series, one block representing the wear-out mode and the other the random puncture mode, as shown next:&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:small5.gif|thumb|center|400px|]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
There are five tires, four active and one standby (represented in the diagram by a standby container with a 4-out-of-5 requirement), as shown next: &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:BStirecontainer.png|thumb|center|400px|]]&lt;br /&gt;
 &lt;br /&gt;
For the standby Wear block, set the active failure and the quiescent distributions, but for the Puncture block, set only the active puncture distribution (because the tire cannot fail due to puncture while stored).  Using BlockSim, the probability of system failure is found to be 0.003 or 0.3%. &lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Note Regarding Numerical Integration Solutions=&lt;br /&gt;
&lt;br /&gt;
Load sharing and standby solutions in BlockSim are performed using numerical integration routines.  As with any numerical analysis routine, the solution error depends on the number of iterations performed, the step size chosen and related factors, plus the behavior of the underlying function.  By default, BlockSim uses a certain set of preset factors.  In general, these defaults are sufficient for most problems.  If a higher precision or verification of the precision for a specific problem is required, BlockSim&#039;s preset options can be modified and/or the integration error can be assessed using the Integration Parameters... option for each container.  For more details, you can refer to the documentation on the Algorithm Setup window in the BlockSim 7 User&#039;s Guide.&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
	<entry>
		<id>https://www.reliawiki.com/index.php?title=File:Fig_5.27.PNG&amp;diff=15558</id>
		<title>File:Fig 5.27.PNG</title>
		<link rel="alternate" type="text/html" href="https://www.reliawiki.com/index.php?title=File:Fig_5.27.PNG&amp;diff=15558"/>
		<updated>2012-02-13T22:08:00Z</updated>

		<summary type="html">&lt;p&gt;Pengying niu: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Pengying niu</name></author>
	</entry>
</feed>