# Multiple Linear Regression Analysis

This chapter expands on the analysis of simple linear regression models and discusses the analysis of multiple linear regression models. A major portion of the results displayed in Weibull++ DOE folios are explained in this chapter because these results are associated with multiple linear regression. One of the applications of multiple linear regression models is Response Surface Methodology (RSM). RSM is a method used to locate the optimum value of the response and is one of the final stages of experimentation. It is discussed in Response Surface Methods. Towards the end of this chapter, the concept of using indicator variables in regression models is explained. Indicator variables are used to represent qualitative factors in regression models. The concept of using indicator variables is important to gain an understanding of ANOVA models, which are the models used to analyze data obtained from experiments. These models can be thought of as first order multiple linear regression models where all the factors are treated as qualitative factors. ANOVA models are discussed in the One Factor Designs and General Full Factorial Designs chapters.

## Multiple Linear Regression Model

A linear regression model that contains more than one predictor variable is called a *multiple linear regression model*. The following model is a multiple linear regression model with two predictor variables, [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math].

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math]

The model is linear because it is linear in the parameters [math]{{\beta }_{0}}\,\![/math], [math]{{\beta }_{1}}\,\![/math] and [math]{{\beta }_{2}}\,\![/math]. The model describes a plane in the three-dimensional space of [math]Y\,\![/math], [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math]. The parameter [math]{{\beta }_{0}}\,\![/math] is the intercept of this plane. Parameters [math]{{\beta }_{1}}\,\![/math] and [math]{{\beta }_{2}}\,\![/math] are referred to as *partial regression coefficients*. Parameter [math]{{\beta }_{1}}\,\![/math] represents the change in the mean response corresponding to a unit change in [math]{{x}_{1}}\,\![/math] when [math]{{x}_{2}}\,\![/math] is held constant. Parameter [math]{{\beta }_{2}}\,\![/math] represents the change in the mean response corresponding to a unit change in [math]{{x}_{2}}\,\![/math] when [math]{{x}_{1}}\,\![/math] is held constant.
Consider the following example of a multiple linear regression model with two predictor variables, [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math] :

- [math]Y=30+5{{x}_{1}}+7{{x}_{2}}+\epsilon \,\![/math]

This regression model is a first order multiple linear regression model. This is because the maximum power of the variables in the model is 1. (The regression plane corresponding to this model is shown in the figure below.) Also shown is an observed data point and the corresponding random error, [math]\epsilon\,\![/math]. The true regression model is usually never known (and therefore the values of the random error terms corresponding to observed data points remain unknown). However, the regression model can be estimated by calculating the parameters of the model for an observed data set. This is explained in Estimating Regression Models Using Least Squares.

One of the following figures shows the contour plot for the regression model the above equation. The contour plot shows lines of constant mean response values as a function of [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math]. The contour lines for the given regression model are straight lines as seen on the plot. Straight contour lines result for first order regression models with no interaction terms.

A linear regression model may also take the following form:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon\,\![/math]

A cross-product term, [math]{{x}_{1}}{{x}_{2}}\,\![/math], is included in the model. This term represents an interaction effect between the two variables [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math]. Interaction means that the effect produced by a change in the predictor variable on the response depends on the level of the other predictor variable(s). As an example of a linear regression model with interaction, consider the model given by the equation [math]Y=30+5{{x}_{1}}+7{{x}_{2}}+3{{x}_{1}}{{x}_{2}}+\epsilon\,\![/math]. The regression plane and contour plot for this model are shown in the following two figures, respectively.

Now consider the regression model shown next:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}x_{1}^{2}+{{\beta }_{3}}x_{1}^{3}+\epsilon\,\![/math]

This model is also a linear regression model and is referred to as a *polynomial regression model*. Polynomial regression models contain squared and higher order terms of the predictor variables making the response surface curvilinear. As an example of a polynomial regression model with an interaction term consider the following equation:

- [math]Y=500+5{{x}_{1}}+7{{x}_{2}}-3x_{1}^{2}-5x_{2}^{2}+3{{x}_{1}}{{x}_{2}}+\epsilon\,\![/math]

This model is a *second order* model because the maximum power of the terms in the model is two. The regression surface for this model is shown in the following figure. Such regression models are used in RSM to find the optimum value of the response, [math]Y\,\![/math] (for details see Response Surface Methods for Optimization). Notice that, although the shape of the regression surface is curvilinear, the regression model is still linear because the model is linear in the parameters. The contour plot for this model is shown in the second of the following two figures.

All multiple linear regression models can be expressed in the following general form:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon\,\![/math]

where [math]k\,\![/math] denotes the number of terms in the model. For example, the model can be written in the general form using [math]{{x}_{3}}=x_{1}^{2}\,\![/math], [math]{{x}_{4}}=x_{2}^{3}\,\![/math] and [math]{{x}_{5}}={{x}_{1}}{{x}_{2}}\,\![/math] as follows:

- [math]Y=500+5{{x}_{1}}+7{{x}_{2}}-3{{x}_{3}}-5{{x}_{4}}+3{{x}_{5}}+\epsilon\,\![/math]

## Estimating Regression Models Using Least Squares

Consider a multiple linear regression model with [math]k\,\![/math] predictor variables:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+...+{{\beta }_{k}}{{x}_{k}}+\epsilon\,\![/math]

Let each of the [math]k\,\![/math] predictor variables, [math]{{x}_{1}}\,\![/math], [math]{{x}_{2}}\,\![/math]... [math]{{x}_{k}}\,\![/math], have [math]n\,\![/math] levels. Then [math]{{x}_{ij}}\,\![/math] represents the [math]i\,\![/math] th level of the [math]j\,\![/math] th predictor variable [math]{{x}_{j}}\,\![/math]. For example, [math]{{x}_{51}}\,\![/math] represents the fifth level of the first predictor variable [math]{{x}_{1}}\,\![/math], while [math]{{x}_{19}}\,\![/math] represents the first level of the ninth predictor variable, [math]{{x}_{9}}\,\![/math]. Observations, [math]{{y}_{1}}\,\![/math], [math]{{y}_{2}}\,\![/math]... [math]{{y}_{n}}\,\![/math], recorded for each of these [math]n\,\![/math] levels can be expressed in the following way:

- [math]\begin{align} {{y}_{1}}= & {{\beta }_{0}}+{{\beta }_{1}}{{x}_{11}}+{{\beta }_{2}}{{x}_{12}}+...+{{\beta }_{k}}{{x}_{1k}}+{{\epsilon }_{1}} \\ {{y}_{2}}= & {{\beta }_{0}}+{{\beta }_{1}}{{x}_{21}}+{{\beta }_{2}}{{x}_{22}}+...+{{\beta }_{k}}{{x}_{2k}}+{{\epsilon }_{2}} \\ & .. \\ {{y}_{i}}= & {{\beta }_{0}}+{{\beta }_{1}}{{x}_{i1}}+{{\beta }_{2}}{{x}_{i2}}+...+{{\beta }_{k}}{{x}_{ik}}+{{\epsilon }_{i}} \\ & .. \\ {{y}_{n}}= & {{\beta }_{0}}+{{\beta }_{1}}{{x}_{n1}}+{{\beta }_{2}}{{x}_{n2}}+...+{{\beta }_{k}}{{x}_{nk}}+{{\epsilon }_{n}} \end{align}\,\![/math]

The system of [math]n\,\![/math] equations shown previously can be represented in matrix notation as follows:

- [math]y=X\beta +\epsilon\,\![/math]

- where

- [math]y=\left[ \begin{matrix} {{y}_{1}} \\ {{y}_{2}} \\ . \\ . \\ . \\ {{y}_{n}} \\ \end{matrix} \right]\text{ }X=\left[ \begin{matrix} 1 & {{x}_{11}} & {{x}_{12}} & . & . & . & {{x}_{1n}} \\ 1 & {{x}_{21}} & {{x}_{22}} & . & . & . & {{x}_{2n}} \\ . & . & . & {} & {} & {} & . \\ . & . & . & {} & {} & {} & . \\ . & . & . & {} & {} & {} & . \\ 1 & {{x}_{n1}} & {{x}_{n2}} & . & . & . & {{x}_{nn}} \\ \end{matrix} \right]\,\![/math]

- [math]\beta =\left[ \begin{matrix} {{\beta }_{0}} \\ {{\beta }_{1}} \\ . \\ . \\ . \\ {{\beta }_{n}} \\ \end{matrix} \right]\text{ and }\epsilon =\left[ \begin{matrix} {{\epsilon }_{1}} \\ {{\epsilon }_{2}} \\ . \\ . \\ . \\ {{\epsilon }_{n}} \\ \end{matrix} \right]\,\![/math]

The matrix [math]X\,\![/math] is referred to as the *design matrix*. It contains information about the levels of the predictor variables at which the observations are obtained. The vector [math]\beta\,\![/math] contains all the regression coefficients. To obtain the regression model, [math]\beta\,\![/math] should be known. [math]\beta\,\![/math] is estimated using least square estimates. The following equation is used:

- [math]\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y\,\![/math]

where [math]^{\prime }\,\![/math] represents the transpose of the matrix while [math]^{-1}\,\![/math] represents the matrix inverse. Knowing the estimates, [math]\hat{\beta }\,\![/math], the multiple linear regression model can now be estimated as:

- [math]\hat{y}=X\hat{\beta }\,\![/math]

The estimated regression model is also referred to as the *fitted model*. The observations, [math]{{y}_{i}}\,\![/math], may be different from the fitted values [math]{{\hat{y}}_{i}}\,\![/math] obtained from this model. The difference between these two values is the residual, [math]{{e}_{i}}\,\![/math]. The vector of residuals, [math]e\,\![/math], is obtained as:

- [math]e=y-\hat{y}\,\![/math]

The fitted model can also be written as follows, using [math]\hat{\beta }={{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y\,\![/math]:

- [math]\begin{align} \hat{y} &= & X\hat{\beta } \\ & = & X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ & = & Hy \end{align}\,\![/math]

where [math]H=X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}\,\![/math]. The matrix, [math]H\,\![/math], is referred to as the hat matrix. It transforms the vector of the observed response values, [math]y\,\![/math], to the vector of fitted values, [math]\hat{y}\,\![/math].

### Example

An analyst studying a chemical process expects the yield to be affected by the levels of two factors, [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math]. Observations recorded for various levels of the two factors are shown in the following table. The analyst wants to fit a first order regression model to the data. Interaction between [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math] is not expected based on knowledge of similar processes. Units of the factor levels and the yield are ignored for the analysis.

The data of the above table can be entered into the Weibull++ DOE folio using the multiple linear regression folio tool as shown in the following figure.

A scatter plot for the data is shown next.

The first order regression model applicable to this data set having two predictor variables is:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math]

where the dependent variable, [math]Y\,\![/math], represents the yield and the predictor variables, [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math], represent the two factors respectively. The [math]X\,\![/math] and [math]y\,\![/math] matrices for the data can be obtained as:

- [math]X=\left[ \begin{matrix} 1 & 41.9 & 29.1 \\ 1 & 43.4 & 29.3 \\ . & . & . \\ . & . & . \\ . & . & . \\ 1 & 77.8 & 32.9 \\ \end{matrix} \right]\text{ }y=\left[ \begin{matrix} 251.3 \\ 251.3 \\ . \\ . \\ . \\ 349.0 \\ \end{matrix} \right]\,\![/math]

The least square estimates, [math]\hat{\beta }\,\![/math], can now be obtained:

- [math]\begin{align} \hat{\beta } &= & {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ & = & {{\left[ \begin{matrix} 17 & 941 & 525.3 \\ 941 & 54270 & 29286 \\ 525.3 & 29286 & 16254 \\ \end{matrix} \right]}^{-1}}\left[ \begin{matrix} 4902.8 \\ 276610 \\ 152020 \\ \end{matrix} \right] \\ & = & \left[ \begin{matrix} -153.51 \\ 1.24 \\ 12.08 \\ \end{matrix} \right] \end{align}\,\![/math]

- Thus:

- [math]\hat{\beta }=\left[ \begin{matrix} {{{\hat{\beta }}}_{0}} \\ {{{\hat{\beta }}}_{1}} \\ {{{\hat{\beta }}}_{2}} \\ \end{matrix} \right]=\left[ \begin{matrix} -153.51 \\ 1.24 \\ 12.08 \\ \end{matrix} \right]\,\![/math]

and the estimated regression coefficients are [math]{{\hat{\beta }}_{0}}=-153.51\,\![/math], [math]{{\hat{\beta }}_{1}}=1.24\,\![/math] and [math]{{\hat{\beta }}_{2}}=12.08\,\![/math]. The fitted regression model is:

- [math]\begin{align} \hat{y} & = & {{{\hat{\beta }}}_{0}}+{{{\hat{\beta }}}_{1}}{{x}_{1}}+{{{\hat{\beta }}}_{2}}{{x}_{2}} \\ & = & -153.5+1.24{{x}_{1}}+12.08{{x}_{2}} \end{align}\,\![/math]

The fitted regression model can be viewed in the Weibull++ DOE folio, as shown next.

A plot of the fitted regression plane is shown in the following figure.

The fitted regression model can be used to obtain fitted values, [math]{{\hat{y}}_{i}}\,\![/math], corresponding to an observed response value, [math]{{y}_{i}}\,\![/math]. For example, the fitted value corresponding to the fifth observation is:

- [math]\begin{align} {{{\hat{y}}}_{i}} &= & -153.5+1.24{{x}_{i1}}+12.08{{x}_{i2}} \\ {{{\hat{y}}}_{5}} & = & -153.5+1.24{{x}_{51}}+12.08{{x}_{52}} \\ & = & -153.5+1.24(47.3)+12.08(29.9) \\ & = & 266.3 \end{align}\,\![/math]

The observed fifth response value is [math]{{y}_{5}}=273.0\,\![/math]. The residual corresponding to this value is:

- [math]\begin{align} {{e}_{i}} & = & {{y}_{i}}-{{{\hat{y}}}_{i}} \\ {{e}_{5}}& = & {{y}_{5}}-{{{\hat{y}}}_{5}} \\ & = & 273.0-266.3 \\ & = & 6.7 \end{align}\,\![/math]

In Weibull++ DOE folios, fitted values and residuals are shown in the Diagnostic Information table of the detailed summary of results. The values are shown in the following figure.

The fitted regression model can also be used to predict response values. For example, to obtain the response value for a new observation corresponding to 47 units of [math]{{x}_{1}}\,\![/math] and 31 units of [math]{{x}_{2}}\,\![/math], the value is calculated using:

- [math]\begin{align} \hat{y}(47,31)& = & -153.5+1.24(47)+12.08(31) \\ & = & 279.26 \end{align}\,\![/math]

### Properties of the Least Square Estimators for Beta

The least square estimates, [math]{{\hat{\beta }}_{0}}\,\![/math], [math]{{\hat{\beta }}_{1}}\,\![/math], [math]{{\hat{\beta }}_{2}}\,\![/math]... [math]{{\hat{\beta }}_{k}}\,\![/math], are unbiased estimators of [math]{{\beta }_{0}}\,\![/math], [math]{{\beta }_{1}}\,\![/math], [math]{{\beta }_{2}}\,\![/math]... [math]{{\beta }_{k}}\,\![/math], provided that the random error terms, [math]{{\epsilon }_{i}}\,\![/math], are normally and independently distributed. The variances of the [math]\hat{\beta }\,\![/math] s are obtained using the [math]{{({{X}^{\prime }}X)}^{-1}}\,\![/math] matrix. The variance-covariance matrix of the estimated regression coefficients is obtained as follows:

- [math]C={{\hat{\sigma }}^{2}}{{({{X}^{\prime }}X)}^{-1}}\,\![/math]

[math]C\,\![/math] is a symmetric matrix whose diagonal elements, [math]{{C}_{jj}}\,\![/math], represent the variance of the estimated [math]j\,\![/math] th regression coefficient, [math]{{\hat{\beta }}_{j}}\,\![/math]. The off-diagonal elements, [math]{{C}_{ij}}\,\![/math], represent the covariance between the [math]i\,\![/math] th and [math]j\,\![/math] th estimated regression coefficients, [math]{{\hat{\beta }}_{i}}\,\![/math] and [math]{{\hat{\beta }}_{j}}\,\![/math]. The value of [math]{{\hat{\sigma }}^{2}}\,\![/math] is obtained using the error mean square, [math]M{{S}_{E}}\,\![/math]. The variance-covariance matrix for the data in the table (see Estimating Regression Models Using Least Squares) can be viewed in the DOE folio, as shown next.

Calculations to obtain the matrix are given in this example. The positive square root of [math]{{C}_{jj}}\,\![/math] represents the estimated standard deviation of the [math]j\,\![/math] th regression coefficient, [math]{{\hat{\beta }}_{j}}\,\![/math], and is called the estimated standard error of [math]{{\hat{\beta }}_{j}}\,\![/math] (abbreviated [math]se({{\hat{\beta }}_{j}})\,\![/math] ).

- [math]se({{\hat{\beta }}_{j}})=\sqrt{{{C}_{jj}}}\,\![/math]

## Hypothesis Tests in Multiple Linear Regression

This section discusses hypothesis tests on the regression coefficients in multiple linear regression. As in the case of simple linear regression, these tests can only be carried out if it can be assumed that the random error terms, [math]{{\epsilon }_{i}}\,\![/math], are normally and independently distributed with a mean of zero and variance of [math]{{\sigma }^{2}}\,\![/math]. Three types of hypothesis tests can be carried out for multiple linear regression models:

- Test for significance of regression: This test checks the significance of the whole regression model.
- [math]t\,\![/math] test: This test checks the significance of individual regression coefficients.
- [math]F\,\![/math] test: This test can be used to simultaneously check the significance of a number of regression coefficients. It can also be used to test individual coefficients.

### Test for Significance of Regression

The test for significance of regression in the case of multiple linear regression analysis is carried out using the analysis of variance. The test is used to check if a linear statistical relationship exists between the response variable and at least one of the predictor variables. The statements for the hypotheses are:

- [math]\begin{align} & {{H}_{0}}:& {{\beta }_{1}}={{\beta }_{2}}=...={{\beta }_{k}}=0 \\ & {{H}_{1}}:& {{\beta }_{j}}\ne 0\text{ for at least one }j \end{align}\,\![/math]

The test for [math]{{H}_{0}}\,\![/math] is carried out using the following statistic:

- [math]{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}\,\![/math]

where [math]M{{S}_{R}}\,\![/math] is the regression mean square and [math]M{{S}_{E}}\,\![/math] is the error mean square. If the null hypothesis, [math]{{H}_{0}}\,\![/math], is true then the statistic [math]{{F}_{0}}\,\![/math] follows the [math]F\,\![/math] distribution with [math]k\,\![/math] degrees of freedom in the numerator and [math]n-\,\![/math] ( [math]k+1\,\![/math] ) degrees of freedom in the denominator. The null hypothesis, [math]{{H}_{0}}\,\![/math], is rejected if the calculated statistic, [math]{{F}_{0}}\,\![/math], is such that:

- [math]{{F}_{0}}\gt {{f}_{\alpha ,k,n-(k+1)}}\,\![/math]

#### Calculation of the Statistic [math]{{F}_{0}}\,\![/math]

To calculate the statistic [math]{{F}_{0}}\,\![/math], the mean squares [math]M{{S}_{R}}\,\![/math] and [math]M{{S}_{E}}\,\![/math] must be known. As explained in Simple Linear Regression Analysis, the mean squares are obtained by dividing the sum of squares by their degrees of freedom. For example, the total mean square, [math]M{{S}_{T}}\,\![/math], is obtained as follows:

- [math]M{{S}_{T}}=\frac{S{{S}_{T}}}{dof(S{{S}_{T}})}\,\![/math]

where [math]S{{S}_{T}}\,\![/math] is the total sum of squares and [math]dof(S{{S}_{T}})\,\![/math] is the number of degrees of freedom associated with [math]S{{S}_{T}}\,\![/math]. In multiple linear regression, the following equation is used to calculate [math]S{{S}_{T}}\,\![/math] :

- [math]S{{S}_{T}}={{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y\,\![/math]

where [math]n\,\![/math] is the total number of observations, [math]y\,\![/math] is the vector of observations (that was defined in Estimating Regression Models Using Least Squares), [math]I\,\![/math] is the identity matrix of order [math]n\,\![/math] and [math]J\,\![/math] represents an [math]n\times n\,\![/math] square matrix of ones. The number of degrees of freedom associated with [math]S{{S}_{T}}\,\![/math], [math]dof(S{{S}_{T}})\,\![/math], is ( [math]n-1\,\![/math] ). Knowing [math]S{{S}_{T}}\,\![/math] and [math]dof(S{{S}_{T}})\,\![/math] the total mean square, [math]M{{S}_{T}}\,\![/math], can be calculated.

The regression mean square, [math]M{{S}_{R}}\,\![/math], is obtained by dividing the regression sum of squares, [math]S{{S}_{R}}\,\![/math], by the respective degrees of freedom, [math]dof(S{{S}_{R}})\,\![/math], as follows:

- [math]M{{S}_{R}}=\frac{S{{S}_{R}}}{dof(S{{S}_{R}})}\,\![/math]

The regression sum of squares, [math]S{{S}_{R}}\,\![/math], is calculated using the following equation:

- [math]S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y\,\![/math]

where [math]n\,\![/math] is the total number of observations, [math]y\,\![/math] is the vector of observations, [math]H\,\![/math] is the hat matrix and [math]J\,\![/math] represents an [math]n\times n\,\![/math] square matrix of ones. The number of degrees of freedom associated with [math]S{{S}_{R}}\,\![/math], [math]dof(S{{S}_{E}})\,\![/math], is [math]k\,\![/math], where [math]k\,\![/math] is the number of predictor variables in the model. Knowing [math]S{{S}_{R}}\,\![/math] and [math]dof(S{{S}_{R}})\,\![/math] the regression mean square, [math]M{{S}_{R}}\,\![/math], can be calculated.
The error mean square, [math]M{{S}_{E}}\,\![/math], is obtained by dividing the error sum of squares, [math]S{{S}_{E}}\,\![/math], by the respective degrees of freedom, [math]dof(S{{S}_{E}})\,\![/math], as follows:

- [math]M{{S}_{E}}=\frac{S{{S}_{E}}}{dof(S{{S}_{E}})}\,\![/math]

The error sum of squares, [math]S{{S}_{E}}\,\![/math], is calculated using the following equation:

- [math]S{{S}_{E}}={{y}^{\prime }}(I-H)y\,\![/math]

where [math]y\,\![/math] is the vector of observations, [math]I\,\![/math] is the identity matrix of order [math]n\,\![/math] and [math]H\,\![/math] is the hat matrix. The number of degrees of freedom associated with [math]S{{S}_{E}}\,\![/math], [math]dof(S{{S}_{E}})\,\![/math], is [math]n-(k+1)\,\![/math], where [math]n\,\![/math] is the total number of observations and [math]k\,\![/math] is the number of predictor variables in the model. Knowing [math]S{{S}_{E}}\,\![/math] and [math]dof(S{{S}_{E}})\,\![/math], the error mean square, [math]M{{S}_{E}}\,\![/math], can be calculated. The error mean square is an estimate of the variance, [math]{{\sigma }^{2}}\,\![/math], of the random error terms, [math]{{\epsilon }_{i}}\,\![/math].

- [math]{{\hat{\sigma }}^{2}}=M{{S}_{E}}\,\![/math]

##### Example

The test for the significance of regression, for the regression model obtained for the data in the table (see Estimating Regression Models Using Least Squares), is illustrated in this example. The null hypothesis for the model is:

- [math]{{H}_{0}}: {{\beta }_{1}}={{\beta }_{2}}=0\,\![/math]

The statistic to test [math]{{H}_{0}}\,\![/math] is:

- [math]{{F}_{0}}=\frac{M{{S}_{R}}}{M{{S}_{E}}}\,\![/math]

To calculate [math]{{F}_{0}}\,\![/math], first the sum of squares are calculated so that the mean squares can be obtained. Then the mean squares are used to calculate the statistic [math]{{F}_{0}}\,\![/math] to carry out the significance test.
The regression sum of squares, [math]S{{S}_{R}}\,\![/math], can be obtained as:

- [math]S{{S}_{R}}={{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y\,\![/math]

The hat matrix, [math]H\,\![/math] is calculated as follows using the design matrix [math]X\,\![/math] from the previous example:

- [math]\begin{align} H & = & X{{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }} \\ & = & \left[ \begin{matrix} 0.27552 & 0.25154 & . & . & -0.04030 \\ 0.25154 & 0.23021 & . & . & -0.029120 \\ . & . & . & . & . \\ . & . & . & . & . \\ -0.04030 & -0.02920 & . & . & 0.30115 \\ \end{matrix} \right] \end{align}\,\![/math]

Knowing [math]y\,\![/math], [math]H\,\![/math] and [math]J\,\![/math], the regression sum of squares, [math]S{{S}_{R}}\,\![/math], can be calculated:

- [math]\begin{align} S{{S}_{R}} & = & {{y}^{\prime }}\left[ H-(\frac{1}{n})J \right]y \\ & = & 12816.35 \end{align}\,\![/math]

The degrees of freedom associated with [math]S{{S}_{R}}\,\![/math] is [math]k\,\![/math], which equals to a value of two since there are two predictor variables in the data in the table (see Multiple Linear Regression Analysis). Therefore, the regression mean square is:

- [math]\begin{align} M{{S}_{R}}& = & \frac{S{{S}_{R}}}{dof(S{{S}_{R}})} \\ & = & \frac{12816.35}{2} \\ & = & 6408.17 \end{align}\,\![/math]

Similarly to calculate the error mean square, [math]M{{S}_{E}}\,\![/math], the error sum of squares, [math]S{{S}_{E}}\,\![/math], can be obtained as:

- [math]\begin{align} S{{S}_{E}} &= & {{y}^{\prime }}\left[ I-H \right]y \\ & = & 423.37 \end{align}\,\![/math]

The degrees of freedom associated with [math]S{{S}_{E}}\,\![/math] is [math]n-(k+1)\,\![/math]. Therefore, the error mean square, [math]M{{S}_{E}}\,\![/math], is:

- [math]\begin{align} M{{S}_{E}} &= & \frac{S{{S}_{E}}}{dof(S{{S}_{E}})} \\ & = & \frac{S{{S}_{E}}}{(n-(k+1))} \\ & = & \frac{423.37}{(17-(2+1))} \\ & = & 30.24 \end{align}\,\![/math]

The statistic to test the significance of regression can now be calculated as:

- [math]\begin{align} {{f}_{0}}& = & \frac{M{{S}_{R}}}{M{{S}_{E}}} \\ & = & \frac{6408.17}{423.37/(17-3)} \\ & = & 211.9 \end{align}\,\![/math]

The critical value for this test, corresponding to a significance level of 0.1, is:

- [math]\begin{align} {{f}_{\alpha ,k,n-(k+1)}} &= & {{f}_{0.1,2,14}} \\ & = & 2.726 \end{align}\,\![/math]

Since [math]{{f}_{0}}\gt {{f}_{0.1,2,14}}\,\![/math], [math]{{H}_{0}}:\,\![/math] [math]{{\beta }_{1}}={{\beta }_{2}}=0\,\![/math] is rejected and it is concluded that at least one coefficient out of [math]{{\beta }_{1}}\,\![/math] and [math]{{\beta }_{2}}\,\![/math] is significant. In other words, it is concluded that a regression model exists between yield and either one or both of the factors in the table. The analysis of variance is summarized in the following table.

### Test on Individual Regression Coefficients (*t* Test)

The [math]t\,\![/math] test is used to check the significance of individual regression coefficients in the multiple linear regression model. Adding a significant variable to a regression model makes the model more effective, while adding an unimportant variable may make the model worse. The hypothesis statements to test the significance of a particular regression coefficient, [math]{{\beta }_{j}}\,\![/math], are:

- [math]\begin{align} & {{H}_{0}}: & {{\beta }_{j}}=0 \\ & {{H}_{1}}: & {{\beta }_{j}}\ne 0 \end{align}\,\![/math]

The test statistic for this test is based on the [math]t\,\![/math] distribution (and is similar to the one used in the case of simple linear regression models in Simple Linear Regression Anaysis):

- [math]{{T}_{0}}=\frac{{{{\hat{\beta }}}_{j}}}{se({{{\hat{\beta }}}_{j}})}\,\![/math]

where the standard error, [math]se({{\hat{\beta }}_{j}})\,\![/math], is obtained. The analyst would fail to reject the null hypothesis if the test statistic lies in the acceptance region:

- [math]-{{t}_{\alpha /2,n-2}}\lt {{T}_{0}}\lt {{t}_{\alpha /2,n-2}}\,\![/math]

This test measures the contribution of a variable while the remaining variables are included in the model. For the model [math]\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{1}}{{x}_{1}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}\,\![/math], if the test is carried out for [math]{{\beta }_{1}}\,\![/math], then the test will check the significance of including the variable [math]{{x}_{1}}\,\![/math] in the model that contains [math]{{x}_{2}}\,\![/math] and [math]{{x}_{3}}\,\![/math] (i.e., the model [math]\hat{y}={{\hat{\beta }}_{0}}+{{\hat{\beta }}_{2}}{{x}_{2}}+{{\hat{\beta }}_{3}}{{x}_{3}}\,\![/math] ). Hence the test is also referred to as partial or marginal test. In DOE folios, this test is displayed in the Regression Information table.

#### Example

The test to check the significance of the estimated regression coefficients for the data is illustrated in this example. The null hypothesis to test the coefficient [math]{{\beta }_{2}}\,\![/math] is:

- [math]{{H}_{0}}:{{\beta }_{2}}=0\,\![/math]

The null hypothesis to test [math]{{\beta }_{1}}\,\![/math] can be obtained in a similar manner. To calculate the test statistic, [math]{{T}_{0}}\,\![/math], we need to calculate the standard error.
In the example, the value of the error mean square, [math]M{{S}_{E}}\,\![/math], was obtained as 30.24. The error mean square is an estimate of the variance, [math]{{\sigma }^{2}}\,\![/math].

- Therefore:

- [math]\begin{align} {{{\hat{\sigma }}}^{2}} &= & M{{S}_{E}} \\ & = & 30.24 \end{align}\,\![/math]

The variance-covariance matrix of the estimated regression coefficients is:

- [math]\begin{align} C &= & {{{\hat{\sigma }}}^{2}}{{({{X}^{\prime }}X)}^{-1}} \\ & = & 30.24\left[ \begin{matrix} 336.5 & 1.2 & -13.1 \\ 1.2 & 0.005 & -0.049 \\ -13.1 & -0.049 & 0.5 \\ \end{matrix} \right] \\ & = & \left[ \begin{matrix} 10176.75 & 37.145 & -395.83 \\ 37.145 & 0.1557 & -1.481 \\ -395.83 & -1.481 & 15.463 \\ \end{matrix} \right] \end{align}\,\![/math]

From the diagonal elements of [math]C\,\![/math], the estimated standard error for [math]{{\hat{\beta }}_{1}}\,\![/math] and [math]{{\hat{\beta }}_{2}}\,\![/math] is:

- [math]\begin{align} se({{{\hat{\beta }}}_{1}}) &= & \sqrt{0.1557}=0.3946 \\ se({{{\hat{\beta }}}_{2}})& = & \sqrt{15.463}=3.93 \end{align}\,\![/math]

The corresponding test statistics for these coefficients are:

- [math]\begin{align} {{({{t}_{0}})}_{{{{\hat{\beta }}}_{1}}}} &= & \frac{{{{\hat{\beta }}}_{1}}}{se({{{\hat{\beta }}}_{1}})}=\frac{1.24}{0.3946}=3.1393 \\ {{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} &= & \frac{{{{\hat{\beta }}}_{2}}}{se({{{\hat{\beta }}}_{2}})}=\frac{12.08}{3.93}=3.0726 \end{align}\,\![/math]

The critical values for the present [math]t\,\![/math] test at a significance of 0.1 are:

- [math]\begin{align} {{t}_{\alpha /2,n-(k+1)}} &= & {{t}_{0.05,14}}=1.761 \\ -{{t}_{\alpha /2,n-(k+1)}} & = & -{{t}_{0.05,14}}=-1.761 \end{align}\,\![/math]

Considering [math]{{\hat{\beta }}_{2}}\,\![/math], it can be seen that [math]{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}}\,\![/math] does not lie in the acceptance region of [math]-{{t}_{0.05,14}}\lt {{t}_{0}}\lt {{t}_{0.05,14}}\,\![/math]. The null hypothesis, [math]{{H}_{0}}:{{\beta }_{2}}=0\,\![/math], is rejected and it is concluded that [math]{{\beta }_{2}}\,\![/math] is significant at [math]\alpha =0.1\,\![/math]. This conclusion can also be arrived at using the [math]p\,\![/math] value noting that the hypothesis is two-sided. The [math]p\,\![/math] value corresponding to the test statistic, [math]{{({{t}_{0}})}_{{{{\hat{\beta }}}_{2}}}} = 3.0726\,\![/math], based on the [math]t\,\![/math] distribution with 14 degrees of freedom is:

- [math]\begin{align} p\text{ }value & = & 2\times (1-P(T\le |{{t}_{0}}|) \\ & = & 2\times (1-0.9959) \\ & = & 0.0083 \end{align}\,\![/math]

Since the [math]p\,\![/math] value is less than the significance, [math]\alpha =0.1\,\![/math], it is concluded that [math]{{\beta }_{2}}\,\![/math] is significant. The hypothesis test on [math]{{\beta }_{1}}\,\![/math] can be carried out in a similar manner.

As explained in Simple Linear Regression Analysis, in DOE folios, the information related to the [math]t\,\![/math] test is displayed in the Regression Information table as shown in the figure below.

In this table, the [math]t\,\![/math] test for [math]{{\beta }_{2}}\,\![/math] is displayed in the row for the term Factor 2 because [math]{{\beta }_{2}}\,\![/math] is the coefficient that represents this factor in the regression model. Columns labeled Standard Error, T Value and P Value represent the standard error, the test statistic for the [math]t\,\![/math] test and the [math]p\,\![/math] value for the [math]t\,\![/math] test, respectively. These values have been calculated for [math]{{\beta }_{2}}\,\![/math] in this example. The Coefficient column represents the estimate of regression coefficients. These values are calculated as shown in this example. The Effect column represents values obtained by multiplying the coefficients by a factor of 2. This value is useful in the case of two factor experiments and is explained in Two-Level Factorial Experiments. Columns labeled Low Confidence and High Confidence represent the limits of the confidence intervals for the regression coefficients and are explained in Confidence Intervals in Multiple Linear Regression. The Variance Inflation Factor column displays values that give a measure of *multicollinearity*. This is explained in Multicollinearity.

### Test on Subsets of Regression Coefficients (Partial *F* Test)

This test can be considered to be the general form of the [math]t\,\![/math] test mentioned in the previous section. This is because the test simultaneously checks the significance of including many (or even one) regression coefficients in the multiple linear regression model. Adding a variable to a model increases the regression sum of squares, [math]S{{S}_{R}}\,\![/math]. The test is based on this increase in the regression sum of squares. The increase in the regression sum of squares is called the *extra sum of squares*.
Assume that the vector of the regression coefficients, [math]\beta\,\![/math], for the multiple linear regression model, [math]y=X\beta +\epsilon\,\![/math], is partitioned into two vectors with the second vector, [math]{{\theta}_{2}}\,\![/math], containing the last [math]r\,\![/math] regression coefficients, and the first vector, [math]{{\theta}_{1}}\,\![/math], containing the first ( [math]k+1-r\,\![/math] ) coefficients as follows:

- [math]\beta =\left[ \begin{matrix} {{\theta}_{1}} \\ {{\theta}_{2}} \\ \end{matrix} \right]\,\![/math]

- with:

- [math]{{\theta}_{1}}=[{{\beta }_{0}},{{\beta }_{1}}...{{\beta }_{k-r}}{]}'\text{ and }{{\theta}_{2}}=[{{\beta }_{k-r+1}},{{\beta }_{k-r+2}}...{{\beta }_{k}}{]}'\text{ }\,\![/math]

The hypothesis statements to test the significance of adding the regression coefficients in [math]{{\theta}_{2}}\,\![/math] to a model containing the regression coefficients in [math]{{\theta}_{1}}\,\![/math] may be written as:

- [math]\begin{align} & {{H}_{0}}: & {{\theta}_{2}}=0 \\ & {{H}_{1}}: & {{\theta}_{2}}\ne 0 \end{align}\,\![/math]

The test statistic for this test follows the [math]F\,\![/math] distribution and can be calculated as follows:

- [math]{{F}_{0}}=\frac{S{{S}_{R}}({{\theta}_{2}}|{{\theta}_{1}})/r}{M{{S}_{E}}}\,\![/math]

where [math]S{{S}_{R}}({{\theta}_{2}}|{{\theta}_{1}})\,\![/math] is the the increase in the regression sum of squares when the variables corresponding to the coefficients in [math]{{\theta}_{2}}\,\![/math] are added to a model already containing [math]{{\theta}_{1}}\,\![/math], and [math]M{{S}_{E}}\,\![/math] is obtained from the equation given in Simple Linear Regression Analysis. The value of the extra sum of squares is obtained as explained in the next section.

The null hypothesis, [math]{{H}_{0}}\,\![/math], is rejected if [math]{{F}_{0}}\gt {{f}_{\alpha ,r,n-(k+1)}}\,\![/math]. Rejection of [math]{{H}_{0}}\,\![/math] leads to the conclusion that at least one of the variables in [math]{{x}_{k-r+1}}\,\![/math], [math]{{x}_{k-r+2}}\,\![/math]... [math]{{x}_{k}}\,\![/math] contributes significantly to the regression model. In a DOE folio, the results from the partial [math]F\,\![/math] test are displayed in the ANOVA table.

### Types of Extra Sum of Squares

The extra sum of squares can be calculated using either the partial (or adjusted) sum of squares or the sequential sum of squares. The type of extra sum of squares used affects the calculation of the test statistic for the partial [math]F\,\![/math] test described above. In DOE folios, selection for the type of extra sum of squares is available as shown in the figure below. The partial sum of squares is used as the default setting. The reason for this is explained in the following section on the partial sum of squares.

#### Partial Sum of Squares

The partial sum of squares for a term is the extra sum of squares when all terms, except the term under consideration, are included in the model. For example, consider the model:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon\,\![/math]

The sum of squares of regression of this model is denoted by [math]S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})\,\![/math]. Assume that we need to know the partial sum of squares for [math]{{\beta }_{2}}\,\![/math]. The partial sum of squares for [math]{{\beta }_{2}}\,\![/math] is the increase in the regression sum of squares when [math]{{\beta }_{2}}\,\![/math] is added to the model. This increase is the difference in the regression sum of squares for the full model of the equation given above and the model that includes all terms except [math]{{\beta }_{2}}\,\![/math]. These terms are [math]{{\beta }_{0}}\,\![/math], [math]{{\beta }_{1}}\,\![/math] and [math]{{\beta }_{12}}\,\![/math]. The model that contains these terms is:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+\epsilon\,\![/math]

The sum of squares of regression of this model is denoted by [math]S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})\,\![/math]. The partial sum of squares for [math]{{\beta }_{2}}\,\![/math]can be represented as [math]S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})\,\![/math] and is calculated as follows:

- [math]\begin{align} S{{S}_{R}}({{\beta }_{2}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}})& = & S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}) \end{align}\,\![/math]

For the present case, [math]{{\theta}_{2}}=[{{\beta }_{2}}{]}'\,\![/math] and [math]{{\theta}_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{12}}{]}'\,\![/math]. It can be noted that for the partial sum of squares [math]{{\beta }_{1}}\,\![/math] contains all coefficients other than the coefficient being tested.

A Weibull++ DOE folio has the partial sum of squares as the default selection. This is because the [math]t\,\![/math] test is a partial test, i.e., the [math]t\,\![/math] test on an individual coefficient is carried by assuming that all the remaining coefficients are included in the model (similar to the way the partial sum of squares is calculated). The results from the [math]t\,\![/math] test are displayed in the Regression Information table. The results from the partial [math]F\,\![/math] test are displayed in the ANOVA table. To keep the results in the two tables consistent with each other, the partial sum of squares is used as the default selection for the results displayed in the ANOVA table. The partial sum of squares for all terms of a model may not add up to the regression sum of squares for the full model when the regression coefficients are correlated. If it is preferred that the extra sum of squares for all terms in the model always add up to the regression sum of squares for the full model then the sequential sum of squares should be used.

##### Example

This example illustrates the [math]F\,\![/math] test using the partial sum of squares. The test is conducted for the coefficient [math]{{\beta }_{1}}\,\![/math] corresponding to the predictor variable [math]{{x}_{1}}\,\![/math] for the data. The regression model used for this data set in the example is:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math]

The null hypothesis to test the significance of [math]{{\beta }_{1}}\,\![/math] is:

- [math]{{H}_{0}}: {{\beta }_{1}}=0\,\![/math]

The statistic to test this hypothesis is:

- [math]{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{1}}|{{\beta }_{2}})/r}{M{{S}_{E}}}\,\![/math]

where [math]S{{S}_{R}}({{\beta }_{1}}|{{\beta }_{2}})\,\![/math] represents the partial sum of squares for [math]{{\beta }_{1}}\,\![/math], [math]r\,\![/math] represents the number of degrees of freedom for [math]S{{S}_{R}}({{\beta }_{1}}|{{\beta }_{2}})\,\![/math] (which is one because there is just one coefficient, [math]{{\beta }_{1}}\,\![/math], being tested) and [math]M{{S}_{E}}\,\![/math] is the error mean square and has been calculated in the second example as 30.24.

The partial sum of squares for [math]{{\beta }_{1}}\,\![/math] is the difference between the regression sum of squares for the full model, [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math], and the regression sum of squares for the model excluding [math]{{\beta }_{1}}\,\![/math], [math]Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math]. The regression sum of squares for the full model has been calculated in the second example as 12816.35. Therefore:

- [math]S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})=12816.35\,\![/math]

The regression sum of squares for the model [math]Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math] is obtained as shown next. First the design matrix for this model, [math]{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}\,\![/math], is obtained by dropping the second column in the design matrix of the full model, [math]X\,\![/math] (the full design matrix, [math]X\,\![/math], was obtained in the example). The second column of [math]X\,\![/math] corresponds to the coefficient [math]{{\beta }_{1}}\,\![/math] which is no longer in the model. Therefore, the design matrix for the model, [math]Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math], is:

- [math]{{X}_{{{\beta }_{0}},{{\beta }_{2}}}}=\left[ \begin{matrix} 1 & 29.1 \\ 1 & 29.3 \\ . & . \\ . & . \\ 1 & 32.9 \\ \end{matrix} \right]\,\![/math]

The hat matrix corresponding to this design matrix is [math]{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}\,\![/math]. It can be calculated using [math]{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}={{X}_{{{\beta }_{0}},{{\beta }_{2}}}}{{(X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{2}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{2}}}^{\prime }\,\![/math]. Once [math]{{H}_{{{\beta }_{0}},{{\beta }_{2}}}}\,\![/math] is known, the regression sum of squares for the model [math]Y={{\beta }_{0}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math], can be calculated as:

- [math]\begin{align} S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) & = & {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{2}}}}-(\frac{1}{n})J \right]y \\ & = & 12518.32 \end{align}\,\![/math]

Therefore, the partial sum of squares for [math]{{\beta }_{1}}\,\![/math] is:

- [math]\begin{align} S{{S}_{R}}({{\beta }_{1}}|{{\beta }_{2}})& = & S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}})-S{{S}_{R}}({{\beta }_{0}},{{\beta }_{2}}) \\ & = & 12816.35-12518.32 \\ & = & 298.03 \end{align}\,\![/math]

Knowing the partial sum of squares, the statistic to test the significance of [math]{{\beta }_{1}}\,\![/math] is:

- [math]\begin{align} {{f}_{0}} &= & \frac{S{{S}_{R}}({{\beta }_{1}}|{{\beta }_{2}})/r}{M{{S}_{E}}} \\ & = & \frac{298.03/1}{30.24} \\ & = & 9.855 \end{align}\,\![/math]

The [math]p\,\![/math] value corresponding to this statistic based on the [math]F\,\![/math] distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is:

- [math]\begin{align} p\text{ }value &= & 1-P(F\le {{f}_{0}}) \\ & = & 1-0.9928 \\ & = & 0.0072 \end{align}\,\![/math]

Assuming that the desired significance is 0.1, since [math]p\,\![/math] value < 0.1, [math]{{H}_{0}}:{{\beta }_{1}}=0\,\![/math] is rejected and it can be concluded that [math]{{\beta }_{1}}\,\![/math] is significant. The test for [math]{{\beta }_{2}}\,\![/math] can be carried out in a similar manner. In the results obtained from the DOE folio, the calculations for this test are displayed in the ANOVA table as shown in the following figure. Note that the conclusion obtained in this example can also be obtained using the [math]t\,\![/math] test as explained in the example in Test on Individual Regression Coefficients (t Test). The ANOVA and Regression Information tables in the DOE folio represent two different ways to test for the significance of the variables included in the multiple linear regression model.

#### Sequential Sum of Squares

The sequential sum of squares for a coefficient is the extra sum of squares when coefficients are added to the model in a sequence. For example, consider the model:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+{{\beta }_{23}}{{x}_{2}}{{x}_{3}}+{{\beta }_{123}}{{x}_{1}}{{x}_{2}}{{x}_{3}}+\epsilon\,\![/math]

The sequential sum of squares for [math]{{\beta }_{13}}\,\![/math] is the increase in the sum of squares when [math]{{\beta }_{13}}\,\![/math] is added to the model observing the sequence of the equation given above. Therefore this extra sum of squares can be obtained by taking the difference between the regression sum of squares for the model after [math]{{\beta }_{13}}\,\![/math] was added and the regression sum of squares for the model before [math]{{\beta }_{13}}\,\![/math] was added to the model. The model after [math]{{\beta }_{13}}\,\![/math] is added is as follows:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+{{\beta }_{13}}{{x}_{1}}{{x}_{3}}+\epsilon\,\![/math]

This is because to maintain the sequence all coefficients preceding [math]{{\beta }_{13}}\,\![/math] must be included in the model. These are the coefficients [math]{{\beta }_{0}}\,\![/math], [math]{{\beta }_{1}}\,\![/math], [math]{{\beta }_{2}}\,\![/math], [math]{{\beta }_{12}}\,\![/math] and [math]{{\beta }_{3}}\,\![/math].
Similarly the model before [math]{{\beta }_{13}}\,\![/math] is added must contain all coefficients of the equation given above except [math]{{\beta }_{13}}\,\![/math]. This model can be obtained as follows:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+{{\beta }_{12}}{{x}_{1}}{{x}_{2}}+{{\beta }_{3}}{{x}_{3}}+\epsilon\,\![/math]

The sequential sum of squares for [math]{{\beta }_{13}}\,\![/math] can be calculated as follows:

- [math]\begin{align} S{{S}_{R}}({{\beta }_{13}}|{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}) & = & S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}},{{\beta }_{13}})- S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}) \end{align}\,\![/math]

For the present case, [math]{{\theta}_{2}}=[{{\beta }_{13}}{]}'\,\![/math] and [math]{{\theta}_{1}}=[{{\beta }_{0}},{{\beta }_{1}},{{\beta }_{2}},{{\beta }_{12}},{{\beta }_{3}}{]}'\,\![/math]. It can be noted that for the sequential sum of squares [math]{{\beta }_{1}}\,\![/math] contains all coefficients proceeding the coefficient being tested.

The sequential sum of squares for all terms will add up to the regression sum of squares for the full model, but the sequential sum of squares are order dependent.

##### Example

This example illustrates the partial [math]F\,\![/math] test using the sequential sum of squares. The test is conducted for the coefficient [math]{{\beta }_{1}}\,\![/math] corresponding to the predictor variable [math]{{x}_{1}}\,\![/math] for the data. The regression model used for this data set in the example is:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon \,\![/math]

The null hypothesis to test the significance of [math]{{\beta }_{1}}\,\![/math] is:

- [math]{{H}_{0}}:{{\beta }_{1}}=0\,\![/math]

The statistic to test this hypothesis is:

- [math]{{F}_{0}}=\frac{S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})/r}{M{{S}_{E}}}\,\![/math]

where [math]S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})\,\![/math] represents the sequential sum of squares for [math]{{\beta }_{1}}\,\![/math], [math]r\,\![/math] represents the number of degrees of freedom for [math]S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})\,\![/math] (which is one because there is just one coefficient, [math]{{\beta }_{1}}\,\![/math], being tested) and [math]M{{S}_{E}}\,\![/math] is the error mean square and has been calculated in the second example as 30.24.

The sequential sum of squares for [math]{{\beta }_{1}}\,\![/math] is the difference between the regression sum of squares for the model after adding [math]{{\beta }_{1}}\,\![/math], [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon\,\![/math], and the regression sum of squares for the model before adding [math]{{\beta }_{1}}\,\![/math], [math]Y={{\beta }_{0}}+\epsilon\,\![/math]. The regression sum of squares for the model [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon\,\![/math] is obtained as shown next. First the design matrix for this model, [math]{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}\,\![/math], is obtained by dropping the third column in the design matrix for the full model, [math]X\,\![/math] (the full design matrix, [math]X\,\![/math], was obtained in the example). The third column of [math]X\,\![/math] corresponds to coefficient [math]{{\beta }_{2}}\,\![/math] which is no longer used in the present model. Therefore, the design matrix for the model, [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon\,\![/math], is:

- [math]{{X}_{{{\beta }_{0}},{{\beta }_{1}}}}=\left[ \begin{matrix} 1 & 41.9 \\ 1 & 43.4 \\ . & . \\ . & . \\ 1 & 77.8 \\ \end{matrix} \right]\,\![/math]

The hat matrix corresponding to this design matrix is [math]{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}\,\![/math]. It can be calculated using [math]{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}={{X}_{{{\beta }_{0}},{{\beta }_{1}}}}{{(X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }{{X}_{{{\beta }_{0}},{{\beta }_{1}}}})}^{-1}}X_{{{\beta }_{0}},{{\beta }_{1}}}^{\prime }\,\![/math]. Once [math]{{H}_{{{\beta }_{0}},{{\beta }_{1}}}}\,\![/math] is known, the regression sum of squares for the model [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+\epsilon\,\![/math] can be calculated as:

- [math]\begin{align} S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})& = & {{y}^{\prime }}\left[ {{H}_{{{\beta }_{0}},{{\beta }_{1}}}}-(\frac{1}{n})J \right]y \\ & = & 12530.85 \end{align}\,\![/math]

The regression sum of squares for the model [math]Y={{\beta }_{0}}+\epsilon\,\![/math] is equal to zero since this model does not contain any variables. Therefore:

- [math]S{{S}_{R}}({{\beta }_{0}})=0\,\![/math]

The sequential sum of squares for [math]{{\beta }_{1}}\,\![/math] is:

- [math]\begin{align} S{{S}_{R}}({{\beta }_{1}}|{{\beta }_{0}}) &= & S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})-S{{S}_{R}}({{\beta }_{0}}) \\ & = & 12530.85-0 \\ & = & 12530.85 \end{align}\,\![/math]

Knowing the sequential sum of squares, the statistic to test the significance of [math]{{\beta }_{1}}\,\![/math] is:

- [math]\begin{align} {{f}_{0}} &= & \frac{S{{S}_{R}}({{\beta }_{0}},{{\beta }_{1}})/r}{M{{S}_{E}}} \\ & = & \frac{12530.85/1}{30.24} \\ & = & 414.366 \end{align}\,\![/math]

The [math]p\,\![/math] value corresponding to this statistic based on the [math]F\,\![/math] distribution with 1 degree of freedom in the numerator and 14 degrees of freedom in the denominator is:

- [math]\begin{align} p\text{ }value &= & 1-P(F\le {{f}_{0}}) \\ & = & 1-0.999999 \\ & = & 8.46\times {{10}^{-12}} \end{align}\,\![/math]

Assuming that the desired significance is 0.1, since [math]p\,\![/math] value < 0.1, [math]{{H}_{0}}:{{\beta }_{1}}=0\,\![/math] is rejected and it can be concluded that [math]{{\beta }_{1}}\,\![/math] is significant. The test for [math]{{\beta }_{2}}\,\![/math] can be carried out in a similar manner. This result is shown in the following figure.

## Confidence Intervals in Multiple Linear Regression

Calculation of confidence intervals for multiple linear regression models are similar to those for simple linear regression models explained in Simple Linear Regression Analysis.

### Confidence Interval on Regression Coefficients

A 100 ([math]1-\alpha\,\![/math]) percent confidence interval on the regression coefficient, [math]{{\beta }_{j}}\,\![/math], is obtained as follows:

- [math]{{\hat{\beta }}_{j}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{C}_{jj}}}\,\![/math]

The confidence interval on the regression coefficients are displayed in the Regression Information table under the Low Confidence and High Confidence columns as shown in the following figure.

Confidence Interval on Fitted Values, [math]{{\hat{y}}_{i}}\,\![/math]
A 100 ([math]1-\alpha\,\![/math]) percent confidence interval on any fitted value, [math]{{\hat{y}}_{i}}\,\![/math], is given by:

- [math]{{\hat{y}}_{i}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}x_{i}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{i}}}\,\![/math]

where:

- [math]{{x}_{i}}=\left[ \begin{matrix} 1 \\ {{x}_{i1}} \\ . \\ . \\ . \\ {{x}_{ik}} \\ \end{matrix} \right]\,\![/math]

In the above example, the fitted value corresponding to the fifth observation was calculated as [math]{{\hat{y}}_{5}}=266.3\,\![/math]. The 90% confidence interval on this value can be obtained as shown in the figure below. The values of 47.3 and 29.9 used in the figure are the values of the predictor variables corresponding to the fifth observation the table.

### Confidence Interval on New Observations

As explained in Simple Linear Regression Analysis, the confidence interval on a new observation is also referred to as the prediction interval. The prediction interval takes into account both the error from the fitted model and the error associated with future observations. A 100 ([math]1-\alpha\,\![/math]) percent confidence interval on a new observation, [math]{{\hat{y}}_{p}}\,\![/math], is obtained as follows:

- [math]{{\hat{y}}_{p}}\pm {{t}_{\alpha /2,n-(k+1)}}\sqrt{{{{\hat{\sigma }}}^{2}}(1+x_{p}^{\prime }{{({{X}^{\prime }}X)}^{-1}}{{x}_{p}})}\,\![/math]

where:

- [math]{{x}_{p}}=\left[ \begin{matrix} 1 \\ {{x}_{p1}} \\ . \\ . \\ . \\ {{x}_{pk}} \\ \end{matrix} \right]\,\![/math]

[math]{{x}_{p1}}\,\![/math],..., [math]{{x}_{pk}}\,\![/math] are the levels of the predictor variables at which the new observation, [math]{{\hat{y}}_{p}}\,\![/math], needs to be obtained.

In multiple linear regression, prediction intervals should only be obtained at the levels of the predictor variables where the regression model applies. In the case of multiple linear regression it is easy to miss this. Having values lying within the range of the predictor variables does not necessarily mean that the new observation lies in the region to which the model is applicable. For example, consider the next figure where the shaded area shows the region to which a two variable regression model is applicable. The point corresponding to [math]p\,\![/math] th level of first predictor variable, [math]{{x}_{1}}\,\![/math], and [math]p\,\![/math] th level of the second predictor variable, [math]{{x}_{2}}\,\![/math], does not lie in the shaded area, although both of these levels are within the range of the first and second predictor variables respectively. In this case, the regression model is not applicable at this point.

## Measures of Model Adequacy

As in the case of simple linear regression, analysis of a fitted multiple linear regression model is important before inferences based on the model are undertaken. This section presents some techniques that can be used to check the appropriateness of the multiple linear regression model.

### Coefficient of Multiple Determination, *R*^{2}

The coefficient of multiple determination is similar to the coefficient of determination used in the case of simple linear regression. It is defined as:

- [math]\begin{align} {{R}^{2}} & = & \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ & = & 1-\frac{S{{S}_{E}}}{S{{S}_{T}}} \end{align}\,\![/math]

[math]{{R}^{2}}\,\![/math] indicates the amount of total variability explained by the regression model. The positive square root of [math]{{R}^{2}}\,\![/math] is called the multiple correlation coefficient and measures the linear association between [math]Y\,\![/math] and the predictor variables, [math]{{x}_{1}}\,\![/math], [math]{{x}_{2}}\,\![/math]... [math]{{x}_{k}}\,\![/math].

The value of [math]{{R}^{2}}\,\![/math] increases as more terms are added to the model, even if the new term does not contribute significantly to the model. An increase in the value of [math]{{R}^{2}}\,\![/math] cannot be taken as a sign to conclude that the new model is superior to the older model. A better statistic to use is the adjusted [math]{{R}^{2}}\,\![/math] statistic defined as follows:

- [math]\begin{align} R_{adj}^{2} &= & 1-\frac{M{{S}_{E}}}{M{{S}_{T}}} \\ & = & 1-\frac{S{{S}_{E}}/(n-(k+1))}{S{{S}_{T}}/(n-1)} \\ & = & 1-(\frac{n-1}{n-(k+1)})(1-{{R}^{2}}) \end{align}\,\![/math]

The adjusted [math]{{R}^{2}}\,\![/math] only increases when significant terms are added to the model. Addition of unimportant terms may lead to a decrease in the value of [math]R_{adj}^{2}\,\![/math].

In a DOE folio, [math]{{R}^{2}}\,\![/math] and [math]R_{adj}^{2}\,\![/math] values are displayed as R-sq and R-sq(adj), respectively. Other values displayed along with these values are S, PRESS and R-sq(pred). As explained in Simple Linear Regression Analysis, the value of S is the square root of the error mean square, [math]M{{S}_{E}}\,\![/math], and represents the "standard error of the model."

PRESS is an abbreviation for prediction error sum of squares. It is the error sum of squares calculated using the PRESS residuals in place of the residuals, [math]{{e}_{i}}\,\![/math], in the equation for the error sum of squares. The PRESS residual, [math]{{e}_{(i)}}\,\![/math], for a particular observation, [math]{{y}_{i}}\,\![/math], is obtained by fitting the regression model to the remaining observations. Then the value for a new observation, [math]{{\hat{y}}_{p}}\,\![/math], corresponding to the observation in question, [math]{{y}_{i}}\,\![/math], is obtained based on the new regression model. The difference between [math]{{y}_{i}}\,\![/math] and [math]{{\hat{y}}_{p}}\,\![/math] gives [math]{{e}_{(i)}}\,\![/math]. The PRESS residual, [math]{{e}_{(i)}}\,\![/math], can also be obtained using [math]{{h}_{ii}}\,\![/math], the diagonal element of the hat matrix, [math]H\,\![/math], as follows:

- [math]{{e}_{(i)}}=\frac{{{e}_{i}}}{1-{{h}_{ii}}}\,\![/math]

R-sq(pred), also referred to as prediction [math]{{R}^{2}}\,\![/math], is obtained using PRESS as shown next:

- [math]R_{pred}^{2}=1-\frac{PRESS}{S{{S}_{T}}}\,\![/math]

The values of R-sq, R-sq(adj) and S are indicators of how well the regression model fits the observed data. The values of PRESS and R-sq(pred) are indicators of how well the regression model predicts new observations. For example, higher values of PRESS or lower values of R-sq(pred) indicate a model that predicts poorly. The figure below shows these values for the data. The values indicate that the regression model fits the data well and also predicts well.

### Residual Analysis

Plots of residuals, [math]{{e}_{i}}\,\![/math], similar to the ones discussed in Simple Linear Regression Analysis for simple linear regression, are used to check the adequacy of a fitted multiple linear regression model. The residuals are expected to be normally distributed with a mean of zero and a constant variance of [math]{{\sigma }^{2}}\,\![/math]. In addition, they should not show any patterns or trends when plotted against any variable or in a time or run-order sequence. Residual plots may also be obtained using standardized and studentized residuals. Standardized residuals, [math]{{d}_{i}}\,\![/math], are obtained using the following equation:

- [math]\begin{align} {{d}_{i}}&= & \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}}} \\ & = & \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}}} \end{align}\,\![/math]

Standardized residuals are scaled so that the standard deviation of the residuals is approximately equal to one. This helps to identify possible outliers or unusual observations. However, standardized residuals may understate the true residual magnitude, hence studentized residuals, [math]{{r}_{i}}\,\![/math], are used in their place. Studentized residuals are calculated as follows:

- [math] \begin{align} {{r}_{i}} & = & \frac{{{e}_{i}}}{\sqrt{{{{\hat{\sigma }}}^{2}}(1-{{h}_{ii}})}} \\ & = & \frac{{{e}_{i}}}{\sqrt{M{{S}_{E}}(1-{{h}_{ii}})}} \end{align} \,\! [/math]

where [math]{{h}_{ii}}\,\![/math] is the [math]i\,\![/math] th diagonal element of the hat matrix, [math]H\,\![/math]. External studentized (or the studentized deleted) residuals may also be used. These residuals are based on the PRESS residuals mentioned in Coefficient of Multiple Determination, *R*^{2}. The reason for using the external studentized residuals is that if the [math]i\,\![/math] th observation is an outlier, it may influence the fitted model. In this case, the residual [math]{{e}_{i}}\,\![/math] will be small and may not disclose that [math]i\,\![/math] th observation is an outlier. The external studentized residual for the [math]i\,\![/math] th observation, [math]{{t}_{i}}\,\![/math], is obtained as follows:

- [math]{{t}_{i}}={{e}_{i}}{{\left[ \frac{n-k}{S{{S}_{E}}(1-{{h}_{ii}})-e_{i}^{2}} \right]}^{0.5}}\,\![/math]

Residual values for the data are shown in the figure below. Standardized residual plots for the data are shown in next two figures. The Weibull++ DOE folio compares the residual values to the critical values on the [math]t\,\![/math] distribution for studentized and external studentized residuals.

For other residuals the normal distribution is used. For example, for the data, the critical values on the [math]t\,\![/math] distribution at a significance of 0.1 are [math]{{t}_{0.05,14}}=1.761\,\![/math] and [math]-{{t}_{0.05,14}}=-1.761\,\![/math] (as calculated in the example, Test on Individual Regression Coefficients (*t* Test)). The studentized residual values corresponding to the 3rd and 17th observations lie outside the critical values. Therefore, the 3rd and 17th observations are outliers. This can also be seen on the residual plots in the next two figures.

### Outlying *x* Observations

Residuals help to identify outlying [math]y\,\![/math] observations. Outlying [math]x\,\![/math] observations can be detected using leverage. Leverage values are the diagonal elements of the hat matrix, [math]{{h}_{ii}}\,\![/math]. The [math]{{h}_{ii}}\,\![/math] values always lie between 0 and 1. Values of [math]{{h}_{ii}}\,\![/math] greater than [math]2(k+1)/n\,\![/math] are considered to be indicators of outlying [math]x\,\![/math] observations.

### Influential Observations Detection

Once an outlier is identified, it is important to determine if the outlier has a significant effect on the regression model. One measure to detect influential observations is Cook's distance measure which is computed as follows:

- [math]{{D}_{i}}=\frac{r_{i}^{2}}{(k+1)}\left[ \frac{{{h}_{ii}}}{(1-{{h}_{ii}})} \right]\,\![/math]

To use Cook's distance measure, the [math]{{D}_{i}}\,\![/math] values are compared to percentile values on the [math]F\,\![/math] distribution with [math](k+1,n-(k+1))\,\![/math] degrees of freedom. If the percentile value is less than 10 or 20 percent, then the [math]i\,\![/math] th case has little influence on the fitted values. However, if the percentile value is close to 50 percent or greater, the [math]i\,\![/math] th case is influential, and fitted values with and without the [math]i\,\![/math] th case will differ substantially.

#### Example

Cook's distance measure can be calculated as shown next. The distance measure is calculated for the first observation of the data. The remaining values along with the leverage values are shown in the figure below (displaying Leverage and Cook's distance measure for the data).

The standardized residual corresponding to the first observation is:

- [math]\begin{align} {{r}_{1}} & = & \frac{{{e}_{1}}}{\sqrt{M{{S}_{E}}(1-{{h}_{11}})}} \\ & = & \frac{1.3127}{\sqrt{30.3(1-0.2755)}} \\ & = & 0.2804 \end{align}\,\![/math]

Cook's distance measure for the first observation can now be calculated as:

- [math]\begin{align} {{D}_{1}} & = & \frac{r_{1}^{2}}{(k+1)}\left[ \frac{{{h}_{11}}}{(1-{{h}_{11}})} \right] \\ & = & \frac{{{0.2804}^{2}}}{(2+1)}\left[ \frac{0.2755}{(1-0.2755)} \right] \\ & = & 0.01 \end{align}\,\![/math]

The 50th percentile value for [math]{{F}_{3,14}}\,\![/math] is 0.83. Since all [math]{{D}_{i}}\,\![/math] values are less than this value there are no influential observations.

### Lack-of-Fit Test

The lack-of-fit test for simple linear regression discussed in Simple Linear Regression Analysis may also be applied to multiple linear regression to check the appropriateness of the fitted response surface and see if a higher order model is required. Data for [math]m\,\![/math] replicates may be collected as follows for all [math]n\,\![/math] levels of the predictor variables:

- [math]\begin{align} & & {{y}_{11}},{{y}_{12}},....,{{y}_{1m}}\text{ }m\text{ repeated observations at the first level } \\ & & {{y}_{21}},{{y}_{22}},....,{{y}_{2m}}\text{ }m\text{ repeated observations at the second level} \\ & & ... \\ & & {{y}_{i1}},{{y}_{i2}},....,{{y}_{im}}\text{ }m\text{ repeated observations at the }i\text{th level} \\ & & ... \\ & & {{y}_{n1}},{{y}_{n2}},....,{{y}_{nm}}\text{ }m\text{ repeated observations at the }n\text{th level } \end{align}\,\![/math]

The sum of squares due to pure error, [math]S{{S}_{PE}}\,\![/math], can be obtained as discussed in the Simple Linear Regression Analysis as:

- [math]S{{S}_{PE}}=\underset{i=1}{\overset{n}{\mathop \sum }}\,\underset{j=1}{\overset{m}{\mathop \sum }}\,{{({{y}_{ij}}-{{\bar{y}}_{i}})}^{2}}\,\![/math]

The number of degrees of freedom associated with [math]S{{S}_{PE}}\,\![/math] are:

- [math] dof(S{S}_{PE}) = nm-n \,\! [/math]

Knowing [math]S{{S}_{PE}}\,\![/math], sum of squares due to lack-of-fit, [math]S{{S}_{LOF}}\,\![/math], can be obtained as:

- [math]S{{S}_{LOF}}=S{{S}_{E}}-S{{S}_{PE}}\,\![/math]

The number of degrees of freedom associated with [math]S{{S}_{LOF}}\,\![/math] are:

- [math] \begin{align} dof(S{{S}_{LOF}}) & = & dof(S{{S}_{E}})-dof(S{{S}_{PE}}) \\ & = & n-(k+1)-(nm-n) \end{align} \,\! [/math]

The test statistic for the lack-of-fit test is:

- [math]\begin{align} {{F}_{0}} & = & \frac{S{{S}_{LOF}}/dof(S{{S}_{LOF}})}{S{{S}_{PE}}/dof(S{{S}_{PE}})} \\ & = & \frac{M{{S}_{LOF}}}{M{{S}_{PE}}} \end{align}\,\![/math]

## Other Topics in Multiple Linear Regression

### Polynomial Regression Models

Polynomial regression models are used when the response is curvilinear. The equation shown next presents a second order polynomial regression model with one predictor variable:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{11}}x_{1}^{2}+\epsilon\,\![/math]

Usually, coded values are used in these models. Values of the variables are coded by centering or expressing the levels of the variable as deviations from the mean value of the variable and then scaling or dividing the deviations obtained by half of the range of the variable.

- [math]coded\text{ }value=\frac{actual\text{ }value-mean}{half\text{ }of\text{ }range}\,\![/math]

The reason for using coded predictor variables is that many times [math]x\,\![/math] and [math]{{x}^{2}}\,\![/math] are highly correlated and, if uncoded values are used, there may be computational difficulties while calculating the [math]{{({{X}^{\prime }}X)}^{-1}}\,\![/math] matrix to obtain the estimates, [math]\hat{\beta }\,\![/math], of the regression coefficients using the equation for the [math]F\,\![/math] distribution given in Statistics Background on DOE.

### Qualitative Factors

The multiple linear regression model also supports the use of qualitative factors. For example, gender may need to be included as a factor in a regression model. One of the ways to include qualitative factors in a regression model is to employ indicator variables. Indicator variables take on values of 0 or 1. For example, an indicator variable may be used with a value of 1 to indicate female and a value of 0 to indicate male.

- [math]{{x}_{1}}=\{\begin{array}{*{35}{l}} 1\text{ Female} \\ 0\text{ Male} \\ \end{array}\,\![/math]

In general ( [math]n-1\,\![/math] ) indicator variables are required to represent a qualitative factor with [math]n\,\![/math] levels. As an example, a qualitative factor representing three types of machines may be represented as follows using two indicator variables:

- [math]\begin{align} {{x}_{1}} & = & 1,\text{ }{{x}_{2}} & = & 0\text{ Machine Type I} \\ {{x}_{1}} & = & 0,\text{ }{{x}_{2}} & = & 1\text{ Machine Type II} \\ {{x}_{1}} & = & 0,\text{ }{{x}_{2}} & = & 0\text{ Machine Type III} \end{align}\,\![/math]

An alternative coding scheme for this example is to use a value of -1 for all indicator variables when representing the last level of the factor:

- [math]\begin{align} {{x}_{1}} & = & 1,\text{ }{{x}_{2}}& = &0\text{ Machine Type I} \\ {{x}_{1}}& = & 0,\text{ }{{x}_{2}}& = &1\text{ Machine Type II} \\ {{x}_{1}}& = & -1,\text{ }{{x}_{2}}& = &-1\text{ Machine Type III} \end{align}\,\![/math]

Indicator variables are also referred to as dummy variables or binary variables.

#### Example

Consider data from two types of reactors of a chemical process shown where the yield values are recorded for various levels of factor [math]{{x}_{1}}\,\![/math]. Assuming there are no interactions between the reactor type and [math]{{x}_{1}}\,\![/math], a regression model can be fitted to this data as shown next.

Since the reactor type is a qualitative factor with two levels, it can be represented by using one indicator variable. Let [math]{{x}_{2}}\,\![/math] be the indicator variable representing the reactor type, with 0 representing the first type of reactor and 1 representing the second type of reactor.

- [math]{{x}_{2}}=\{\begin{array}{*{35}{l}} 0\text{ Reactor Type I} \\ 1\text{ Reactor Type II} \\ \end{array}\,\![/math]

Data entry in the DOE folio for this example is shown in the figure after the table below. The regression model for this data is:

- [math]Y={{\beta }_{0}}+{{\beta }_{1}}{{x}_{1}}+{{\beta }_{2}}{{x}_{2}}+\epsilon\,\![/math]

The [math]X\,\![/math] and [math]y\,\![/math] matrices for the given data are:

The estimated regression coefficients for the model can be obtained as:

- [math]\begin{align} \hat{\beta }& = & {{({{X}^{\prime }}X)}^{-1}}{{X}^{\prime }}y \\ & = & \left[ \begin{matrix} 153.7 \\ 2.4 \\ -27.5 \\ \end{matrix} \right] \end{align}\,\![/math]

Therefore, the fitted regression model is:

- [math]\hat{y}=153.7+2.4{{x}_{1}}-27.5{{x}_{2}}\,\![/math]

Note that since [math]{{x}_{2}}\,\![/math] represents a qualitative predictor variable, the fitted regression model cannot be plotted simultaneously against [math]{{x}_{1}}\,\![/math] and [math]{{x}_{2}}\,\![/math] in a two-dimensional space (because the resulting surface plot will be meaningless for the dimension in [math]{{x}_{2}}\,\![/math] ). To illustrate this, a scatter plot of the data against [math]{{x}_{2}}\,\![/math] is shown in the following figure.

It can be noted that, in the case of qualitative factors, the nature of the relationship between the response (yield) and the qualitative factor (reactor type) cannot be categorized as linear, or quadratic, or cubic, etc. The only conclusion that can be arrived at for these factors is to see if these factors contribute significantly to the regression model. This can be done by employing the partial [math]F\,\![/math] test discussed in Multiple Linear Regression Analysis (using the extra sum of squares of the indicator variables representing these factors). The results of the test for the present example are shown in the ANOVA table. The results show that [math]{{x}_{2}}\,\![/math] (reactor type) contributes significantly to the fitted regression model.

### Multicollinearity

At times the predictor variables included in a multiple linear regression model may be found to be dependent on each other. Multicollinearity is said to exist in a multiple regression model with strong dependencies between the predictor variables. Multicollinearity affects the regression coefficients and the extra sum of squares of the predictor variables. In a model with multicollinearity the estimate of the regression coefficient of a predictor variable depends on what other predictor variables are included the model. The dependence may even lead to change in the sign of the regression coefficient. In a such models, an estimated regression coefficient may not be found to be significant individually (when using the [math]t\,\![/math] test on the individual coefficient or looking at the [math]p\,\![/math] value) even though a statistical relation is found to exist between the response variable and the set of the predictor variables (when using the [math]F\,\![/math] test for the set of predictor variables). Therefore, you should be careful while looking at individual predictor variables in models that have multicollinearity. Care should also be taken while looking at the extra sum of squares for a predictor variable that is correlated with other variables. This is because in models with multicollinearity the extra sum of squares is not unique and depends on the other predictor variables included in the model.

Multicollinearity can be detected using the variance inflation factor (abbreviated [math]VIF\,\![/math] ). [math]VIF\,\![/math] for a coefficient [math]{{\beta }_{j}}\,\![/math] is defined as:

- [math]VIF=\frac{1}{(1-R_{j}^{2})}\,\![/math]

where [math]R_{j}^{2}\,\![/math] is the coefficient of multiple determination resulting from regressing the [math]j\,\![/math] th predictor variable, [math]{{x}_{j}}\,\![/math], on the remaining [math]k\,\![/math] -1 predictor variables. Mean values of [math]VIF\,\![/math] considerably greater than 1 indicate multicollinearity problems.
A few methods of dealing with multicollinearity include increasing the number of observations in a way designed to break up dependencies among predictor variables, combining the linearly dependent predictor variables into one variable, eliminating variables from the model that are unimportant or using coded variables.

#### Example

Variance inflation factors can be obtained for the data below.

To calculate the variance inflation factor for [math]{{x}_{1}}\,\![/math], [math]R_{1}^{2}\,\![/math] has to be calculated.[math]R_{1}^{2}\,\![/math] is the coefficient of determination for the model when [math]{{x}_{1}}\,\![/math] is regressed on the remaining variables. In the case of this example there is just one remaining variable which is [math]{{x}_{2}}\,\![/math]. If a regression model is fit to the data, taking [math]{{x}_{1}}\,\![/math] as the response variable and [math]{{x}_{2}}\,\![/math] as the predictor variable, then the design matrix and the vector of observations are:

- [math]{{X}_{{{R}_{1}}}}=\left[ \begin{matrix} 1 & 29.1 \\ 1 & 29.3 \\ . & . \\ . & . \\ . & . \\ 1 & 32.9 \\ \end{matrix} \right]\text{ }{{y}_{{{R}_{1}}}}=\left[ \begin{matrix} 41.9 \\ 43.4 \\ . \\ . \\ . \\ 77.8 \\ \end{matrix} \right]\,\![/math]

The regression sum of squares for this model can be obtained as:

- [math]\begin{align} S{{S}_{R}}= & y_{{{R}_{1}}}^{\prime }\left[ {{H}_{{{R}_{1}}}}-(\frac{1}{n})J \right]{{y}_{{{R}_{1}}}} \\ = & 1988.6 \end{align}\,\![/math]

where [math]{{H}_{{{R}_{1}}}}\,\![/math] is the hat matrix (and is calculated using [math]{{H}_{{{R}_{1}}}}={{X}_{{{R}_{1}}}}{{(X_{{{R}_{1}}}^{\prime }{{X}_{{{R}_{1}}}})}^{-1}}X_{{{R}_{1}}}^{\prime }\,\![/math] ) and [math]J\,\![/math] is the matrix of ones. The total sum of squares for the model can be calculated as:

- [math]\begin{align} S{{S}_{T}}= & {{y}^{\prime }}\left[ I-(\frac{1}{n})J \right]y \\ = & 2182.9 \end{align}\,\![/math]

where [math]I\,\![/math] is the identity matrix. Therefore:

- [math]\begin{align} R_{1}^{2}= & \frac{S{{S}_{R}}}{S{{S}_{T}}} \\ = & \frac{1988.6}{2182.9} \\ = & 0.911 \end{align}\,\![/math]

Then the variance inflation factor for [math]{{x}_{1}}\,\![/math] is:

- [math]\begin{align} VI{{F}_{1}}= & \frac{1}{(1-R_{1}^{2})} \\ = & \frac{1}{1-0.911} \\ = & 11.2 \end{align}\,\![/math]

The variance inflation factor for [math]{{x}_{2}}\,\![/math], [math]VI{{F}_{2}}\,\![/math], can be obtained in a similar manner. In the DOE folios, the variance inflation factors are displayed in the VIF column of the Regression Information table as shown in the following figure. Since the values of the variance inflation factors obtained are considerably greater than 1, multicollinearity is an issue for the data.