Definition:

  • Linear regression but in multiple dimensions
  • With independent observations on and the associated values of , the complete model becomes:
  • Error term, :
    • (constant)
    • for
  • Then we have where:
    • is design matrix
  • We introduce the artificial variable so that

Least square estimation:

  • We must determine the values for the regression coefficients and the error variance consistent with the available data.
  • Let be trial values for . The method of least squares selects so as to minimize the sum of the squares of the differences:
  • The coefficients chosen by the least squares criterion are called least squares estimates of , denoted by
  • The deviations are called residuals. The vector of residuals contains the information about the remaining unknown parameter
  • Proposition:
    • Let have full rank . The least squares estimate of is given by
    • Let denote the fitted values of , where is called hat matrix.
    • Then the residuals satisfy and .
    • The residual sum of squares
  • The coefficient of determination gives the proportion of the total variation in the ‘s explained by, or attributable to, the predictor variables .
    • Here equals 1 if the fitted equation passes through all the data points, so that for all .
    • In addition, if and . The predictor variables have no influence on the response.
  • The least squares estimator has
    • The residuals have the properties
    • Also, . Defining
    • We have
    • Moreover, and are uncorrelated.

Inferences about the regression model:

  • Let where has full rank and is distributed as . Then the maximum likelihood estimator of is the same as the least squares estimator .
    • Moreover, and is distributed independently of the residuals .
    • Further, where is the maximum likelihood estimator of
  • Let , where has full rank and is . Then a confidence region for is given by
    • Also, simultaneous confidence intervals for the are given by where is the diagonal element of corresponding to .
  • The confidence ellipsoid is centered at the maximum likelihood estimate and its orientation and size are determined by the eigenvalues and eigenvectors of .
    • If an eigenvalue is nearly zero, the confidence ellipsoid will be very long in the direction of the corresponding eigenvector.
    • Practitioners often use the intervals when searching for important predictor variables.

Likelihood ratio tests for regression parameters:

  • Part of regression analysis is concerned with assessing the effects of particular predictor variables on the response variable. One null hypothesis of interest states that certain of the ‘s do not influence the response .
  • These predictors will be labeled . The statement that do not influence translates into the statistical hypothesis where .
  • Setting
    • we can express the general linear model as
  • Define extra sum of squares to be
    • where
  • Let have full rank and be distributed as The likelihood ratio test rejects if

Inferences from the estimated regression function

  • Once an investigator is satisfied with the fitted regression model, it can be used to solve two prediction problems.
  • Let be selected values for the predictor variables.
  • Let denote the value of the response when the predictor variables have values . According to the classical linear regression model,
  • Its least squares estimate is
  • is the unbiased linear estimator of with minimum variance
    • If the errors are normally distributed, then a confidence interval for is
  • Prediction of a new observation, such as at , is more uncertain than estimating the expected value of .
    • where and is independent of and, hence, of and .
  • A new observation has the unbiased predictor
    • The variance of the forecast error is
    • When the errors have a normal distribution, a prediction interval for is given by
  • The prediction interval for is wider than the confidence interval for estimating the value of the regression function .
    • The additional uncertainty in forecasting , which is represented by the extra term in the expression , comes from the presence of the unknown error term .