Least Squares
Least squares linear and quadratic models create a polynomial that approximates the data sample globally as a best-fit.
For example, for one design study containing two input parameters and , a linear polynomial for predicted value including constant and linear terms is defined as:
A quadratic polynomial including constant, linear terms, interaction, and squared terms is defined as:
The fit quality of a polynomial to a data point is measured by its residual defined as the difference between the actual value and the predicted value :
In matrix notation, the equation is:
Here, represents the design matrix. For example, the design matrix of a quadratic model with two input parameters and from m simulations is:
The vector of the unknown coefficients is:
Best possible fit means that the residual sum of squares is minimized for the sample data:
In the form of linear approximation:
After expanding the linear approximation:
Due to the dimensions of the vector , , and the matrix , and are two scalars. Therefore, . The equation can be re-written as:
The minimum of the sum of squares is found when the gradient of linear approximation is zero:
Least squares surrogates usually do not pass through the sample points directly. They go beside the sample points in the design space to obtain a best global fit. The least squares solution can be computed using the singular value decomposition of the matrix . This method is efficient with a large number of sample data.
Number of Evaluations
To determine the vector of unknown coefficients, you are advised to run twice as many simulations as the dimension of the vector .
- For a linear model, the dimension of
is n+1. n
indicates the number of input parameters in the design study.(5162)
- For a quadratic model, the dimension of
is 1+2n+n(n-1)/2.
n indicates the number of input parameters in the design study.(5163)