Linear Partial Least Squares Fit
This function computes the Partial Least Squares solution and the first derivative of the regression coefficients. This implementation scales mostly in the number of variables
linear.pls.fit( X, y, m = ncol(X), compute.jacobian = FALSE, DoF.max = min(ncol(X) + 1, nrow(X) - 1) )
X |
matrix of predictor observations. |
y |
vector of response observations. The length of |
m |
maximal number of Partial Least Squares components. Default is
|
compute.jacobian |
Should the first derivative of the regression
coefficients be computed as well? Default is |
DoF.max |
upper bound on the Degrees of Freedom. Default is
|
We first standardize X
to zero mean and unit variance.
coefficients |
matrix of regression coefficients |
intercept |
vector of regression intercepts |
DoF |
Degrees of Freedom |
sigmahat |
vector of estimated model error |
Yhat |
matrix of fitted values |
yhat |
vector of squared length of fitted values |
RSS |
vector of residual sum of error |
covariance
if
compute.jacobian
is TRUE
, the function returns the array of
covariance matrices for the PLS regression coefficients.
TT |
matrix of normalized PLS components |
Nicole Kraemer
Kraemer, N., Sugiyama M. (2011). "The Degrees of Freedom of Partial Least Squares Regression". Journal of the American Statistical Association 106 (494) https://www.tandfonline.com/doi/abs/10.1198/jasa.2011.tm10107
n<-50 # number of observations p<-5 # number of variables X<-matrix(rnorm(n*p),ncol=p) y<-rnorm(n) pls.object<-linear.pls.fit(X,y,m=5,compute.jacobian=TRUE)
Please choose more modern alternatives, such as Google Chrome or Mozilla Firefox.