Skip to content

Instantly share code, notes, and snippets.

@sebstud
Created May 6, 2014 09:02
Show Gist options
  • Save sebstud/e723c1042413c33f2c69 to your computer and use it in GitHub Desktop.
Save sebstud/e723c1042413c33f2c69 to your computer and use it in GitHub Desktop.
Collection export small problem #50
@Article{Jong2001,
Title = {Canonical partial least squares and continuum power regression},
Author = {Jong, Sijmen De and Wise, Barry M and Ricker, N Lawrence},
Journal = {Journal of Chemometrics},
Pages = {85-100},
Volume = {15},
Year = {2001},
Abstractnote = {A method, canonical PLS, is proposed for performing the basic PLS calculations in the canonical co-ordinate system of the predictor matrix X. This reduces the size of the problem to its smallest possible dimension as determined by the rank of X. The computation is further simplified since the cross-product matrices XTX and XXT are symmetric. PLS weights, scores and loadings referring to the canonical co-ordinate system can be easily back-transformed to the original co-ordinate system. The method offers an ideal setting to carry out the continuum regression approach to PLS introduced by Wise and Ricker. By raising the singular values to some power gamma, one may artificially decrease (gamma < 1) or increase (gamma > 1) the degree of multicollinearity in the X data. One may investigate a series of models by considering various values of the power gamma. This offers a means to push the model into the direction of ordinary least squares (gamma = 0) or principal components regression (gammararrinfin), with PLS regression as an intermediate case (gamma = 1). Since all these computations are mainly performed in canonical space, obtained after one singular value decomposition, a considerable gain in speed is achieved. This is demonstrated over a wide range of data set sizes (number of X and Y variables, number of samples) and model parameters (number of latent variables and number of powers considered). The gains in computational efficiency (as measured by the ratio of the number of floating point operations required relative to the original algorithm) range from a factor of 39 to over 100. Copyright 2000 John Wiley & Sons, Ltd.},
Number = {2},
Url = {http://www3.interscience.wiley.com/cgi-bin/abstract/75503122/START}
}
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment