Partial inverse regression
In regression with a vector of quantitative predictors, sufficient dimension reduction methods can effectively reduce the predictor dimension, while preserving full regression information and assuming no parametric model. However, all current reduction methods require the sample size n to be greater than the number of predictors p. It is well known that partial least squares can deal with problems with n < p. We first establish a link between partial least squares and sufficient dimension reduction. Motivated by this link, we then propose a new dimension reduction method, entitled partial inverse regression. We show that its sample estimator is consistent, and that its performance is similar to or superior to partial least squares when n < p, especially when the regression model is nonlinear or heteroscedastic. An example involving the spectroscopy analysis of biscuit dough is also given. Copyright 2007, Oxford University Press.
Year of publication: |
2007
|
---|---|
Authors: | Li, Lexin ; Cook, R. Dennis ; Tsai, Chih-Ling |
Published in: |
Biometrika. - Biometrika Trust, ISSN 0006-3444. - Vol. 94.2007, 3, p. 615-625
|
Publisher: |
Biometrika Trust |
Saved in:
Saved in favorites
Similar items by person
-
Li, Lexin, (2005)
-
Longitudinal data model selection
Azari, Rahman, (2006)
-
Tobit Model Estimation and Sliced Inverse Regression
Li, Lexin, (2008)
- More ...