Bayesian multiscale smoothing in supervised and semi-supervised kernel discriminant analysis
In kernel discriminant analysis, it is common practice to select the smoothing parameter (bandwidth) based on the training data and use it for classifying all unlabeled observations. But this method of selecting a single scale of smoothing ignores the major issue of model uncertainty. Moreover, in addition to depending on the training sample, a good choice of bandwidth may also depend on the observation to be classified, and a fixed level of smoothing may not work well in all parts of the measurement space. So, instead of using a single smoothing parameter, it may be more useful in practice to study classification results for multiple scales of smoothing and judiciously aggregate them to arrive at the final decision. This paper adopts a Bayesian approach to carry out one such multiscale analysis using a probabilistic framework. This framework also helps us to extend our multiscale method for semi-supervised classification, where, in addition to the training sample, one uses unlabeled test set observations to form the decision rule. Some well-known benchmark data sets are analyzed to show the utility of these proposed methods.
Year of publication: |
2011
|
---|---|
Authors: | Mukhopadhyay, Subhadeep ; Ghosh, Anil K. |
Published in: |
Computational Statistics & Data Analysis. - Elsevier, ISSN 0167-9473. - Vol. 55.2011, 7, p. 2344-2353
|
Publisher: |
Elsevier |
Keywords: | Bayes risk Gibbs sampling Kernel density estimation Misclassification rate Markov chain Monte Carlo Non-informative prior Transductive learning |
Saved in:
Online Resource
Saved in favorites
Similar items by person
-
Nonlinear time series modeling: A unified perspective, algorithm and application
Mukhopadhyay, Subhadeep, (2018)
-
Nonparametric universal copula modeling
Mukhopadhyay, Subhadeep, (2020)
-
Applied MANOVA and Discriminant Analysis (2nd ed.). Carl J. Huberty and Stephen Olejnik
Ghosh, Anil K., (2007)
- More ...