2010 CompressedFisherLinearDiscrimin

From GM-RKB
Jump to navigation Jump to search

Subject Headings:

Notes

Cited By

Quotes

Author Keywords

Dimensionality Reduction, Random Projection, Compressed Learning, Linear Discriminant Analysis, Classification

Abstract

We consider random projections in conjunction with classification, specifically the analysis of Fisher's Linear Discriminant (FLD) classifier in randomly projected data spaces.

Unlike previous analyses of other classifiers in this setting, we avoid the unnatural effects that arise when one insists that all pairwise distances are approximately preserved under projection. We impose no sparsity or underlying low-dimensional structure constraints on the data; we instead take advantage of the class structure inherent in the problem. We obtain a reasonably tight upper bound on the estimated misclassification error on average over the random choice of the projection, which, in contrast to early distance preserving approaches, tightens in a natural way as the number of training examples increases. It follows that, for good generalisation of FLD, the required projection dimension grows logarithmically with the number of classes. We also show that the error contribution of a covariance misspecification is always no worse in the low-dimensional space than in the initial high-dimensional space. We contrast our findings to previous related work, and discuss our insights.

References

,

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2010 CompressedFisherLinearDiscriminRobert J. Durrant
Ata Kaban
Compressed Fisher Linear Discriminant Analysis: Classification of Randomly Projected DataKDD-2010 Proceedings10.1145/1835804.18359452010