Random Subspace Method Algorithm: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
m (Text replacement - " [[" to " [[")
m (Text replacement - " [[" to " [[")
 
Line 29: Line 29:
=== 2017 ===
=== 2017 ===
* ([[Sammut & Webb, 2017]]) ⇒ [[Claude Sammut]], and [[Geoffrey I. Webb]]. ([[2017]]). [https://link.springer.com/referenceworkentry/10.1007/978-1-4899-7687-1_696 “Random Subspace Method”]. In: ([[Sammut & Webb, 2017]]). [https://doi.org/10.1007/978-1-4899-7687-1_696 DOI: 10.1007/978-1-4899-7687-1_696]
* ([[Sammut & Webb, 2017]]) ⇒ [[Claude Sammut]], and [[Geoffrey I. Webb]]. ([[2017]]). [https://link.springer.com/referenceworkentry/10.1007/978-1-4899-7687-1_696 “Random Subspace Method”]. In: ([[Sammut & Webb, 2017]]). [https://doi.org/10.1007/978-1-4899-7687-1_696 DOI: 10.1007/978-1-4899-7687-1_696]
** QUOTE: The [[Random Subspace Method Algorithm|random subspace method]] is an [[ensemble learning technique]]. The principle is to increase [[diversity]] between [[members of the ensemble]] by restricting [[classifier]]s to work on different [[random subset]]s of the full [[feature space]]. Each [[classifier]] [[learn]]s with a [[subset]] of [[size]] <i>n</i>, [[chosen uniformly at random]] from the full [[set]] of size <i>N</i>. [[Empirical studi]]es have suggested [[good result]]s can be obtained with the [[rule-of-thumb]] to choose <i>n = N∕2</i> [[feature]]s. The [[method]] is generally found to [[perform]] best when there are a large number of [[feature]]s (large N), and the [[discriminative information]] is spread across them. The [[method]] can [[underperform]] in the converse situation, when there are few [[informative feature]]s, and a large number of [[noisy]]/[[irrelevant feature]]s. [[Random Forest]]s is an [[algorithm]] combining [[Random Subspace Method Algorithm|RSM]] with the [[Bagging algorithm]], which can provide [[significant gain]]s over each used separately.
** QUOTE: The [[Random Subspace Method Algorithm|random subspace method]] is an [[ensemble learning technique]]. The principle is to increase [[diversity]] between [[members of the ensemble]] by restricting [[classifier]]s to work on different [[random subset]]s of the full [[feature space]]. Each [[classifier]] [[learn]]s with a [[subset]] of [[size]] <i>n</i>, [[chosen uniformly at random]] from the full [[set]] of size <i>N</i>. [[Empirical studi]]es have suggested [[good result]]s can be obtained with the [[rule-of-thumb]] to choose <i>n = N∕2</i> [[feature]]s. The [[method]] is generally found to [[perform]] best when there are a large number of [[feature]]s (large N), and the [[discriminative information]] is spread across them. The [[method]] can [[underperform]] in the converse situation, when there are few [[informative feature]]s, and a large number of [[noisy]]/[[irrelevant feature]]s. [[Random Forest]]s is an [[algorithm]] combining [[Random Subspace Method Algorithm|RSM]] with the [[Bagging algorithm]], which can provide [[significant gain]]s over each used separately.


=== 2009 ===
=== 2009 ===

Latest revision as of 17:07, 1 June 2024

A Random Subspace Method (RMS) Algorithm is an Ensemble Learning Algorithm that aims to increase an ensemble's diversity by restricting classifiers to work on different random subsets of the full feature space.



References

2019

  • (Wikipedia, 2019) ⇒ https://en.wikipedia.org/wiki/Random_subspace_method Retrieved:2019-8-18.
    • In machine learning the random subspace method,[1] also called attribute bagging or feature bagging, is an ensemble learning method that attempts to reduce the correlation between estimators in an ensemble by training them on random samples of features instead of the entire feature set (...)

      An ensemble of models employing the random subspace method can be constructed using the following algorithm:

      • Let the number of training points be N and the number of features in the training data be D.
      • Choose L to be the number of individual models in the ensemble.
      • For each individual model l, choose n<dub>l (nl < N) to be the number of input points for l. It is common to have only one value of nl for all the individual models.
      • For each individual model l, create a training set by choosing dl features from D with replacement and train the model.
Now, to apply the ensemble model to an unseen point, combine the outputs of the L individual models by majority voting or by combining the posterior probabilities.

2017

2009


  1. Ho, Tin Kam (1998). “The Random Subspace Method for Constructing Decision Forests" (PDF). IEEE Transactions on Pattern Analysis and Machine Intelligence. 20 (8): 832–844. doi:10.1109/34.709601.