Regression Tree Learning Algorithm: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
m (Text replacement - "]] Category" to " [[Category")
m (Text replacement - "<P> [[" to "<P>  [[")
Line 40: Line 40:
=== 1999 ===
=== 1999 ===
* ([[1999_InductiveLearningofTreebasedReg|Torgo, 1999]]) ⇒ [[Luis Torgo]]. ([[1999]]). “[http://www.dcc.fc.up.pt/~ltorgo/PhD/ Inductive Learning of Tree-based Regression Models]." Ph.D. Thesis, Thesis, Faculty of Sciences, University of Porto
* ([[1999_InductiveLearningofTreebasedReg|Torgo, 1999]]) ⇒ [[Luis Torgo]]. ([[1999]]). “[http://www.dcc.fc.up.pt/~ltorgo/PhD/ Inductive Learning of Tree-based Regression Models]." Ph.D. Thesis, Thesis, Faculty of Sciences, University of Porto
** QUOTE: [[Regression tree]]s are constructed using a [[recursive partitioning (RP) algorithm]]. This [[algorithm]] builds a [[Decision Treet|tree]] by [[recursively splitting]] the [[training sample]] into smaller [[subset]]s. We give below a [[high level description]] of the [[algorithm]]. The [[RP algorithm]] receives as [[input]] a [[set]] of <math>n</math> [[data point]]s, <math>D_t=\{\langle \mathbf{x_i},y_i\rangle\}^{n_t}_{i=1}</math>, and if certain [[termination criteria]] are not met it generates a [[test node]] <math>t</math>, whose [[Decision Treet Branch|branches]] are obtained by applying the same [[algorithm]] with two [[subset]]s of the [[input]] [[data point]]s. These [[subset]]s consist of the cases that logically entail the [[split test]] <math>s^*</math> in the [[node]] <math>t</math>, <math> D_{t_L}=\{\langle \mathbf{x_i},y_i\rangle\} \in D_t : \mathbf{x_i} \to s^*\} </math>, and the remaining cases, <math> D_{t_R}=\{\langle \mathbf{x_i},y_i\rangle\} \in D_t : \mathbf{x_i} \not \to s^*\}</math>. At each node the best [[split test]] is chosen according to some local [[criterion]], which means that this is a [[greedy hill-climbing algorithm]].        <P>       [[File:1999_InductiveLearningofTreebasedReg_Algorithm3.1.jpg]]<P>        The [[algorithm]] has three main components:
** QUOTE: [[Regression tree]]s are constructed using a [[recursive partitioning (RP) algorithm]]. This [[algorithm]] builds a [[Decision Treet|tree]] by [[recursively splitting]] the [[training sample]] into smaller [[subset]]s. We give below a [[high level description]] of the [[algorithm]]. The [[RP algorithm]] receives as [[input]] a [[set]] of <math>n</math> [[data point]]s, <math>D_t=\{\langle \mathbf{x_i},y_i\rangle\}^{n_t}_{i=1}</math>, and if certain [[termination criteria]] are not met it generates a [[test node]] <math>t</math>, whose [[Decision Treet Branch|branches]] are obtained by applying the same [[algorithm]] with two [[subset]]s of the [[input]] [[data point]]s. These [[subset]]s consist of the cases that logically entail the [[split test]] <math>s^*</math> in the [[node]] <math>t</math>, <math> D_{t_L}=\{\langle \mathbf{x_i},y_i\rangle\} \in D_t : \mathbf{x_i} \to s^*\} </math>, and the remaining cases, <math> D_{t_R}=\{\langle \mathbf{x_i},y_i\rangle\} \in D_t : \mathbf{x_i} \not \to s^*\}</math>. At each node the best [[split test]] is chosen according to some local [[criterion]], which means that this is a [[greedy hill-climbing algorithm]].        <P>         [[File:1999_InductiveLearningofTreebasedReg_Algorithm3.1.jpg]]<P>        The [[algorithm]] has three main components:
*** A way to select a [[split test]] (the [[splitting rule]]).
*** A way to select a [[split test]] (the [[splitting rule]]).
*** A [[rule]] to determine when a [[tree node]] is [[terminal]] ([[termination criterion]]).
*** A [[rule]] to determine when a [[tree node]] is [[terminal]] ([[termination criterion]]).

Revision as of 01:50, 27 February 2024

A Regression Tree Learning Algorithm is a decision tree learning algorithm that is a model-based regression algorithm.



References

2017a

  • (Wikipedia, 2017) ⇒ https://en.wikipedia.org/wiki/Decision_tree_learning Retrieved:2017-10-15.
    • Decision tree learning uses a decision tree (as a predictive model) to go from observations about an item (represented in the branches) to conclusions about the item's target value (represented in the leaves). It is one of the predictive modelling approaches used in statistics, data mining and machine learning. Tree models where the target variable can take a discrete set of values are called classification trees ; in these tree structures, leaves represent class labels and branches represent conjunctions of features that lead to those class labels. Decision trees where the target variable can take continuous values (typically real numbers) are called regression trees.

      In decision analysis, a decision tree can be used to visually and explicitly represent decisions and decision making. In data mining, a decision tree describes data (but the resulting classification tree can be an input for decision making). This page deals with decision trees in data mining.

2017b

2013

2006

1999

1997

1992

1984