Eager Learning Algorithm

From GM-RKB
Jump to: navigation, search

An Eager Learning Algorithm is a learning algorithm that explores an entire training record set during a training phase to build a decision structure that it can exploit during the testing phase.



References

1997

  • (Mitchell, 1997) ⇒ Tom M. Mitchell. (1997). “Machine Learning." McGraw-Hill. . ISBN:0070428077
    • QUOTE: Section 8.6 Remarks on Lazy and Eager Learning: In this chapter we considered three lazy learning methods: the k-Nearest Neighbor algorithm, locally weighted regression, and case-based reasoning. We call these methods lazy because they defer the decision of how to generalize beyond the training data until each new query instance in encountered. We also discussed an eager learning method the method for learning radial basis function networks. We call this method eager because it generalize beyond the training data before observe the new query, committing at training time to the network structure and weights that define its approximation to the target function. In this same sense, every other algorithm discussed elsewhere in this book (e.g., Backpropagation, C4.5) is an eager learning algorithm. ... Lazy methods may consider the query instance xq when deciding how to generalize beyond the training data D. ... Eager methods cannot. By the time they observe the query instance xq they have already chosen their (global) approximation to the target function. ... The key point in the above paragraph is that a lazy learning has the option of (implicitly) representing the target function by a combination of many local approximations, whereas an eager learner must commit at training time to a single global approximation. The distinction between eager and lazy learning is thus related to the distinction between global and local approximations to the target function.