Error Rate Measure

From GM-RKB
Jump to navigation Jump to search

An Error Rate Measure is a statistical measure that quantifies the frequency or proportion of errors in a system, process, or statistical test relative to the total number of observations or decisions.



References

2023

  • (James et al., 2023) ⇒ Gareth James, Daniela Witten, Trevor Hastie, and Robert Tibshirani. (2023). "An Introduction to Statistical Learning." Springer.
    • QUOTE: The test error rate is the average error that results from using a statistical learning method to predict the response on a new observation—one that was not used in training the method. In contrast, the training error rate is computed using the observations that were used to fit the model.

2020

  • (Murphy, 2020) ⇒ Kevin P. Murphy. (2020). "Machine Learning: A Probabilistic Perspective." MIT Press.
    • QUOTE: The error rate is simply one minus the accuracy. For binary classification, the error rate equals the sum of false positive and false negative rates. This decomposition helps identify whether errors are primarily due to false alarms or missed detections.

2016

  • (Hastie et al., 2016) ⇒ Trevor Hastie, Robert Tibshirani, and Jerome Friedman. (2016). "The Elements of Statistical Learning." Springer.
    • QUOTE: The expected prediction error (generalization error) can be decomposed into irreducible error (noise), squared bias, and variance. This bias-variance tradeoff is fundamental to understanding how model complexity affects error rates.

2012

  • (Kohavi & Provost, 2012) ⇒ Ron Kohavi and Foster Provost. (2012). "Glossary of Terms." Machine Learning.
    • QUOTE: Error rate: The proportion of instances for which the system produces an incorrect output. For classification, this is the number of misclassified instances divided by the total number of instances. The complement of accuracy.

2006

  • (Japkowicz & Shah, 2006) ⇒ Nathalie Japkowicz and Mohak Shah. (2006). "Evaluating Learning Algorithms." Cambridge University Press.
    • QUOTE: Error rate alone can be misleading, particularly with imbalanced datasets. A classifier that always predicts the majority class will have low error rate but poor performance on the minority class. Multiple metrics are needed for comprehensive evaluation.

1997

  • (Efron & Tibshirani, 1997) ⇒ Bradley Efron and Robert Tibshirani. (1997). "An Introduction to the Bootstrap." Chapman & Hall.
    • QUOTE: The apparent error rate (resubstitution error) is typically over-optimistic because the same data is used for both model fitting and error estimation. Cross-validation and bootstrap methods provide more realistic error rate estimates.

1995

  • (Kohavi, 1995) ⇒ Ron Kohavi. (1995). "A Study of Cross-Validation and Bootstrap for Accuracy Estimation." IJCAI.
    • QUOTE: Error estimation is crucial for model selection and assessment. The true error rate is the expected disagreement between the classifier and the target function over the entire distribution of instances.