2011 InternetScaleDataAnalysis

From GM-RKB
Jump to navigation Jump to search

Subject Headings:

Notes

Cited By

Quotes

Author Keywords

Abstract

This talk covers techniques for analyzing data sets with up to trillions of examples with billions of features, using thousands of computers. To operate at this scale requires an understanding of an increasing complex hardware hierarchy (e.g. cache, memory, SSD, another machine in the rack, disk, a machine in another data center, ...); a model for recovering from inevitable hardware and software failures; a machine learning model that allows for efficient computation over large, continuously updated data sets; and a way to visualize and share the results.

References

,

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2011 InternetScaleDataAnalysisPeter NorvigInternet Scale Data Analysis10.1145/2020408.2020412