Distributed Machine Learning System: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
m (Text replacement - "\<P\>([\s]{1,7})([^\s])" to "<P> $2")
m (Text replacement - ". " to ". ")
Line 1: Line 1:
A [[Distributed Machine Learning System]] is [[Distributed Computing System]] for implementing and developing [[machine learning algorithm]]s.
A [[Distributed Machine Learning System]] is [[Distributed Computing System]] for implementing and developing [[machine learning algorithm]]s.
* <B>Context:</B>
* <B>Context:</B>
** It can range from being a standard [[Distributed ML System]], to being a [[Distributed Reinforcement Learning System]], to being a [[Distributed Deep Learning System]].  
** It can range from being a standard [[Distributed ML System]], to being a [[Distributed Reinforcement Learning System]], to being a [[Distributed Deep Learning System]].
* <B>Example(s):</B>
* <B>Example(s):</B>
** [[MLlib]],
** [[MLlib]],
Line 34: Line 34:


=== 2014 ===
=== 2014 ===
* ([[2014_AReliableEffectiveTerascaleLine|Agarwal et al., 2014]]) ⇒ [[Alekh Agarwal]], [[Olivier Chapelle]], [[Miroslav Dudík]], and [[John Langford]]. ([[2014]]). &ldquo;[http://www.jmlr.org/papers/volume15/agarwal14a/agarwal14a.pdf A Reliable Effective Terascale Linear Learning System].&rdquo; In: The Journal of Machine Learning Research, 15(1).  
* ([[2014_AReliableEffectiveTerascaleLine|Agarwal et al., 2014]]) ⇒ [[Alekh Agarwal]], [[Olivier Chapelle]], [[Miroslav Dudík]], and [[John Langford]]. ([[2014]]). &ldquo;[http://www.jmlr.org/papers/volume15/agarwal14a/agarwal14a.pdf A Reliable Effective Terascale Linear Learning System].&rdquo; In: The Journal of Machine Learning Research, 15(1).
** QUOTE: Perhaps the simplest [[strategy]] when the number of [[example]]s n is too large for a given [[learning algorithm]] is to reduce the data set size by [[subsampling]]. However, this [[strategy]] only works if the [[problem]] is simple enough or the number of [[parameter]]s is very small. The [[setting]] of interest here is when a large number of [[example]]s is really needed to [[learn]] a [[good model]]. [[Distributed algorithm]]s are a natural choice for such scenarios.
** QUOTE: Perhaps the simplest [[strategy]] when the number of [[example]]s n is too large for a given [[learning algorithm]] is to reduce the data set size by [[subsampling]]. However, this [[strategy]] only works if the [[problem]] is simple enough or the number of [[parameter]]s is very small. The [[setting]] of interest here is when a large number of [[example]]s is really needed to [[learn]] a [[good model]]. [[Distributed algorithm]]s are a natural choice for such scenarios.



Revision as of 18:52, 1 August 2022

A Distributed Machine Learning System is Distributed Computing System for implementing and developing machine learning algorithms.



References

2018

2016

2014