Distributed Machine Learning System: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
m (Text replacement - ". " to ". ")
m (Text replacement - " " to " ")
 
Line 26: Line 26:


=== 2018 ===
=== 2018 ===
* (Nishihara & Moritz, 2018) ⇒ [[Robert Nishihara]], and [[Philipp Moritz]] (Jan 9, 2018). [https://bair.berkeley.edu/blog/2018/01/09/ray/ "Ray: A Distributed System for AI"] Retrieved on 2019-04-14
* (Nishihara & Moritz, 2018) ⇒ [[Robert Nishihara]], and [[Philipp Moritz]] (Jan 9, 2018). [https://bair.berkeley.edu/blog/2018/01/09/ray/ "Ray: A Distributed System for AI"] Retrieved on 2019-04-14
** QUOTE: One of [[Ray]]’s goals is to enable practitioners to turn a [[prototype algorithm]] that runs on a [[laptop]] into a [[high-performance distributed application]] that runs efficiently on a [[cluster]] (or on a [[single multi-core machine]]) with relatively few additional lines of [[code]]. Such a framework should include the [[performance]] benefits of a [[hand-optimized system]] without requiring the user to reason about [[scheduling]], [[data transfer]]s, and [[machine failure]]s.        <P>        (...) There are two main ways of using [[Ray]]: through its [[lower-level API]]s and [[higher-level libraries]]. The [[higher-level libraries]] are built on top of the [[lower-level API]]s. Currently these include [[Ray RLlib]], a [[scalable reinforcement learning library]] and [[Ray.tune]], an efficient [[distributed hyperparameter search library]].
** QUOTE: One of [[Ray]]’s goals is to enable practitioners to turn a [[prototype algorithm]] that runs on a [[laptop]] into a [[high-performance distributed application]] that runs efficiently on a [[cluster]] (or on a [[single multi-core machine]]) with relatively few additional lines of [[code]]. Such a framework should include the [[performance]] benefits of a [[hand-optimized system]] without requiring the user to reason about [[scheduling]], [[data transfer]]s, and [[machine failure]]s.        <P>        (...) There are two main ways of using [[Ray]]: through its [[lower-level API]]s and [[higher-level libraries]]. The [[higher-level libraries]] are built on top of the [[lower-level API]]s. Currently these include [[Ray RLlib]], a [[scalable reinforcement learning library]] and [[Ray.tune]], an efficient [[distributed hyperparameter search library]].



Latest revision as of 07:55, 19 June 2023

A Distributed Machine Learning System is Distributed Computing System for implementing and developing machine learning algorithms.



References

2018

2016

2014