Gradient Descent-based Learning Algorithm: Difference between revisions
Jump to navigation
Jump to search
(Created page with "A Gradient Descent-based Learning Algorithm is a supervised learning algorithm that is a gradient descent algorithm. * <B>Counter-Example(s):</B> ** Convex Optim...") |
m (Text replacement - "QUOTE: " to "QUOTE: ") |
||
(25 intermediate revisions by 4 users not shown) | |||
Line 1: | Line 1: | ||
A [[Gradient Descent-based Learning Algorithm]] is a [[supervised learning algorithm]] that is a [[gradient descent algorithm]]. | A [[Gradient Descent-based Learning Algorithm]] is a [[supervised learning algorithm]] that is a [[gradient-descent optimization algorithm]]. | ||
* <B>Context:</B> | |||
** … | |||
* <B>Example(s):</B> | |||
** an [[Accelerated Gradient Descent (AGD)]], | |||
** an [[Adaptive Gradient Algorithm (AdaGrad)]], | |||
** an [[Adaptive Learning Rate Algorithm (ADADELTA)]], | |||
** an [[Adaptive Moment Estimation Algorithm (Adam)]], | |||
** a [[Gradient-Descent Boosted Learning Algorithm]]. | |||
** a [[Mini-Batch Gradient Descent Algorithm (MBGD)]]. | |||
** a [[Momentum Gradient Descent (MGD)]], | |||
** a [[Root Mean Square Propagation Algorithm (RMSprop)]]; | |||
** a [[Stochastic Gradient Descent Algorithm]]. | |||
** … | |||
* <B>Counter-Example(s):</B> | * <B>Counter-Example(s):</B> | ||
** [[Convex Optimization-based Learning Algorithm]]. | ** [[Convex Optimization-based Learning Algorithm]]. | ||
* <B>See:</B> [[Perceptron Algorithm]], [[Gradient Descent-based Learning System]]. | * <B>See:</B> [[Perceptron Algorithm]], [[Gradient Descent-based Learning System]]. | ||
---- | ---- | ||
---- | ---- | ||
===1998=== | == References == | ||
* ([[1998_GradientbasedLearningAppliedtoD|LeCun | |||
=== 2018 === | |||
* (Wijaya et al., 2018) ⇒ Galih Praja Wijaya, Dendi Handian, Imam Fachmi Nasrulloh, Lala Septem Riza, Rani Megasari, Enjun Junaeti (2018), [https://cran.r-project.org/web/packages/gradDescent/index.html "gradDescent: Gradient Descent for Regression Tasks"], [https://cran.r-project.org/web/packages/gradDescent/gradDescent.pdf "Reference manual (PDF)]. | |||
** QUOTE: An implementation of various [[learning algorithm]]s based on [[Gradient Descent]] for dealing with [[regression task]]s. The variants of [[gradient descent algorithm]] are: [[Mini-Batch Gradient Descent (MBGD)]], which is an [[optimization]] to use [[training data]] partially to reduce the [[computation load]]. [[Stochastic Gradient Descent (SGD)]], which is an [[optimization]] to use a [[random data]] in [[learning]] to reduce the computation load drastically. [[Stochastic Average Gradient (SAG)]], which is a [[SGD-based algorithm]] to minimize [[stochastic step]] to [[average]]. [[Momentum Gradient Descent (MGD)]], which is an optimization to speed-up [[Gradient Descent-based Learning Algorithm|gradient descent learning]]. [[Accelerated Gradient Descent (AGD)]], which is an [[optimization]] to accelerate [[Gradient Descent-based Learning Algorithm|gradient descent learning]]. [[Adagrad]], which is a [[gradient-descent-based algorithm]] that accumulate previous [[cost]] to do [[adaptive learning]]. [[Adadelta]], which is a [[gradient-descent-based algorithm]] that use [[hessian approximation]] to do [[adaptive learning]]. [[RMSprop]], which is a [[gradient-descent-based algorithm]] that combine [[Adagrad]] and [[Adadelta]] [[adaptive learning]] ability. [[Adam]], which is a [[gradient-descent-based algorithm]] that [[mean]] and [[variance moment]] to do [[adaptive learning]]. [[Stochastic Variance Reduce Gradient (SVRG)]], which is an [[optimization]] [[SGD-based algorithm]] to accelerates the process toward converging by reducing the gradient. [[Semi Stochastic Gradient Descent (SSGD)]],which is a [[SGD-based algorithm]] that combine [[GD]] and [[SGD]] to accelerates the process toward converging by choosing one of the gradients at a time. [[Stochastic Recursive Gradient Algorithm (SARAH)]], which is an optimization algorithm similarly [[SVRG]] to accelerates the process toward converging by accumulated stochastic information. [[Stochastic Recursive Gradient Algorithm+ (SARAHPlus)]], which is a [[SARAH]] practical variant algorithm to accelerates the process toward converging provides a possibility of earlier termination. | |||
=== 1998 === | |||
* ([[1998_GradientbasedLearningAppliedtoD|LeCun et al., 1998]]) ⇒ [[Yann LeCun]], [[Léon Bottou]], [[Yoshua Bengio]], and [[Patrick Haffner]]. ([[1998]]). “[http://yann.lecun.com/exdb/publis/pdf/lecun-01a.pdf Gradient-based Learning Applied to Document Recognition]." [http://dx.doi.org/10.1109/5.726791 doi:10.1109/5.726791] | |||
=== 1990 === | === 1990 === | ||
* Williams, Ronald J., and Jing Peng. | * ([[Williams et al., 1990]]) ⇒ [[Ronald J. Williams]], and [[Jing Peng]]. ([[1990]]). “An Efficient Gradient-based Algorithm for on-line Training of Recurrent Network Trajectories." Neural computation 2, no. 4 </s> | ||
---- | ---- | ||
__NOTOC__ | __NOTOC__ | ||
[[Category:Concept]] |
Latest revision as of 20:45, 29 December 2022
A Gradient Descent-based Learning Algorithm is a supervised learning algorithm that is a gradient-descent optimization algorithm.
- Context:
- …
- Example(s):
- an Accelerated Gradient Descent (AGD),
- an Adaptive Gradient Algorithm (AdaGrad),
- an Adaptive Learning Rate Algorithm (ADADELTA),
- an Adaptive Moment Estimation Algorithm (Adam),
- a Gradient-Descent Boosted Learning Algorithm.
- a Mini-Batch Gradient Descent Algorithm (MBGD).
- a Momentum Gradient Descent (MGD),
- a Root Mean Square Propagation Algorithm (RMSprop);
- a Stochastic Gradient Descent Algorithm.
- …
- Counter-Example(s):
- See: Perceptron Algorithm, Gradient Descent-based Learning System.
References
2018
- (Wijaya et al., 2018) ⇒ Galih Praja Wijaya, Dendi Handian, Imam Fachmi Nasrulloh, Lala Septem Riza, Rani Megasari, Enjun Junaeti (2018), "gradDescent: Gradient Descent for Regression Tasks", "Reference manual (PDF).
- QUOTE: An implementation of various learning algorithms based on Gradient Descent for dealing with regression tasks. The variants of gradient descent algorithm are: Mini-Batch Gradient Descent (MBGD), which is an optimization to use training data partially to reduce the computation load. Stochastic Gradient Descent (SGD), which is an optimization to use a random data in learning to reduce the computation load drastically. Stochastic Average Gradient (SAG), which is a SGD-based algorithm to minimize stochastic step to average. Momentum Gradient Descent (MGD), which is an optimization to speed-up gradient descent learning. Accelerated Gradient Descent (AGD), which is an optimization to accelerate gradient descent learning. Adagrad, which is a gradient-descent-based algorithm that accumulate previous cost to do adaptive learning. Adadelta, which is a gradient-descent-based algorithm that use hessian approximation to do adaptive learning. RMSprop, which is a gradient-descent-based algorithm that combine Adagrad and Adadelta adaptive learning ability. Adam, which is a gradient-descent-based algorithm that mean and variance moment to do adaptive learning. Stochastic Variance Reduce Gradient (SVRG), which is an optimization SGD-based algorithm to accelerates the process toward converging by reducing the gradient. Semi Stochastic Gradient Descent (SSGD),which is a SGD-based algorithm that combine GD and SGD to accelerates the process toward converging by choosing one of the gradients at a time. Stochastic Recursive Gradient Algorithm (SARAH), which is an optimization algorithm similarly SVRG to accelerates the process toward converging by accumulated stochastic information. Stochastic Recursive Gradient Algorithm+ (SARAHPlus), which is a SARAH practical variant algorithm to accelerates the process toward converging provides a possibility of earlier termination.
1998
- (LeCun et al., 1998) ⇒ Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. (1998). “Gradient-based Learning Applied to Document Recognition." doi:10.1109/5.726791
1990
- (Williams et al., 1990) ⇒ Ronald J. Williams, and Jing Peng. (1990). “An Efficient Gradient-based Algorithm for on-line Training of Recurrent Network Trajectories." Neural computation 2, no. 4