Exploding Gradient Problem: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
m (Text replacement - " " to " ")
m (Text replacement - "<P> [[" to "<P>  [[")
 
Line 17: Line 17:
=== 2021 ===
=== 2021 ===
* (DeepAI, 2021) ⇒ https://deepai.org/machine-learning-glossary-and-terms/exploding-gradient-problem Retrieved:2021-01-23.
* (DeepAI, 2021) ⇒ https://deepai.org/machine-learning-glossary-and-terms/exploding-gradient-problem Retrieved:2021-01-23.
** QUOTE: In [[machine learning]], the [[exploding gradient problem]] is an issue found in [[Neural Network Training System|training artificial neural network]]s with [[gradient-based learning method]]s and [[backpropagation]]. An [[artificial neural network]] is a [[learning algorithm]], also called [[ANN|neural network]] or [[neural net]], that uses a [[network]] of [[function]]s to understand and translate [[data input]] into a specific [[output]]. This type of [[learning algorithm]] is designed to mimic the way [[neuron]]s [[function]] in the [[human brain]]. [[Exploding gradient]]s are a [[problem]] when large [[error gradient]]s accumulate and result in very large [[update]]s to [[neural network model]] weights during [[NN Training System|training]]. [[Gradient]]s are used during [[NN Training System|training]] to [[update]] the [[network weight]]s, but when the typically this process works best when these [[update]]s are small and controlled. When the [[magnitude]]s of the [[gradient]]s accumulate, an [[unstable network]] is likely to occur, which can cause poor [[predicition]] results or even a [[model]] that reports nothing useful what so ever. There are methods to fix exploding [[gradient]]s, which include [[gradient clipping]] and [[weight regularization]], among others.        <P><div style="text-align:center">(...)</div>        <P>       [[Exploding gradient]]s can cause problems in the [[training of artificial neural network]]s. When there are [[exploding gradient]]s, an [[unstable network]] can result and the [[learning]] cannot be completed. The values of the [[NN Weight|weight]]s can also become so large as to overflow and result in something called [[NaN value]]s. [[NaN value]]s, which stands for not a number, are values that represent an undefined or unrepresentable values. It is useful to know how to identify [[exploding gradient]]s in order to correct the [[NN Training System|training]].
** QUOTE: In [[machine learning]], the [[exploding gradient problem]] is an issue found in [[Neural Network Training System|training artificial neural network]]s with [[gradient-based learning method]]s and [[backpropagation]]. An [[artificial neural network]] is a [[learning algorithm]], also called [[ANN|neural network]] or [[neural net]], that uses a [[network]] of [[function]]s to understand and translate [[data input]] into a specific [[output]]. This type of [[learning algorithm]] is designed to mimic the way [[neuron]]s [[function]] in the [[human brain]]. [[Exploding gradient]]s are a [[problem]] when large [[error gradient]]s accumulate and result in very large [[update]]s to [[neural network model]] weights during [[NN Training System|training]]. [[Gradient]]s are used during [[NN Training System|training]] to [[update]] the [[network weight]]s, but when the typically this process works best when these [[update]]s are small and controlled. When the [[magnitude]]s of the [[gradient]]s accumulate, an [[unstable network]] is likely to occur, which can cause poor [[predicition]] results or even a [[model]] that reports nothing useful what so ever. There are methods to fix exploding [[gradient]]s, which include [[gradient clipping]] and [[weight regularization]], among others.        <P><div style="text-align:center">(...)</div>        <P>         [[Exploding gradient]]s can cause problems in the [[training of artificial neural network]]s. When there are [[exploding gradient]]s, an [[unstable network]] can result and the [[learning]] cannot be completed. The values of the [[NN Weight|weight]]s can also become so large as to overflow and result in something called [[NaN value]]s. [[NaN value]]s, which stands for not a number, are values that represent an undefined or unrepresentable values. It is useful to know how to identify [[exploding gradient]]s in order to correct the [[NN Training System|training]].


=== 2017 ===
=== 2017 ===

Latest revision as of 01:50, 27 February 2024

An Exploding Gradient Problem is a Neural Network Training Algorithm problem that arises when using gradient descent and backpropagation.



References

2021

2017

Grosse2017 RNN.png
Figure 3 shows the function computed at each time step, as well as the function computed by the network as a whole. From this figure, you can see which regions have exploding or vanishing gradients.

Grosse2017 FIG3.png
Figure 3: (left) The function computed by the RNN at each time step, (right) the function computed by the network.