1999 LinearNeuralNetworks
		
		
		
		
		
		Jump to navigation
		Jump to search
		
		
	
- (Orr, 1999a) ⇒ Genevieve Orr. (1999). “Linear Neural Networks.” In: CS-449: Neural Networks" Fall 99
 
Subject Headings: Course, Presentation Slides, Gradient Descent Algorithm.
Notes
Cited By
Quotes
- To find the gradient G for the entire data set, we sum at each weight the contribution given by equation 6 over all the data points. We can then subtract a small proportion µ (called the learning rate) of G from the weights to perform gradient descent.
 - 1. Initialize all weights to small random values.
 - 2. REPEAT until done
- 1. For each weight wij set
 - 2. For each data point (x, t)p
- 1. set input units to x
 - 2. compute value of output units
 - 3. For each weight wij set
 
 
 - 3. For each weight wij set
 - An alternative approach is online learning, where the weights are updated immediately after seeing each data point. Since the gradient for a single data point can be considered a noisy approximation to the overall gradient G (Fig. 5), this is also called stochastic (noisy) gradient descent.
 
References
,
| Author | volume | Date Value | title | type | journal | titleUrl | doi | note | year | |
|---|---|---|---|---|---|---|---|---|---|---|
| 1999 LinearNeuralNetworks | Genevieve Orr | Linear Neural Networks | http://www.willamette.edu/~gorr/classes/cs449/linear2.html |