Règle delta généralisée


Révision datée du 13 septembre 2019 à 11:38 par Pitpitt (discussion | contributions) (Page créée avec « == en construction == Catégorie:Vocabulary Catégorie:Apprentissage machine Catégorie:UNSW == Définition == xxxxxxx == Français == xxxxxxx == Anglais... »)
(diff) ← Version précédente | Voir la version actuelle (diff) | Version suivante → (diff)

en construction


Définition

xxxxxxx

Français

xxxxxxx

Anglais

generalized delta rule An improvement on the in error backpropagation learning. If the learning rate (often denoted by η) is small, the backprop algorithm proceeds slowly, but accurately follows the path of steepest descent on the error surface. If η is too large, the algorithm may "bounce off the canyon walls of the error surface" - i.e. not work well. This can be largely avoided by modifying the delta rule to include a momentum term:

Δwji(n) = α Δwji(n–1) + η δj(n) yi(n)

in the notation of Haykin's text (Neural networks - a comprehensive foundation). The constant α is a termed the momentum constant and can be adjusted to achieve the best effect. The second summand corresponds to the standard delta rule, while the first summand says "add α × the previous change to this weight."

This new rule is called the generalized delta rule. The effect is that if the basic delta rule would be consistently pushing a weight in the same direction, then it gradually gathers "momentum" in that direction.


Source : INWS machine learning dictionary ]



Contributeurs: Imane Meziani, wiki