PDF4PRO ⚡AMP

Modern search engine that looking for books and documents around the web

Example: dental hygienist

Understanding the difficulty of training deep feedforward ...

249 Understanding the difficulty of training deep feedforward neural networksXavier GlorotYoshua BengioDIRO, Universit e de Montr eal, Montr eal, Qu ebec, CanadaAbstractWhereas before 2006 it appears that deep multi-layer neural networks were not successfullytrained, since then several algorithms have beenshown to successfully train them, with experi-mental results showing the superiority of deepervs less deep architectures. All these experimen-tal results were obtained with new initializationor training mechanisms. Our objective here is tounderstand better why standard gradient descentfrom random initialization is doing so poorlywith deep neural networks, to better understandthese recent relative successes and help designbetter algorithms in the future. We first observethe influence of the non-linear activations func-tions. We find that the logistic sigmoid activationis unsuited for deep networks with random ini-tialization because of its mean value, which candrive especially the top hidden layer into satu-ration.

deep networks with sigmoids but initialized from unsuper-vised pre-training (e.g. from RBMs) do not suffer from this saturation behavior. Our proposed explanation rests on the hypothesis that the transformation that the lower layers of the randomly initialized network computes initially is

Tags:

  Network, Deep, Deep networks

Information

Domain:

Source:

Link to this page:

Please notify us if you found a problem with this document:

Spam in document Broken preview Other abuse

Transcription of Understanding the difficulty of training deep feedforward ...

Related search queries