I'm training a DBN for classification with CD1 pretraining and Conjugate Gradient fine tuning. The CG implementation is based on Carl Edward Rasmussen code (http://learning.eng.cam.ac.uk/carl/code/minimize/minimize.m). It works quite well with sigmoid binary hidden units, but does not seem to work on Rectified Linear Hidden Unit.

Is there any reason why CG should not work on RLU hidden units ? Is there a better method for fine tuning a DBN with RLU hidden units ?

asked Jun 17 '14 at 02:30

Baptiste%20Wicht's gravatar image

Baptiste Wicht
31121315

Be the first one to answer this question!
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.