Revision history[back]
click to hide/show revision 1
Revision n. 1

Jun 30 '10 at 17:26

osdf's gravatar image

osdf
67031119

You might want to look at PGPE. While the cited paper presents the main idea in the context of Reinforcement Learning (particularly Policy Gradients), PGPE is by no means limited to that field. E.g. an ICANN-paper this year shows how it is used to break a cryptographic system (much faster than standard Evolutionary Search). You also might want to give Stochastic Search a try, though one might find out that with parameter sizes >> 50, PGPE is faster.

click to hide/show revision 2
Revision n. 2

Jul 01 '10 at 16:51

osdf's gravatar image

osdf
67031119

You might want to look at PGPE. While the cited paper presents the main idea in the context of Reinforcement Learning (particularly Policy Gradients), PGPE is by no means limited to that field. E.g. an ICANN-paper this year ICANN 2010 paper shows how it is used to break a cryptographic system (much faster than standard Evolutionary Search). You also might want to give Stochastic Search a try, though one might find out that with parameter sizes >> 50, PGPE is faster.

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.