Submitted by nicku_a t3_120h120 in MachineLearning
nicku_a OP t1_jdllrfv wrote
Reply to comment by jomobro117 in [P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up by nicku_a
Hey! Yes, there are similarities to PBT, but there are a few differences here. Firstly, the mutations implemented with AgileRL are much more dynamic. Rather than only mutating hyperparameters, we’re allowing any part of the algorithm/model to mutate - HPs, network architecture (layers and nodes), activation functions and network weights themselves. We also train the population in one go, and offer efficient learning by sharing experience within the population.
Viewing a single comment thread. View all comments