Viewing a single comment thread. View all comments

jomobro117 t1_jdljb97 wrote

Thanks for sharing! Just a couple of questions. Is the evolutionary algorithm you use similar to PBT or fundamentally different in some way? And is there a plan to implement distributed training and HPO (similar to Ray RLlib with PBT from Tune)?

2

nicku_a OP t1_jdllrfv wrote

Hey! Yes, there are similarities to PBT, but there are a few differences here. Firstly, the mutations implemented with AgileRL are much more dynamic. Rather than only mutating hyperparameters, we’re allowing any part of the algorithm/model to mutate - HPs, network architecture (layers and nodes), activation functions and network weights themselves. We also train the population in one go, and offer efficient learning by sharing experience within the population.

1

nicku_a OP t1_jdlluja wrote

And yes the plan is to offer distributed training! As you can imagine there are about a million things we want/need to add! If you would like to get involved in the project and help out, please do

1