our datset take close to a day to finish training, if we have 5x the data it won't work with our application, thats why we are trying to see if distributed training would help lower training time
but the training process would be slower than parallelization? is batch gradient descent the industry standard for handling large dataset in nn training?
IdeaEnough443 OP t1_izgyjq8 wrote
Reply to comment by SwordOfVarjo in [D] What is the recommended approach to training NN on big data set? by IdeaEnough443
our datset take close to a day to finish training, if we have 5x the data it won't work with our application, thats why we are trying to see if distributed training would help lower training time