I am currently trying to come up with a novel structure for a CLDNN (Convolutional, LSTM, Deep Neural Network)
Just as any other networks, I am having a difficult time optimizing the hyper parameters.
I would like to try grid search and random search to get an optimal set of hyperparameters but I am not clear on few things.
If I run a simulation of the network with a temporary set of hyperparameters, how do I measure "goodness" of the hyperparameters? I was thinking about recording the cost and training accuracy after N number of epochs for each simulations.
Since each simulation takes relatively long time (for my network it takes about 70 seconds to train for one epoch), is there a faster way to check the "goodness" of the hyperparameters without actually running the full training?
Is there a general tip/advice for hyperparameter-optimization?