Made the check in dnn_trainer for convergence more robust. Previously, if we
encountered a bad mini-batch that made the loss value suddenly jump up by a larger than normal value it could make the trainer think we converged. Now the test is robust to recent spikes in loss value.
Showing
Please
register
or
sign in
to comment