Meet DiffGrad: New Deep Learning Optimizer that solves Adam’s ‘overshoot’ issue
5 min readDec 26, 2019
DiffGrad, a new optimizer introduced in the paper “diffGrad: An optimizer for CNN’s” by Dubey, et al, builds on the proven Adam optimizer by developing an adaptive ‘friction clamp’ and monitoring the local change in gradients in order to automatically lock in optimal parameter values that Adam can skip over.