Accelerating Deep Learning by Focusing on the Biggest Losers
Abstract:
This paper introduces Selective-Backprop, a technique that accelerates the training of deep neural networks (DNNs) by prioritizing examples with high loss at each iteration. Selective-Backprop uses the output of a training example's forward pass to decide whether to use that example to compute gradients and update parameters, or to skip...More
Code:
Data:
Full Text
Tags
Comments