Oxford University
Dropout With a Difference: Reduce neural net overfitting without impacting accuracy
The technique known as dropout discourages neural networks from overfitting by deterring them from reliance on particular features. A new approach reorganizes the process to run efficiently on the chips that typically run neural network calculations.