[link]
Summary by Léo Paillier 7 years ago
_Objective:_ Theoretical study of Deep Neural Network, their expressivity and regularizations.
## Results:
The key findings of the article are:
### A. Deep neural networks easily fit random labels.
Both when randomizing labels, replacing images with raw noise or all situations in-between.
1. The effective capacity of neural networks is sufficient for memorizing the entire data set.
2. Even optimization on random labels remains easy. In fact, training time increases only by a small constant factor compared with training on the true labels.
3. Randomizing labels is solely a data transformation, leaving all other properties of the learning problem unchanged.
### B. Explicit regularization may improve generalization performance, but is neither necessary nor by itself sufficient for controlling generalization error.
By explicit regularization they mean batch normalisation, weight decay, dropout, data augmentation, etc.
### C. Generically large neural networks can express any labeling of the training data.
More formally, a very simple two-layer ReLU network with `p = 2n + d` parameters can express any labeling of any sample of size `n` in `d` dimensions.
### D. The optimization algorithm itself is implicitly regularizing the solution.
SGD acts as an implicit regularizer and properties are inherited by models that were trained using SGD.

more
less