Poster
in
Workshop: INTERPOLATE — First Workshop on Interpolation Regularizers and Beyond
Overparameterization Implicitly Regularizes Input-Space Smoothness
Matteo Gamba · Hossein Azizpour · Mårten Björkman
Keywords: [ Smoothness ] [ Double descent ] [ deep networks ] [ Lipschitz constant ]
Existing bounds on the generalization error of deep networks assume some form of smooth or bounded dependence on the input variable and intermediate activations, falling short of investigating the mechanisms controlling such factors in practice. In this work, we present an empirical study of the Lipschitz constant of networks trained in practice, as the number of model parameters and training epochs vary. We present non-monotonic trends for the Lipschitz constant, strongly correlating with double descent for the test error. Our findings highlight a theoretical shortcoming in modeling input-space smoothness via monotonic bounds.