Skip to yearly menu bar Skip to main content


Poster

ReLU Characteristic Activation Analysis

Wenlin Chen · Hong Ge

[ ]
Fri 13 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

We introduce a novel approach for analyzing the training dynamics of ReLU networks by examining the characteristic activation boundaries of individual ReLU neurons. Our proposed analysis reveals a critical instability in common neural network parameterizations and normalizations during stochastic optimization, which impedes fast convergence and hurts generalization performance. Addressing this, we propose Geometric Parameterization (GmP), a novel neural network parameterization technique that effectively separates the radial and angular components of weights in the hyperspherical coordinate system. We show theoretically that GmP resolves the aforementioned instability issue. We report empirical results on various models and benchmarks to verify GmP's theoretical advantages of optimization stability, convergence speed and generalization performance.

Live content is unavailable. Log in and register to view live content