Skip to yearly menu bar Skip to main content


Oral
in
Workshop: ML with New Compute Paradigms

Casting hybrid digital-analog training into hierarchical energy-based learning

Timothy Nest · Maxence Ernoult

[ ] [ Project Page ]
Sun 15 Dec 9:40 a.m. PST — 9:50 a.m. PST
 
presentation: ML with New Compute Paradigms
Sun 15 Dec 9 a.m. PST — 5 p.m. PST

Abstract:

Deep learning requires new approaches to combat the rising cost of training large models. The combination of energy-based analog circuits and the Equilibrium Propagation (EP) algorithm offers one compelling alternative to BP for gradient-based optimization of neural nets. In this work, we introduce a hybrid framework comprising feedforward (FF) and energy-based (EB) blocks housed on digital and analog circuits. We derive a novel algorithm to compute gradients end-to-end via BP \emph{and} EP, through FF and EB parts respectively, enabling EP to be applied to much more flexible and realistic architectures as analog units are incorporated into digital circuitry over time. We demonstrate the effectiveness of the proposed approach, showing that a standard Hopfield Network can be split into any shape while maintaining automatic differentiation performance. We apply it to ImageNet32 where we establish new SOTA in the EP and BP-alternative literature (46\% top-1).

Chat is not available.