Poster
in
Workshop: ML with New Compute Paradigms
Improving Analog Neural Network Robustness: A Noise-Agnostic Approach with Explainable Regularizations
Alice Duque · Pedro Freire · Egor Manuylovich · Sergei K. Turitsyn · Jaroslaw Prilepsky · Dmitrii Stoliarov
This work tackles the critical challenge of mitigating "hardware noise" in deep analog neural networks, a major obstacle in advancing analog signal processing devices. We propose a comprehensive, hardware-agnostic solution to address both correlated and uncorrelated noise affecting the activation layers of deep neural models. The novelty of our approach lies in its ability to demystify the "black box" nature of noise-resilient networks by revealing the underlying mechanisms that reduce sensitivity to noise. In doing so, we introduce a new explainable regularization framework that harnesses these mechanisms to significantly enhance noise robustness in deep neural architectures, obtaining over 53\% accuracy improvement in noisy environments, when compared to models with standard training.