Functional Error Correction for Reliable Neural Networks Conference Paper uri icon

abstract

  • When neural networks (NeuralNets) are implemented in hardware, their weights need to be stored in memory devices. As noise accumulates in the stored weights, the NeuralNets performance will degrade. This paper studies how to use error correcting codes (ECCs) to protect the weights. Different from classic error correction in data storage, the optimization objective is to optimize the NeuralNets performance after error correction, instead of minimizing the Uncorrectable Bit Error Rate in the protected bits. That is, by seeing the NeuralNet as a function of its input, the error correction scheme is function-oriented. A main challenge is that a deep NeuralNet often has millions to hundreds of millions of weights, causing a large redundancy overhead for ECCs, and the relationship between the weights and its NeuralNets performance can be highly complex. To address the challenge, we propose a Selective Protection (SP) scheme, which chooses only a subset of important bits for ECC protection. To find such bits and achieve an optimized tradeoff between ECCs redundancy and NeuralNets performance, we present an algorithm based on deep reinforcement learning. Experimental results verify that compared to the natural baseline scheme, the proposed algorithm achieves substantially better performance for the functional error correction task.

name of conference

  • 2020 IEEE International Symposium on Information Theory (ISIT)

published proceedings

  • 2020 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT)

author list (cited authors)

  • Huang, K., Siegel, P. H., & Jiang, A. A.

citation count

  • 4

complete list of authors

  • Huang, Kunping||Siegel, Paul H||Jiang, Anxiao Andrew

publication date

  • June 2020