Compute-in-Memory Upside Down: A Learning Operator Co-Design Perspective for Scalability

Shamma Nasrin, Priyesh Shukla, Shruthi Jaisimha and Amit Ranjan Trivedi
Department of Electrical and Computer Engineering, University of Illinois at Chicago

ABSTRACT


This paper discusses the potential of modelhardware co-design to simplify the implementation complexity of compute-in-SRAM deep learning considerably. Although compute-in-SRAM has emerged as a promising approach to improve the energy efficiency of DNN processing, current implementations suffer due to complex and excessive mixedsignal peripherals, such as the need for parallel digital-toanalog converters (DACs) at each input port. Comparatively, our approach inherently obviates complex peripherals by codesigning learning operators to SRAM’s operational constraints. For example, our co-designed implementation is DAC-free even for multibit precision DNN processing. Additionally, we also discuss the interaction of our compute-in-SRAM operator with Bayesian inference of DNNs. We show a synergistic interaction of Bayesian inference with our framework, where Bayesian methods allow achieving similar accuracy with much smaller network size. Although each iteration of sample-based Bayesian inference is computationally expensive, the cost is minimized by our computein- SRAM approach. Meanwhile, by reducing the network size, Bayesian methods reduce the footprint cost of compute-in-SRAM implementation, which is a crucial concern for the method. We characterize this interaction for deep learning-based pose (position and orientation) estimation for a drone.

Keywords: Deep neural networks, Compute-in-memory, Pose-estimation, Nanodrone.



Full Text (PDF)