Design of Many-Core Big Little μBrains for Energy-Efficient Embedded Neuromorphic Computing

M. Lakshmi Varshika1,a, Adarsha Balaji1, Federico Corradi2, Anup Das1, Jan Stuijt2 and Francky Catthoor3
1Electrical and Computer Engineering, Drexel University, USA
2Stichting IMEC Nederland, Netherlands
3IMEC, Belgium
al.luo@nudt.edu.cn

ABSTRACT


As spiking-based deep learning inference applications are increasing in embedded systems, these systems tend to integrate neuromorphic accelerators such as μBrain to improve energy efficiency. We propose a μBrain-based scalable many-core neuromorphic hardware design to accelerate the computations of spiking deep convolutional neural networks (SDCNNs). To increase energy efficiency, cores are designed to be heterogeneous in terms of their neuron and synapse capacity (i.e., big vs. little cores), and they are interconnected using a parallel segmented bus interconnect, which leads to lower latency and energy compared to a traditional mesh-based Network-on-Chip (NoC). We propose a system software framework called SentryOS to map SDCNN inference applications to the proposed design. SentryOS consists of a compiler and a run-time manager. The compiler compiles an SDCNN application into sub-networks by exploiting the internal architecture of big and little μBrain cores. The run-time manager schedules these sub-networks onto cores and pipeline their execution to improve throughput. We evaluate the proposed big little many-core neuromorphic design and the system software framework with five commonly-used SDCNN inference applications and show that the proposed solution reduces energy (between 37% and 98%), reduces latency (between 9% and 25%), and increases application throughput (between 20% and 36%).We also show that SentryOS can be easily extended for other spiking neuromorphic accelerators such as Loihi and DYNAPs.

Keywords: Neuromorphic Computing, Spiking Deep Convolutional Neural Networks, Many-Core, Embedded Systems, μBrain.



Full Text (PDF)