Learning to infer: RL-based search for DNN primitive selection on Heterogeneous Embedded Systems
Miguel de Prado1,2,a, Nuria Pazos1,b and Luca Benini2,c
1He-Arc Ingenierie, HES-SO
amiguel.deprado@he-arc.ch
bnuria.pazos@he-arc.ch
2Integrated Systems Laboratory, ETH Zurich
clbenini@iis.ee.ethz.ch
ABSTRACT
Deep Learning is increasingly being adopted by industry for computer vision applications running on embedded devices. While Convolutional Neural Networks’ accuracy has achieved a mature and remarkable state, inference latency and throughput are a major concern especially when targeting low-cost and low-power embedded platforms. CNNs’ inference latency may become a bottleneck for Deep Learning adoption by industry, as it is a crucial specification for many real-time processes. Furthermore, deployment of CNNs across heterogeneous platforms presents major compatibility issues due to vendor-specific technology and acceleration libraries.
In this work, we present QS-DNN, a fully automatic search based on Reinforcement Learning which, combined with an inference engine optimizer, efficiently explores through the design space and empirically finds the optimal combinations of libraries and primitives to speed up the inference of CNNs on heterogeneous embedded devices. We show that, an optimized combination can achieve 45x speedup in inference latency on CPU compared to a dependency-free baseline and 2x on average on GPGPU compared to the best vendor library. Further, we demonstrate that, the quality of results and time “to-solution” is much better than with Random Search and achieves up to 15x better results for a short-time search.