12.6 Special Session: Computing with Emerging Memories: How Good can it be?

Printer-friendly version PDF version

Date: Thursday 22 March 2018
Time: 16:00 - 17:30
Location / Room: Konf. 4

Chair:
Pierre-Emmanuel Gaillardon, University of Utah, US

Co-Chair:
Ian O’Connor, Ecole Centrale de Lyon, FR

With the recent evolutions of nanometer transistor technologies, power consumption emerged as the most critical limitation. Within advanced processors and computing architectures, the processor-memory communication accounts for a significant part of the energy requirement. While alternative design approaches, such as the use of optimized accelerators or advanced power management techniques are successfully employed in contemporary designs, the trend keeps worsening due to the ever-increasing gap between on-chip and off-chip memory data rates. This trend, known as Von Neumann bottleneck, not only limits the system performance, but also acts nowadays as a limiter of the energy scaling. The quest towards more energy-efficiency requires solutions that solve the Von Neumann bottleneck by tightly intertwining computing with memories. In this hot topic session, we intend to elaborate on in-memory computing by identifying and compare the latest computing models in light of conventional, e.g., SRAMs, and emerging memory technologies, e.g., RRAMs, STT-MRAMs. In-memory computing is considered here in the general sense of computing information locally within large data storage.

TimeLabelPresentation Title
Authors
16:0012.6.1PRACTICAL CHALLENGES IN DELIVERING THE PROMISES OF REAL PROCESSING-IN-MEMORY MACHINES
Speaker:
Nishil Talati, Technion - Israel Institute of Technology, IL
Authors:
Nishil Talati1, Ameer Haj Ali2, Rotem Ben Hur2, Nimrod Wald2, Ronny Ronen2, Pierre-Emmanuel Gaillardon3 and Shahar Kvatinsky1
1Technion, IL; 2Technion - Israel Institute of Technology, IL; 3University of Utah, US
Abstract
Processing-in-Memory (PiM) machines promise to overcome the von Neumann bottleneck in order to further scale performance and energy efficiency of computing systems by reducing the extent of data transfer and offering ample parallelism. In this paper, we take the memristive Memory Processing Unit (mMPU) as a case study of a PiM machine and scrutinize it in practical scenarios. Specifically, we explore the limitations of parallelism and data transfer elimination. We argue that lack of operand locality and arrangement might make data transfer inevitable in the mMPU. We then devise techniques to move data within the mMPU, without transferring it off-chip, and quantify their costs. Additionally, we present electrical parameters that might limit the parallelism offered by the mMPU and evaluate their impact. Using benchmarks from the LGsynth91 suite, their vector extensions, and a few synthetic data-parallel workloads, we show that the internal data transfer results in an increase of up to 1.5x in the execution time, while the limited parallelism increases it by 1.1x to 2x.

Download Paper (PDF; Only available from the DATE venue WiFi)
16:3012.6.2SMART INSTRUCTION CODES FOR IN-MEMORY COMPUTING ARCHITECTURES COMPATIBLE WITH STANDARD SRAM INTERFACES
Speaker:
Maha Kooli, CEA-Leti, FR
Authors:
Maha Kooli1, Henri-Pierre CHARLES2, Bastien Giraud3 and Jean-Philippe Noel2
1CEA/LETI, FR; 2CEA, FR; 3CEA LETI, FR
Abstract
This paper presents the computing model for the In-Memory Computing architecture based on SRAM memory that embeds computing abilities. This memory concept offers significant performance gains in terms of energy consumption and execution time. To handle the interaction between the memory and the CPU, new memory instruction codes were designed. Those instructions are communicated by the CPU to the memory, using standard SRAM buses. This implementation allows (1) to embed In-Memory Computing capabilities on a system without Instruction Set Architecture (ISA) modification, and (2) to finely interlace CPU instructions and in-memory computing instructions.

Download Paper (PDF; Only available from the DATE venue WiFi)
17:0012.6.4MEMRISTIVE DEVICES FOR COMPUTATION-IN-MEMORY
Speaker:
Said Hamdioui, Delft University of Technology, NL
Authors:
Jintao Yu, HoangAnh DuNguyen, Mottaqiallah Taouil and Said Hamdioui, TU Delft, NL
Abstract
CMOS technology and its continuous scaling have made electronics and computers accessible and affordable for almost everyone on the globe; in addition, they have enabled the solutions of a wide range of societal problems and applications. Today, however, both the technology and the computer architectures are facing severe challenges/walls making them incapable of providing the demanded computing power with tight constraints. This motivates the need for the exploration of novel architectures based on new device technologies; not only to sustain the financial benefit of technology scaling, but also to develop solutions for extremely demanding emerging applications. This paper presents two computation-in-memory based accelerators making use of emerging memristive devices; they are Memristive Vector Processor and RRAM Automata Processor. The preliminary results of these two accelerators show significant improvement in terms of latency, energy and area as compared to today's architectures and design.

Download Paper (PDF; Only available from the DATE venue WiFi)
17:1512.6.3COMPUTING-IN-MEMORY WITH SPINTRONICS
Speaker:
Shubham Jain, Purdue University, US
Authors:
Shubham Jain1, Sachin Sapatnekar2, Jian-Ping Wang2, Kaushik Roy1 and Anand Raghunathan1
1Purdue University, US; 2Department of Electrical and Computer Engineering, University of Minnesota, US
Abstract
In-memory computing is a promising approach to alleviating the processor-memory data transfer bottleneck in computing systems. While spintronics has attracted great interest as a non-volatile memory technology, recent work has shown that its unique properties can also enable in-memory computing. We summarize efforts in this direction, and describe three different designs that enhance STT-MRAM to perform logic, arithmetic, and vector operations and evaluate transcendental functions within memory arrays.

Download Paper (PDF; Only available from the DATE venue WiFi)
17:30End of session