Skip to main navigation Skip to search Skip to main content

NAND-SPIN-based processing-in-MRAM architecture for convolutional neural network acceleration

  • Beihang University
  • Capital Normal University

Research output: Contribution to journalArticlepeer-review

Abstract

The performance and efficiency of running large-scale datasets on traditional computing systems exhibit critical bottlenecks due to the existing “power wall” and “memory wall” problems. To resolve those problems, processing-in-memory (PIM) architectures are developed to bring computation logic in or near memory to alleviate the bandwidth limitations during data transmission. NAND-like spintronics memory (NAND-SPIN) is one kind of promising magnetoresistive random-access memory (MRAM) with low write energy and high integration density, and it can be employed to perform efficient in-memory computation operations. In this study, we propose a NAND-SPIN-based PIM architecture for efficient convolutional neural network (CNN) acceleration. A straightforward data mapping scheme is exploited to improve parallelism while reducing data movements. Benefiting from the excellent characteristics of NAND-SPIN and in-memory processing architecture, experimental results show that the proposed approach can achieve ∼2.6× speedup and ∼1.4× improvement in energy efficiency over state-of-the-art PIM solutions.

Original languageEnglish
Article number142401
JournalScience China Information Sciences
Volume66
Issue number4
DOIs
StatePublished - Apr 2023

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 7 - Affordable and Clean Energy
    SDG 7 Affordable and Clean Energy

Keywords

  • NAND-like spintronics memory
  • convolutional neural network
  • magnetic tunnel junction
  • nonvolatile memory
  • processing-in-memory

Fingerprint

Dive into the research topics of 'NAND-SPIN-based processing-in-MRAM architecture for convolutional neural network acceleration'. Together they form a unique fingerprint.

Cite this