Abstract:
In this paper, the system level evaluation is performed for DNN inference engines using SOT-MRAM, which includes compute-in-memory (CIM) paradigm and near-memory systolic...Show MoreMetadata
Abstract:
In this paper, the system level evaluation is performed for DNN inference engines using SOT-MRAM, which includes compute-in-memory (CIM) paradigm and near-memory systolic array. The write performance of the SOT materials is projected to 7nm with a macrospin model. For read-intensive CIM, SOT-MRAM with increased on-resistance can achieve 51% and 93% higher energy efficiency than 8T-SRAM at 22nm and 7nm nodes, respectively. For write-intensive systolic array at 7nm node, SOT-MRAM with PtCu track shows 17% higher energy efficiency than SRAM global buffer, respectively.
Published in: 2022 IEEE International Memory Workshop (IMW)
Date of Conference: 15-18 May 2022
Date Added to IEEE Xplore: 25 May 2022
ISBN Information: