Mobile QR Code QR CODE
Title In-depth Survey of Processing-in-memory Architectures for Deep Neural Networks
Authors (Ji-Hoon Jang) ; (Jin Shin) ; (Jun-Tae Park) ; (In-Seong Hwang) ; (Hyun Kim)
DOI https://doi.org/10.5573/JSTS.2023.23.5.322
Page pp.322-339
ISSN 1598-1657
Keywords Processing-in-memory; deep learning; next-generation memory; near-memory computing; deep neural network
Abstract Processing-in-Memory (PIM) is an emerging computing architecture that has gained significant attention in recent times. It aims to maximize data movement efficiency by moving away from the traditional von Neumann architecture. PIM is particularly well-suited for handling deep neural networks (DNNs) that require significant data movement between the processing unit and the memory device. As a result, there has been substantial research in this area. To optimally handle DNNs with diverse structures and inductive biases, such as convolutional neural networks, graph convolutional networks, recurrent neural networks, and transformers, within a PIM architecture, careful consideration should be given to how data mapping and data flow are processed in PIM. This paper aims to provide insight into these aspects by analyzing the characteristics of various DNNs and providing detailed explanations of how they have been implemented with PIM architectures using commercially available memory technologies like DRAM and next-generation memory technologies like ReRAM.