Quant-PIM: An Energy-Efficient Processing-in-Memory Accelerator for Layerwise Quantized Neural Networks
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lee, Young Seo | - |
dc.contributor.author | Chung, Eui-Young | - |
dc.contributor.author | Gong, Young-Ho | - |
dc.contributor.author | Chung, Sung Woo | - |
dc.date.accessioned | 2022-02-13T12:40:33Z | - |
dc.date.available | 2022-02-13T12:40:33Z | - |
dc.date.created | 2022-01-20 | - |
dc.date.issued | 2021-12 | - |
dc.identifier.issn | 1943-0663 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/135621 | - |
dc.description.abstract | Layerwise quantized neural networks (QNNs), which adopt different precisions for weights or activations in a layerwise manner, have emerged as a promising approach for embedded systems. The layerwise QNNs deploy only required number of data bits for the computation (e.g., convolution of weights and activations), which in turn reduces computation energy compared to the conventional QNNs. However, the layerwise QNNs still cause a large amount of energy in the conventional memory systems, since memory accesses are not optimized for the required precision of each layer. To address this problem, we propose Quant-PIM, an energy-efficient processing-in-memory (PIM) accelerator for layerwise QNNs. Quant-PIM selectively reads only required data bits within a data word depending on the precision, by deploying the modified I/O gating logics in a 3-D stacked memory. Thus, Quant-PIM significantly reduces energy consumption for memory accesses. In addition, Quant-PIM improves the performance of layerwise QNNs. When the required precision is half of the weight (or activation) size or less, Quant-PIM reads two data blocks in a single read operation by exploiting the saved memory bandwidth from the selective memory access, thus providing higher compute-throughput. Our simulation results show that Quant-PIM reduces system energy by 39.1%similar to 50.4% compared to the PIM system with 16-bit quantized precision, without accuracy loss. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Quant-PIM: An Energy-Efficient Processing-in-Memory Accelerator for Layerwise Quantized Neural Networks | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Chung, Sung Woo | - |
dc.identifier.doi | 10.1109/LES.2021.3050253 | - |
dc.identifier.scopusid | 2-s2.0-85099546052 | - |
dc.identifier.wosid | 000721999200007 | - |
dc.identifier.bibliographicCitation | IEEE EMBEDDED SYSTEMS LETTERS, v.13, no.4, pp.162 - 165 | - |
dc.relation.isPartOf | IEEE EMBEDDED SYSTEMS LETTERS | - |
dc.citation.title | IEEE EMBEDDED SYSTEMS LETTERS | - |
dc.citation.volume | 13 | - |
dc.citation.number | 4 | - |
dc.citation.startPage | 162 | - |
dc.citation.endPage | 165 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Hardware & Architecture | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Software Engineering | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.subject.keywordAuthor | Accelerator | - |
dc.subject.keywordAuthor | energy efficiency | - |
dc.subject.keywordAuthor | layerwise quantization | - |
dc.subject.keywordAuthor | processing-in-memory (PIM) | - |
dc.subject.keywordAuthor | quantized neural network (QNN) | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.