Low-overhead inverted LUT design for bounded DNN activation functions on floating-point vector ALUs
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kim, Seok Young | - |
dc.contributor.author | Kim, Chang Hyun | - |
dc.contributor.author | Lee, Won Joon | - |
dc.contributor.author | Park, Il | - |
dc.contributor.author | Kim, Seon Wook | - |
dc.date.accessioned | 2022-08-10T09:40:44Z | - |
dc.date.available | 2022-08-10T09:40:44Z | - |
dc.date.created | 2022-08-10 | - |
dc.date.issued | 2022-09 | - |
dc.identifier.issn | 0141-9331 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/142722 | - |
dc.description.abstract | An inference engine uses floating-point numbers to provide high accuracy in deep neural network computing despite its computing resource limitations. However, the computation for non-linear activation functions occurs the performance bottleneck, and we may alleviate it by adopting a lookup table (LUT) method. However, the floating-point number system's characteristic, where intervals between mantissa numbers differ depending on their exponent values, makes it challenging to calculate LUT index values and produce the error-tolerant outputs. This paper proposes a floating-point-based lookup table (FP-LUT) that produces minimal errors and requires negligible hardware cost, especially for vector arithmetic logic units (ALUs), using bfloat16 recently proposed for both inference and training processes. Instead of calculating the index using the function input value, we apply the principle of an inverse function for our design, especially targeting bounded DNN activation functions. We divide a range of function output values linearly by the number of LUT entry sizes and store the corresponding input values in the LUT. Then, we compare the incoming input value with the stored LUT values, find the corresponding address, and convert it into an FP format for the output. We applied our 32-entry FP-LUT to the in-house 8-way bfloat16 MAC unit to support four DNN activation functions: logistic sigmoid, hyper-tangent, soft sign, and ISRU, which incurs only 1.22% and 0.46% of the area and power consumption overhead. Our accuracy analysis shows that with only an entry size of 1/8 compared to state-of-the-art 16-bit fixed-point LUT methods and the small logic overhead, FP-LUT reduces the average errors by 51.8%, 28.4%, 14.4%, and 26.1% in those functions on our test datasets, respectively. Additionally, we show that our scheme satisfies all application-defined accuracy. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | ELSEVIER | - |
dc.title | Low-overhead inverted LUT design for bounded DNN activation functions on floating-point vector ALUs | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Kim, Seon Wook | - |
dc.identifier.doi | 10.1016/j.micpro.2022.104592 | - |
dc.identifier.scopusid | 2-s2.0-85133870087 | - |
dc.identifier.wosid | 000826734500001 | - |
dc.identifier.bibliographicCitation | MICROPROCESSORS AND MICROSYSTEMS, v.93 | - |
dc.relation.isPartOf | MICROPROCESSORS AND MICROSYSTEMS | - |
dc.citation.title | MICROPROCESSORS AND MICROSYSTEMS | - |
dc.citation.volume | 93 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Hardware & Architecture | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Theory & Methods | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.subject.keywordAuthor | Lookuptable | - |
dc.subject.keywordAuthor | Bfloat16 | - |
dc.subject.keywordAuthor | Activationfunctions | - |
dc.subject.keywordAuthor | Deepneuralnetworks | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.