Cited 8 time in
Multipurpose Deep-Learning Accelerator for Arbitrary Quantization With Reduction of Storage, Logic, and Latency Waste
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Moon, Seunghyun | - |
| dc.contributor.author | Mun, Han-Gyeol | - |
| dc.contributor.author | Son, Hyunwoo | - |
| dc.contributor.author | Sim, Jae-Yoon | - |
| dc.date.accessioned | 2023-11-07T04:40:23Z | - |
| dc.date.available | 2023-11-07T04:40:23Z | - |
| dc.date.issued | 2024-01 | - |
| dc.identifier.issn | 0018-9200 | - |
| dc.identifier.issn | 1558-173X | - |
| dc.identifier.uri | https://scholarworks.gnu.ac.kr/handle/sw.gnu/68356 | - |
| dc.description.abstract | Various pruning and quantization heuristics have been proposed to compress recent deep-learning models. However, the rapid development of new optimization techniques makes it difficult for domain-specific accelerators to efficiently process various models showing irregularly stored parameters or nonlinear quantization. This article presents a scalable-precision deep-learning accelerator that supports multiply-and-accumulate operations (MACs) with two arbitrarily quantized data sequences. The proposed accelerator includes three main features. To minimize logic overhead when processing arbitrarily quantized 8-bit precision data, a lookup table (LUT)-based runtime reconfiguration is proposed. The use of bit-serial execution without unnecessary computations enables the multiplication of data with non-equal precision while minimizing logic and latency waste. Furthermore, two distinct data formats, raw and run-length compressed, are supported by a zero-eliminator (ZE) and runtime-density detector (RDD) that are compatible with both formats, delivering enhanced storage and performance. For a precision range of 1–8 bit and fixed sparsity of 30%, the accelerator implemented in 28 nm low-power (LP) CMOS shows a peak performance of 0.87–5.55 TOPS and a power efficiency of 15.1–95.9 TOPS/W. The accelerator supports processing with arbitrary quantization (AQ) while achieving state-of-the-art (SOTA) power efficiency. IEEE | - |
| dc.format.extent | 14 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | Institute of Electrical and Electronics Engineers | - |
| dc.title | Multipurpose Deep-Learning Accelerator for Arbitrary Quantization With Reduction of Storage, Logic, and Latency Waste | - |
| dc.type | Article | - |
| dc.publisher.location | 미국 | - |
| dc.identifier.doi | 10.1109/JSSC.2023.3312615 | - |
| dc.identifier.scopusid | 2-s2.0-85174825344 | - |
| dc.identifier.wosid | 001088286600001 | - |
| dc.identifier.bibliographicCitation | IEEE Journal of Solid-State Circuits, v.59, no.1, pp 1 - 14 | - |
| dc.citation.title | IEEE Journal of Solid-State Circuits | - |
| dc.citation.volume | 59 | - |
| dc.citation.number | 1 | - |
| dc.citation.startPage | 1 | - |
| dc.citation.endPage | 14 | - |
| dc.type.docType | Article | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.relation.journalResearchArea | Engineering | - |
| dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
| dc.subject.keywordAuthor | Arbitrary quantization (AQ) | - |
| dc.subject.keywordAuthor | bit-serial processing | - |
| dc.subject.keywordAuthor | Computer architecture | - |
| dc.subject.keywordAuthor | Decoding | - |
| dc.subject.keywordAuthor | deep neural network (DNN) accelerator | - |
| dc.subject.keywordAuthor | Hardware | - |
| dc.subject.keywordAuthor | lookup table (LUT) | - |
| dc.subject.keywordAuthor | Moon | - |
| dc.subject.keywordAuthor | precision scalability | - |
| dc.subject.keywordAuthor | Quantization (signal) | - |
| dc.subject.keywordAuthor | run-length compression (RLC) | - |
| dc.subject.keywordAuthor | Table lookup | - |
| dc.subject.keywordAuthor | Task analysis | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
Gyeongsang National University Central Library, 501, Jinju-daero, Jinju-si, Gyeongsangnam-do, 52828, Republic of Korea+82-55-772-0532
COPYRIGHT 2022 GYEONGSANG NATIONAL UNIVERSITY LIBRARY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
