Design and Implementation of Reconfigurable Neural Network Accelerator

dc.contributor.authorShenoy, M.S.
dc.contributor.authorRamesh Kini, M.
dc.date.accessioned2026-02-06T06:35:19Z
dc.date.issued2022
dc.description.abstractGeneral-purpose CPUs are sluggish and inefficient when used for computationally intensive applications including in neural networks. It is preferable to develop specialized hardware that can do a large number of multiply-accumulate operations rapidly and efficiently to execute such applications. The Re-configurable Neural Network Accelerator (RNNA) architecture that has been designed is appropriate for a variety of neural network applications. The computational resource requirements vary depending on the application; hence, mapping the application to the available set of resources requires reconfigurability. The fundamental unit of the RNNA is composed of a variety of Multiply-Accumulate (MAC) units, registers, and Address Generation Units (AGU). When compared to the computation performed by a single MAC array, the RNNA with four MAC arrays reduces the time required by approximately 75%. On the Nexys4 DDR Artix-7 FPGA board, RNNA was tested and implemented with a clock frequency of up to 60MHz and power consumption of 0.243W. © 2022 IEEE.
dc.identifier.citation7th IEEE International Conference on Recent Advances and Innovations in Engineering, ICRAIE 2022 - Proceedings, 2022, Vol., , p. 377-381
dc.identifier.urihttps://doi.org/10.1109/ICRAIE56454.2022.10054301
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/29780
dc.publisherInstitute of Electrical and Electronics Engineers Inc.
dc.subjectBatch Processing
dc.subjectConvolutional Neural Net-work
dc.subjectDeep Learning Accelerator
dc.subjectMultiply-Accumulate
dc.subjectNeural Networks
dc.subjectReconfigurability
dc.subjectTensor Processing Unit
dc.titleDesign and Implementation of Reconfigurable Neural Network Accelerator

Files