Title: High-speed pre-accumulator and post-multiplier for convolution neural networks with low power consumption

Authors: K. Mariya Priyadarshini; R.S. Ernest Ravindran; M. Sujatha; K.T.P.S. Kumar

Addresses: Department of Electronics and Communication Engineering, K.L. Deemed to be University, Guntur District, A.P., India ' Department of Electronics and Communication Engineering, K.L. Deemed to be University, Guntur District, A.P., India ' Department of Electronics and Communication Engineering, K.L. Deemed to be University, Guntur District, A.P., India ' Department of Electronics and Communication Engineering, K.L. Deemed to be University, Guntur District, A.P., India

Abstract: In today's phase of growing technology Convolution Neural Networks (CNNs) are all over the places. It is a thriving segment in machine learning and Artificial Intelligences (AI) techniques. CNN needs bulk amount of computing competence and memory with higher frequency range. In this present investigation, Pre-Accumulator and Post-Multipliers (PAPM) are proposed which accelerate the speed of processor. 4-bit multiplier using Carry Save Adder (CSA) is built with 6Transistors-Adder and sutras of Vedic mathematics is constructed. Accumulator of multiplier and accumulator are designed with Two Level Edge Triggering Flip-Flops (TLET-FF) to increase bandwidth of the memory. The proposed architecture of Multiply Accumulate (MAC) circuit consumes very less power when compared to existing high speed MACs. Performance of accumulator is contrasted with three different, two-level triggered flip-flops namely 16TLET-FF, 14TLET-FF and 12TLET-FFs. The projected MAC replaces the existing multipliers due its low power together with high frequency of operation.

Keywords: convolution neural networks; Vedic sutras; carry save adder; flip-flop; multiplier and accumulator.

DOI: 10.1504/IJIPT.2022.125962

International Journal of Internet Protocol Technology, 2022 Vol.15 No.3/4, pp.139 - 147

Received: 05 Jan 2021
Accepted: 17 May 2021

Published online: 05 Oct 2022 *

Full-text access for editors Full-text access for subscribers Purchase this article Comment on this article