Mobile QR Code
Title Low-Complexity Inference Accelerator for Binarized Convolutional Neural Networks
Authors 최경찬(Gyungchan Choi) ; 신지훈(Jihoon Shin) ; 김태환(Tae-Hwan Kim)
DOI https://doi.org/10.5573/ieie.2020.57.1.53
Page pp.53-60
ISSN 2287-5026
Keywords BCNN ; Inference accelerator ; FPGA ; Embedded system ; Low complexity ;
Abstract This paper proposes a low-complexity processor that accelerates the inference of the binarized convolutional neural networks. The proposed processor performs overall inference by the block, whose processing steps are convolution, binarization, and pooling. Each block is formulated to have a binary input and binary output, so that additional buffers to store inter-step temporary results can be eliminated effectively. The proposed processor processes each block by decomposing it to multiple 1×1 convolutions, thereby accelerates any blocks including convolutional layer and fully-connected layer, in a consistent way. It is implemented on Cyclone V FPGA with only 1.05 K adaptive logic module (ALM) and 2304 Kbit on-chip memory, showing the complexity efficiency of 22.514 GOP/s/KALM.