TY - JOUR
T1 - Binary Neural Networks in FPGAs
T2 - Architectures, Tool Flows and Hardware Comparisons
AU - Su, Yuanxin
AU - Seng, Kah Phooi
AU - Ang, Li Minn
AU - Smith, Jeremy
N1 - Publisher Copyright:
© 2023 by the authors.
PY - 2023/11
Y1 - 2023/11
N2 - Binary neural networks (BNNs) are variations of artificial/deep neural network (ANN/DNN) architectures that constrain the real values of weights to the binary set of numbers {−1,1}. By using binary values, BNNs can convert matrix multiplications into bitwise operations, which accelerates both training and inference and reduces hardware complexity and model sizes for implementation. Compared to traditional deep learning architectures, BNNs are a good choice for implementation in resource-constrained devices like FPGAs and ASICs. However, BNNs have the disadvantage of reduced performance and accuracy because of the tradeoff due to binarization. Over the years, this has attracted the attention of the research community to overcome the performance gap of BNNs, and several architectures have been proposed. In this paper, we provide a comprehensive review of BNNs for implementation in FPGA hardware. The survey covers different aspects, such as BNN architectures and variants, design and tool flows for FPGAs, and various applications for BNNs. The final part of the paper gives some benchmark works and design tools for implementing BNNs in FPGAs based on established datasets used by the research community.
AB - Binary neural networks (BNNs) are variations of artificial/deep neural network (ANN/DNN) architectures that constrain the real values of weights to the binary set of numbers {−1,1}. By using binary values, BNNs can convert matrix multiplications into bitwise operations, which accelerates both training and inference and reduces hardware complexity and model sizes for implementation. Compared to traditional deep learning architectures, BNNs are a good choice for implementation in resource-constrained devices like FPGAs and ASICs. However, BNNs have the disadvantage of reduced performance and accuracy because of the tradeoff due to binarization. Over the years, this has attracted the attention of the research community to overcome the performance gap of BNNs, and several architectures have been proposed. In this paper, we provide a comprehensive review of BNNs for implementation in FPGA hardware. The survey covers different aspects, such as BNN architectures and variants, design and tool flows for FPGAs, and various applications for BNNs. The final part of the paper gives some benchmark works and design tools for implementing BNNs in FPGAs based on established datasets used by the research community.
KW - binary neural network (BNN)
KW - computational modeling
KW - field-programmable gate array (FPGA)
KW - latency reduction
UR - http://www.scopus.com/inward/record.url?scp=85177754824&partnerID=8YFLogxK
U2 - 10.3390/s23229254
DO - 10.3390/s23229254
M3 - Review article
C2 - 38005640
AN - SCOPUS:85177754824
SN - 1424-8220
VL - 23
JO - Sensors
JF - Sensors
IS - 22
M1 - 9254
ER -