Accelerating Deep Neural Networks Using FPGAs and ZYNQ

Han Sung Lee, Jae Wook Jeon

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

4 Scopus citations

Abstract

This article aims at implementing a Deep Neural Network (DNN) using Field Programmable Gate Arrays (FPGAs) for real time deep learning inference in embedded systems. In now days DNNs are widely used where high accuracy is required. However, due to the structural complexity, deep learning models are highly computationally intensive. To improve the system performance, optimization techniques such as weight quantization and pruning are commonly adopted. Another approach to improve the system performance is by applying heterogeneous architectures. Processor with Graphics Processing Unit (GPU) architectures are commonly used for deep learning training and inference acceleration. However, GPUs are expensive and consume much power that not a perfect solution for embedded systems. In this paper, we implemented a deep neural network on a Zynq SoC which is a heterogenous system integrated of ARM processor and FPGA. We trained the model with MNIST database, quantized the model's 32-bit floating point weights and bias into integer and implemented model to inference in FPGA. As a result, we deployed a network on an embedded system while maintaining inference accuracy and accelerated the system performance with using less resources.

Original languageEnglish
Title of host publicationTENSYMP 2021 - 2021 IEEE Region 10 Symposium
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781665400268
DOIs
StatePublished - 23 Aug 2021
Externally publishedYes
Event2021 IEEE Region 10 Symposium, TENSYMP 2021 - Jeju, Korea, Republic of
Duration: 23 Aug 202125 Aug 2021

Publication series

NameTENSYMP 2021 - 2021 IEEE Region 10 Symposium

Conference

Conference2021 IEEE Region 10 Symposium, TENSYMP 2021
Country/TerritoryKorea, Republic of
CityJeju
Period23/08/2125/08/21

Keywords

  • AI
  • Deep Learning
  • Deep Neural Networks
  • FPGA
  • Quantization
  • ZYNQ

Fingerprint

Dive into the research topics of 'Accelerating Deep Neural Networks Using FPGAs and ZYNQ'. Together they form a unique fingerprint.

Cite this