eprintid: 3721 rev_number: 9 eprint_status: archive userid: 267 dir: disk0/00/00/37/21 datestamp: 2019-12-06 07:55:14 lastmod: 2020-07-14 09:40:56 status_changed: 2019-12-06 07:55:14 type: conference_item metadata_visibility: show creators_name: Tran, Xuan Tuyen creators_name: Nguyen, Duy Anh creators_name: Bui, Duy Hieu creators_name: Tran, Xuan Tu creators_id: xuantuyen2901@gmail.com creators_id: danguyen@vnu.edu.vn creators_id: hieubd@vnu.edu.vn creators_id: tutx@vnu.edu.vn title: A Variable Precision Approach for Deep Neural Networks ispublished: pub subjects: ElectronicsandComputerEngineering divisions: fac_fet divisions: lab_sis abstract: Deep Neural Network (DNN) architectures have been recently considered as the big breakthrough for a variety of applications. Because of the high computing capabilities required, DNN has been unsuitable for various embedded applications. Many works have been trying to optimize the key operations, which are multiply-and-add, in hardware for a smaller area, higher throughput, and lower power consumption. One way to optimize these factors is to use the reduced bit accuracy; for examples, Google's TPU used only 8-bit integer operations for DNN inference. Based on the characteristics of different layers in DNN, further bit accuracy can be changed to preserve the hardware area, power consumption, and throughput. In this work, the thesis investigates a hardware implementation of multiply-and-add with variable bit precision which can be adjusted at the computation time. The proposed design can calculate the sum of several products with the bit precision ranging from 1 to 16 bits. The hardware implementation results on Xilinx FPGA Virtex 707 development kit show that our design occupies smaller hardware and can run at a higher frequency of 310 MHz, while the same functionality implemented with and without DSP48 blocks can only run at a frequency of 102 MHz. In addition, to demonstrate that the proposed design is applicable effectively for deep neural network architecture, the paper also integrated the new design in the MNIST network. The simulation and verification results show that the proposed system can achieve the accuracy up of to 88%. date: 2019-10-17 date_type: published full_text_status: none pres_type: paper event_title: International Conference on Advanced Technologies for Communications (ATC) 2019 event_location: Ha Noi event_dates: 17-18 October 2019 event_type: conference refereed: TRUE citation: Tran, Xuan Tuyen and Nguyen, Duy Anh and Bui, Duy Hieu and Tran, Xuan Tu (2019) A Variable Precision Approach for Deep Neural Networks. In: International Conference on Advanced Technologies for Communications (ATC) 2019, 17-18 October 2019, Ha Noi.