Abstract

This research presents the extension and application of a voltage and frequency scaling framework called Elongate to a high-performance and reconfigurable binarized neural network. The neural network is created in the FPGA reconfigurable fabric and coupled to a multiprocessor host that controls the operational point to obtain energy proportionality. Elongate instruments a design netlist by inserting timing detectors to enable the exploitation of the operating margins of a device reliably. The elongated neural network is re-targeted to devices with different nominal operating voltages and fabricated with 28 nm (i.e., Zynq) and 16nm (i.e., Zynq Ultrascale) feature sizes showing the portability of the framework to advanced process nodes. New hardware and software components are created to support the 16nm fabric microarchitecture and a comparison in terms of power, energy and performance with the older 28 nm process is performed. The results show that Elongate can obtain new performance and energy points that are up to 86 percent better than nominal at the same level of classification accuracy. Trade-offs between energy and performance are also possible with a large dynamic range of valid working points available. The results also indicate that the built-in neural network robustness allows operation beyond the first point of error while maintaining the classification accuracy largely unaffected.

Highlights

  • FULLY binarized neural networks are a type of convolutional neural networks that reduce the precision of weights and activations from floating point to binary values

  • The conclusion is that the binarized neural network (BNN) built-in error tolerance could be exploited to increase Elongate performance/energy efficiency higher than the error-free value of 86 percent if slight variations of classification accuracy are acceptable in the application

  • In this paper we extended the Elongate framework originally created for Zynq devices to the Ultrascale Zynq devices and integrate it with the SDx toolset that enables hardware design based on C/C++

Read more

Summary

INTRODUCTION

FULLY binarized neural networks are a type of convolutional neural networks that reduce the precision of weights and activations from floating point to binary values. The results show that Elongate can determine extended operating points of voltage and frequency, enabling higher performance, lower power or trade-offs between performance and power so the amount of computation and energy usage adapts to the workload requirements at run-time. This adaptation maximizes the performance/power and improves the energy proportionality of the system as defined in [3] by eliminating the waste incurred when the system operates at maximum performance and idles when no more work is available.

Convolutional Neural Network Accelerators
Adaptive Voltage and Frequency Scaling
Volt Up to 4 64-bit HP ports 1 64-bit ACP coherent port
HARDWARE PLATFORMS SPECIFICATION
ELONGATE FRAMEWORK
Elongate Interfacing and Control
BINARISED NEURAL NETWORK APPLICATION
ELONGATE OVERHEADS
POWER SCALING
32 K 36 K 131
ENERGY AND PERFORMANCE ANALYSIS
ACCURACY ANALYSIS
10 ENERGY PROPORTIONAL COMPUTING ANALYSIS
Findings
11 CONCLUSIONS
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call