Abstract
The use of machine learning is expanding in various applications such as image processing in data centers. With the spread of deep learning, neural-network-based models have frequently been adopted in recent years. Because the processing speed is slow when evaluating machine learning on a CPU, a fast-dedicated hardware accelerator is often used. In particular, the demand for hardware accelerators in data centers is increasing. A low power consumption and high-speed processing are required in a limited space. An implementation method for a ternary neural network utilizing the rewritable look-up table (LUT) of a field-programmable gate array (FPGA) is proposed. Binary/ternary neural networks, which are quantized to 1–2 bits for mapping to LUTs, suffer from a poor recognition accuracy. To prevent a decrease in the recognition accuracy, let <tex xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">$q$</tex> be the number of quantization bits to be stored in the LUT. The memory size of the LUT becomes <tex xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">$O(2^{q})$</tex> bits, and it tends to be exponential in size. We improved the accuracy using an ensemble ternary neural network. There are various ways to select data for ensembles during training, and various ways to select branch pruning for trivialized neural networks. We chose the greedy method for our design. An evaluation using various benchmark datasets showed that the ensemble approach achieved a recognition accuracy equivalent to that of the 32-bit float model. We also estimated the amount of memory required to implement an LUT for an ensemble ternary neural network. The size of the LUT is 1.9 Mbit, which can be realized on the current FPGAs.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.