Abstract

This paper proposes low-biased probabilistic multipliers and applies the proposed designs to the inference stage of convolutional neural networks (CNNs). Highly inaccurate compressors in the probabilistic multiplier make the error distribution unbalanced and produce significant relative errors. We describe design rules for applying different compressors to the probabilistic multiplier. Besides, the proposed design rules enhance the error characteristics by determining the input values of each compressor, suppressing the significant relative error. Notably, motivated by the design rules, we propose a novel inaccurate 4:2 compressor suitable for our new design. The design rules and proposed 4:2 compressor are adopted in 8-bit probabilistic multipliers. The error analysis shows that the new designs can enhance error characteristics without increasing hardware costs. Furthermore, the proposed approximate multipliers are applied to the convolutional layers of CNNs. We describe the fine-tuning scheme to reduce the retraining time of approximation-aware training. The experiments show that proposed designs can produce acceptable classification results compared with those using the floating-point and 8-bit exact multiplications on CIFAR and ImageNet datasets.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call