Abstract

Convolutional Neural Networks (CNN) have made splendid achievements in various object detection tasks. To extend the applications of CNN detection models, the implementation of model inference on edge platforms, such as ASIC, FPGA and other embedded systems, has been intensively investigated in recent years. However, the huge model size and its enormous overhead constrain the deployment of detection model on edge platform which always has limited computational capability. Quantized inference of CNN model is one of the efficient approach to running model on edge platform. In this paper, we develop a hardware-friendly quantized inference scheme of detection model that is used for efficient inference on embedded FPGA systems. The proposed method contains several techniques that are able to optimize the quantized inference of detection model on FPGA device. The experimental results demonstrate that not only make the detection model quantized inference more efficient but also maintain the accuracy of object detection.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.