Abstract
Object detection methods based on neural networks have made considerable progress. However, methods like Faster RCNN and SSD that adopt large neural networks as the base models. It’s still a challenge to deploy such large detection networks in mobile or embedded devices. In this paper, we propose a low bit-width weight optimization approach to train Binary Neural Network for object detection using binary weights in training and testing. We introduce a greedy layer-wise method to train the detection network. This method boosts the performance instead of training the entire network at the same time. Our binary detection neural network (BDNN) can reduce the computational requirements and storage with competitive performance. For example, the binary network based on Faster RCNN with VGG16 can save 95% compression. In our experiments, BDNN achieves comparable performance with mAP 63.3% and outperforms SPPNet by 4.4% on PASCAL VOC 2007.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.