Abstract

Accurately identifying all objects of interest inside the specified frame of reference is crucial for object identification techniques to allow machine vision to interpret images successfully. Theoretical frameworks from computer vision and deep learning have informed many potential solutions to this problem. However, current approaches frequently fail when faced with objects going through random geometric changes and continually show shortcomings when identifying small, dense objects. This research looks at state-of-the-art object detection methods, compares them, and then suggests a convolutional object detection network that has been tweaked to fix the problems with the existing methods. Comparing our research to existing approaches, we find that they perform better. We accomplish this by training deep convolutional networks to detect geometric transformations and adjusting the networks to handle multi-scaled features. The results achieved after the optimization of the You only look once (YOLO) V5 & V7 are stated in this paper. The experiments demonstrated that YOLO v8 achieved higher accuracy than the hyper-tuned YOLO V5 and V7, with an average Precision (mAP) score of 52.8% on the customized military dataset running at 60 Frames Per Second (FPS). Object detection using deep neural networks (DNNs) poses a significant challenge due to the high computational and power requirements, mainly when deployed on general-purpose platforms such as CPUs and GPUs. However, addressing these limitations becomes crucial for efficient edge computing tasks like object detection, where faster, smaller, and energy-efficient solutions are essential. To overcome these challenges, system-on-chip (SoC) designs emerge as a promising solution.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call