Abstract
In this study, we implemented YOLO (You Only Look Once) for real-time object detection and evaluated its performance based on key metrics such as processing speed, frame rate, and object detection accuracy. Our approach emphasizes both the precision and efficiency of YOLO, focusing on its ability to detect objects in real-world scenarios while maintaining a low computational cost. To identify and count objects, the YOLO algorithm was applied to analyze three images. It divided each image into a grid, and each cell predicted bounding boxes and confidence scores for potential objects. Following the processing of these predictions using non-max suppression to remove duplicates, each image contained an accurate count of the items that were detected. The model achieved a processing time of 17.68 seconds, with an average of 0.25 seconds per frame, demonstrating the system's capability for rapid object detection in near real-time applications. On average, 1.32 objects were detected per frame, with a maximum of 1.67 objects in a single frame and a minimum of 1 object per frame, indicating consistent detection across the dataset. The standard deviation of objects per frame (0.113) shows a low variability in object detection rates, reflecting the robustness of the model in handling diverse input frames. The achieved frame rate of 4.2 FPS demonstrates the model's potential for real-time applications, particularly in environments where processing speed is critical. The scientific novelty of this work lies in demonstrating YOLO’s adaptability for efficient object detection while maintaining high detection rates and consistent performance across varying scenarios. This study contributes to the field by showcasing YOLO's applicability in real-time systems, where object detection speed and accuracy are paramount. Our findings provide a foundation for further optimization in high-performance, low-latency object detection tasks, as well as its scalability for more complex detection systems. The results underscore YOLO's potential in both academic and industrial settings.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have