Abstract

Port surveillance videos provide rich and intuitive temporal and spatial information and motion information, which is conducive to path planning, obstacle avoidance and subsequent risk prediction of automated guided vehicle (AGV) in automated container terminal (ACT). Extracting high-fidelity kinematic traffic relevant data from port surveillance videos become an active yet important research task in the ACT community. To that aim, the study proposes a deep-learning based framework to exploit spatial-temporal information from port videos with steps of object detection, object moving displacement mapping and speed estimation. First, the proposed framework detects both AGVs and people in the video with you only look once (YOLO) based model (i.e., YOLO V5 model). Second, we map the AGV (and people) moving distance from videos into physical world with the help of pinhole imaging rule. Third, we estimate AGV and people moving speed in the video with the help of linear regression model. The experiment results suggested that the proposed framework obtains satisfied performance considering that the average object moving displacement error is 0.19 m and the speed error is 0.08 m/s. The research findings help ACT management departments and regulations with high-fidelity traffic relevant data to further port safety and management efficiency.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.