Abstract

This paper proposes a method of automatically annotating tennis action through the integrated use of audio and video information. The proposed method extracts ball-hitting times called “impact times” using audio information, and evaluates the position relations between the player and the ball at the impact time to identify the player’s basic actions, such as forehand swing, overhead swing, etc. Simulation results show that the detection rate for impact time influences the recognition rate of the player’s basic actions. They also reveal that using audio information avoids some event recognition failures that cannot be averted when using only video information, demonstrating the performance and the validity of our approach.KeywordsBasic ActionBall PositionAutomatic AnnotationAudio DataImpact TimeThese keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call