Abstract

It is difficult to create sample datasets for semantic identification for animals with quick behavior traits, such as cats, and the cost of dataset gathering is greatly increased by high-speed motion capture cameras. This paper takes the common Chinese dragon-li as an example, carries on the solid modeling and the gait movement design in the digital 3D modeling software, then constructs the virtual image dataset vCAT through the image rendering technology. To account for changing elements like the body shape, and scene and camera perspective, the vCAT dataset has numerous data sub-categories. In addition, we accomplished the pixel-level segmentation test of synthetic images and actual scene images using deep learning on the vCAT dataset based on the attention mechanism. During the basis of training, the recognition accuracy rate can reach 84% when only 726 images with real scene modeling were used. In addition to providing a fully automated technical method for creating the original image of the virtual image dataset with its corresponding labels, we also thoroughly test and discuss elements like scenes, perspectives, textures, and lighting in the virtual dataset to serve as a foundation for its practical use in deep learning training.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.