Abstract

In traditional 3D reconstruction methods, using a single view to predict the 3D structure of an object is a very difficult task. This research mainly discusses human pose recognition and estimation based on 3D multiview basketball sports dataset. The convolutional neural network framework used in this research is VGG11, and the basketball dataset Image Net is used for pretraining. This research uses some modules of the VGG11 network. For different feature fusion methods, different modules of the VGG11 network are used as the feature extraction network. In order to be efficient in computing and processing, the multilayer perceptron in the network model is implemented by a one‐dimensional convolutional network. The input is a randomly sampled point set, and after a layer of perceptron, it outputs a feature set of n × 16. Then, the feature set is sent to two network branches, one is to continue to use the perceptron method to generate the feature set of n × 1024, and the other network is used to extract the local features of points. After the RGB basketball sports picture passes through the semantic segmentation network, a picture containing the target object is obtained, and the picture is input to the constructed feature fusion network model. After feature extraction is performed on the RGB image and the depth image, respectively, the RGB feature, the local feature of the point cloud, and the global feature are spliced and fused to form a feature vector of N × 1152. There are three branches for this vector network, which, respectively, predict the object position, rotation, and confidence. Among them, the feature dimensionality reduction is realized by one‐dimensional convolution, and the activation function is the ReLU function. After removing the feature mapping module, the accuracy of VC‐CNN_v1 dropped by 0.33% and the accuracy of VC‐CNN_v2 dropped by 0.55%. It can be seen from the research results that the addition of the feature mapping module improves the recognition effect of the network to a certain extent

Highlights

  • Introduction e3D reconstruction based on the depth map requires the input of RGB image and the corresponding depth image

  • After the RGB basketball sports picture passes through the semantic segmentation network, a picture containing the target object is obtained, and the picture is input to the constructed feature fusion network model

  • After feature extraction is performed on the RGB image and the depth image, respectively, the RGB feature, the local feature of the point cloud, and the global feature are spliced and fused to form a feature vector of N × 1152. ere are three branches for this vector network, which, respectively, predict the object position, rotation, and confidence

Read more

Summary

Introduction

Introduction e3D reconstruction based on the depth map requires the input of RGB image and the corresponding depth image. E framework he proposed is “AR-NUI-REHAB-MDSS,” which uses natural user interface- (NUI-) based physical therapy rehabilitation, a personalized exercise presentation and monitoring system for patients, and a mobile decision support system (MDSS) for therapists He proposed AR, a less entertaining adjuvant therapy environment, the research lacks comparative data [3]. E Brito mission augmented reality (AR) platform is used in various applications His experimental design compared two different optical tracking systems based on ARa-marked AR (MB) and unmarked AR (ML) for two types of interfaces: a tangible interface based on gesture recognition and a multimodal interface. Both AR technologies allow consumers to visually observe the function of sports shoes. He compared AR (MB) and AR (ML), the research method is too complicated [4]

Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.