Lightweight Human Behavior Recognition Method for Visual Communication AGV Based on CNN-LSTM

  • Abstract
  • Highlights & Summary
  • References
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Lightweight Human Behavior Recognition Method for Visual Communication AGV Based on CNN-LSTM

Highlights

  • Behavior recognition uses deep learning network model to automatically extract the deep features of data, but traditional machine learning algorithms have some problems such as manual feature extraction and poor generalization ability of models

  • According to different data sources, human behavior recognition can be divided into human behavior recognition based on video images and human behavior recognition based on wearable sensors[3]

  • Human behavior recognition based on video images is to make use of related technologies such as image processing and video processing, and to realize human behavior recognition and specific target detection by analyzing the set of human moving images or video clips acquired by camera equipment[4, 5]

Read more Highlights Expand/Collapse icon

Summary

Behavior recognition modelExpand/Collapse icon

In the 3D convolution behavior recognition model, the extraction and effective utilization of time information is an important factor. In order to reduce the parameters without reducing the accuracy of the model, we propose a new network structure[21]. For the original video information of indefinite duration, the sparse time sampling strategy proposed by Temporal Segment Network (TSN) is used for sampling[23]. The network is reorganized in tail module: the original spatial dimensions are merged into one dimension, a 2D convolutional neural network is obtained. Continue convolution in this 2D network to complete feature extraction.

Convolutional feature maps aligned with LSTM sequence featuresExpand/Collapse icon
Experimental environmentExpand/Collapse icon
Backbone network selectionExpand/Collapse icon
Experimental resultExpand/Collapse icon
Ablation experimentExpand/Collapse icon
FindingsExpand/Collapse icon
ConclusionExpand/Collapse icon
ReferencesShowing 10 of 12 papers
  • Cite Count Icon 52
  • 10.1021/acs.inorgchem.1c03023
CdS@CdSe Core/Shell Quantum Dots for Highly Improved Self-Powered Photodetection Performance.
  • Dec 3, 2021
  • Inorganic Chemistry
  • You Zi + 9 more

  • Cite Count Icon 294
  • 10.1007/s10462-020-09904-8
A survey on video-based Human Action Recognition: recent updates, datasets, challenges, and applications
  • Sep 25, 2020
  • Artificial Intelligence Review
  • Preksha Pareek + 1 more

  • Open Access Icon
  • Cite Count Icon 170
  • 10.1109/iccv.2017.236
Lattice Long Short-Term Memory for Human Action Recognition
  • Oct 1, 2017
  • Lin Sun + 5 more

  • Cite Count Icon 152
  • 10.1002/smll.201900902
Enhanced Photodetection Properties of Tellurium@Selenium Roll-to-Roll Nanotube Heterojunctions.
  • Apr 24, 2019
  • Small
  • Weichun Huang + 13 more

  • Open Access Icon
  • Cite Count Icon 400
  • 10.1007/s11263-022-01594-9
Human Action Recognition and Prediction: A Survey
  • Mar 28, 2022
  • International Journal of Computer Vision
  • Yu Kong + 1 more

  • Open Access Icon
  • Cite Count Icon 604
  • 10.1109/iccv.2011.6126386
Human action recognition by learning bases of action attributes and parts
  • Nov 1, 2011
  • Bangpeng Yao + 5 more

  • Cite Count Icon 119
  • 10.1016/j.cogsys.2018.04.002
Deep learning approach for human action recognition in infrared images
  • Apr 17, 2018
  • Cognitive Systems Research
  • Aparna Akula + 2 more

  • Cite Count Icon 224
  • 10.1109/cvpr.2018.00054
MiCT: Mixed 3D/2D Convolutional Tube for Human Action Recognition
  • May 20, 2018
  • Yizhou Zhou + 3 more

  • Cite Count Icon 40
  • 10.1021/acsanm.3c02317
MXene V2CTx Nanosheet/Bismuth Quantum Dot-Based Heterostructures for Enhanced Flexible Photodetection and Nonlinear Photonics
  • Jul 13, 2023
  • ACS Applied Nano Materials
  • Jun Zhu + 8 more

  • Open Access Icon
  • Cite Count Icon 174
  • 10.1109/tpami.2017.2691768
Learning a Deep Model for Human Action Recognition from Novel Viewpoints.
  • Apr 6, 2017
  • IEEE Transactions on Pattern Analysis and Machine Intelligence
  • Hossein Rahmani + 2 more

Similar Papers
  • Research Article
  • Cite Count Icon 2
  • 10.1007/3dres.02(2013)3
A novel and stable human detection and behavior recognition method based on depth sensor
  • May 8, 2013
  • 3D Research
  • Shuqiang Yang + 1 more

To improve traditional video surveillance systems' performance on human behavior recognition, a new system has been built. Not visual camera but depth camera is chosen as sensor. To adapt to the most common forward oblique view of camera, a normalized digital elevation map, whose pixel intensity indicates the elevation of the scene, is built from the depth image. Coordinates and intensity in the digital elevation map represent the position information about the scene. Oriented templates are proposed to match and detect the human head robustly in the elevation map. As for the excellent visibility of human head in the elevation map, we track the human head to get trajectory. By combining the trajectory information of human head with the elevation map, several predefined human behaviors are recognized. Our behavior recognition method is straightforward and robust. This uniqueness has no similar with the traditional machine learning and classification framework about human behavior recognition.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 16
  • 10.1371/journal.pone.0124640
Uniform Local Binary Pattern Based Texture-Edge Feature for 3D Human Behavior Recognition.
  • May 5, 2015
  • PLOS ONE
  • Yue Ming + 2 more

With the rapid development of 3D somatosensory technology, human behavior recognition has become an important research field. Human behavior feature analysis has evolved from traditional 2D features to 3D features. In order to improve the performance of human activity recognition, a human behavior recognition method is proposed, which is based on a hybrid texture-edge local pattern coding feature extraction and integration of RGB and depth videos information. The paper mainly focuses on background subtraction on RGB and depth video sequences of behaviors, extracting and integrating historical images of the behavior outlines, feature extraction and classification. The new method of 3D human behavior recognition has achieved the rapid and efficient recognition of behavior videos. A large number of experiments show that the proposed method has faster speed and higher recognition rate. The recognition method has good robustness for different environmental colors, lightings and other factors. Meanwhile, the feature of mixed texture-edge uniform local binary pattern can be used in most 3D behavior recognition.

  • Research Article
  • Cite Count Icon 2
  • 10.1142/s0218001420560121
Video Human Behavior Recognition Based on ISA Deep Network Model
  • May 18, 2020
  • International Journal of Pattern Recognition and Artificial Intelligence
  • Xian Zhong + 3 more

Vision-based behavior recognition is the analysis and recognition of human behavior in video. It has been widely used in many aspects such as multimedia information retrieval, behavior monitoring, and robot perception. This paper uses the Independent Subspace Analysis (ISA) deep network model feature extraction method, which is based on the ISA model and neural network theory, and combines data preprocessing methods, [Formula: see text]-means clustering methods, and Support Vector Machine (SVM) classifiers to achieve video classification and identification of human behavior. The ISA-based deep network model feature extraction method is an unsupervised learning method that can obtain behavior characteristics with good invariance and characterization capabilities in video human behavior. The experiment was conducted on the basis of the Hollywood2 human behavior data set. This experiment was compared with other commonly used human behavior feature extraction and recognition methods. The experimental results validated the effectiveness and advantages of this method in the classification and recognition of human behavior.

  • Conference Article
  • 10.1117/12.2643962
A method for multi-target human behavior recognition in small and medium scenes
  • Dec 19, 2022
  • Tao Yang + 5 more

Aiming at the low accuracy of behavior recognition technology for multi-target human behavior recognition in small and medium scenes, a method for multi-target human behavior recognition in small and medium scenes is proposed. In this paper, YOLOv5 and DeepSort are used to detect, track and locate human targets in the video stream. According to the detection frame, the appropriate size of the human target is cropped as the input image of the behavior recognition module to reduce the interference of human behavior background, and finally realize the multi-target human body behavior recognition. The behavior recognition module is composed of an improved C3D network, and the features extracted by YOLOv5 are shared with the behavior recognition module to reduce the amount of computation. Experiments show that this method achieves end-to-end recognition,and can recognize the behavior of different target human bodies in small and medium scenes, and achieves comparable results.

  • Research Article
  • Cite Count Icon 15
  • 10.1016/j.procs.2022.01.045
Human behavior recognition model based on improved EfficientNet
  • Jan 1, 2022
  • Procedia Computer Science
  • Chen-Yun Luo + 3 more

Human behavior recognition model based on improved EfficientNet

  • Research Article
  • 10.3390/s24237595
Research on Multi-Scale Spatio-Temporal Graph Convolutional Human Behavior Recognition Method Incorporating Multi-Granularity Features.
  • Nov 28, 2024
  • Sensors (Basel, Switzerland)
  • Yulin Wang + 3 more

Aiming at the problem that the existing human skeleton behavior recognition methods are insensitive to human local movements and show inaccurate recognition in distinguishing similar behaviors, a multi-scale spatio-temporal graph convolution method incorporating multi-granularity features is proposed for human behavior recognition. Firstly, a skeleton fine-grained partitioning strategy is proposed, which initializes the skeleton data into data streams of different granularities. An adaptive cross-scale feature fusion layer is designed using a normalized Gaussian function to perform feature fusion among different granularities, guiding the model to focus on discriminative feature representations among similar behaviors through fine-grained features. Secondly, a sparse multi-scale adjacency matrix is introduced to solve the bias weighting problem that amplifies the multi-scale spatial domain modeling process under multi-granularity conditions. Finally, an end-to-end graph convolutional neural network is constructed to improve the feature expression ability of spatio-temporal receptive field information and enhance the robustness of recognition between similar behaviors. The feasibility of the proposed algorithm was verified on the public behavior recognition dataset MSR Action 3D, with a accuracy of 95.67%, which is superior to existing behavior recognition methods.

  • Research Article
  • Cite Count Icon 1
  • 10.1155/je/5604741
SABO–LSTM: A Novel Human Behavior Recognition Method for Wearable Devices
  • Jan 1, 2024
  • Journal of Engineering
  • Wei Zhang + 2 more

With the popularity of wearable devices, human behavior recognition technology is becoming increasingly important in social surveillance, health monitoring, smart home, and traffic management. However, traditional human behavior recognition methods rely too much on the subjective experience of managers in hyperparameter selection, resulting in an inefficient parameter optimization process. To address this problem, this paper proposes a long short–term memory (LSTM) neural network model based on a subtraction‐average–based optimizer (SABO) for human behavior recognition in wearable devices. Compared to the traditional method, the SABO–LSTM model significantly improves the recognition accuracy by automatically finding the optimal hyperparameters, which proves its innovation and superiority in practical applications. To demonstrate the effectiveness of the method, four evaluation metrics, including F1 score, precision, recall, and accuracy, are used to validate it on the UCI‐HAR dataset and the WISDM dataset, and control groups are introduced for comparison. The experimental results show that SABO–LSTM can accurately perform the human behavior recognition task with an accuracy of 98.84% and 96.37% on the UCI‐HAR dataset and the WISDM dataset, respectively. In addition, the experimental model outperforms the control model on all four evaluation metrics and outperforms existing recognition methods in terms of accuracy.

  • Conference Article
  • Cite Count Icon 5
  • 10.1109/icsess47205.2019.9040826
Human Behavior Recognition Algorithm Based on HOG Feature and SVM Classifier
  • Oct 1, 2019
  • Qing Cai

Human behavior analysis is a hot research in the field of computer vision. It has broad application prospects in the fields of intelligent monitoring, human-computer interaction, motion analysis and virtual reality. In order to improve the accuracy of human behavior recognition, a human behavior recognition method based on HOG feature and SVM classifier is proposed. First, the HOG features of the training set and the test set are extracted. Then, the multi-class problem is transformed into multiple dual-class problems, and multiple SVM classifiers are trained by using the HOG features. Finally, the trained classifiers are employed to recognize the human walking and waving behavior. Experimental results show that the recognition rate of walking and waving is 87.5% for the UIUC database. The human behavior recognition method proposed in this paper can effectively improve the recognition rate.

  • Conference Article
  • Cite Count Icon 5
  • 10.1109/iaeac47372.2019.8997727
Research on Human Motion Recognition System Based on MEMS Sensor Network
  • Dec 1, 2019
  • Hao Ma + 1 more

The traditional human motion perception and behavior recognition methods are based on machine vision and image processing. This method has the disadvantages of strong intrusion of individual data privacy, large environmental impact and tedious data processing, and cannot meet the practical requirements in real time. This paper introduced a solution of accurately capture human motion by using multi-sensor of MEMS to construct Zigbee wireless network, and utilizing SVM classification algorithm to realize human action recognition efficiently. The system is used to detect the abnormal human fall behavior, which is perfect reliability and real-time performance and the recognition accuracy of the action is above to 90%.

  • Research Article
  • Cite Count Icon 7
  • 10.1016/j.ijleo.2015.08.103
3D Human behavior recognition based on binocular vision and face–hand feature
  • Aug 29, 2015
  • Optik - International Journal for Light and Electron Optics
  • Qing Ye + 2 more

3D Human behavior recognition based on binocular vision and face–hand feature

  • PDF Download Icon
  • Research Article
  • 10.3390/bios14070337
Research on the Human Motion Recognition Method Based on Wearable.
  • Jul 10, 2024
  • Biosensors
  • Zhao Wang + 3 more

The accurate analysis of human dynamic behavior is very important for overcoming the limitations of movement diversity and behavioral adaptability. In this paper, a wearable device-based human dynamic behavior recognition method is proposed. The method collects acceleration and angular velocity data through a six-axis sensor to identify information containing specific behavior characteristics in a time series. A human movement data acquisition platform, the DMP attitude solution algorithm, and the threshold algorithm are used for processing. In this experiment, ten volunteers wore wearable sensors on their bilateral forearms, upper arms, thighs, calves, and waist, and movement data for standing, walking, and jumping were collected in school corridors and laboratory environments to verify the effectiveness of this wearable human movement recognition method. The results show that the recognition accuracy for standing, walking, and jumping reaches 98.33%, 96.67%, and 94.60%, respectively, and the average recognition rate is 96.53%. Compared with similar methods, this method not only improves the recognition accuracy but also simplifies the recognition algorithm and effectively saves computing resources. This research is expected to provide a new perspective for the recognition of human dynamic behavior and promote the wider application of wearable technology in the field of daily living assistance and health management.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 2
  • 10.1109/access.2022.3209816
Human Behavior Recognition Based on Multiscale Convolutional Neural Network
  • Jan 1, 2023
  • IEEE Access
  • Yan Chen

The key problem in human behavior recognition is how to build a spatiotemporal feature extraction and classification network. Aiming at the problem that the existing channel attention mechanism directly pools the global average information of each channel and ignores its local spatial information, this paper proposes two improved channel attention modules, namely the space-time (ST) interaction module of matrix operation and the depth separable convolution module, combined with the research of human behavior recognition. Combined with the superior performance of convolutional neural network (CNN) in image and video processing, a multi-scale convolutional neural network method for human behavior recognition is proposed. Firstly, the behavior video is segmented, and low rank learning is performed on each video segment to extract the corresponding Low rank behavior information, and then these Low rank behavior information are connected on the time axis to obtain the Low rank behavior information of the whole video, so as to effectively capture the behavior information in the video, avoiding tedious extraction steps and various assumptions. The ability of neural network to model human behavior can be transferred and reused in networks with different structures. According to the different characteristics of data features at different network levels, two effective feature difference measurement functions are introduced to reduce the difference between features extracted from different network structures. Experiments on several public datasets show that the proposed method has a good classification effect. The experimental results show that the method has a good accuracy in human behavior recognition. It is proved that the proposed model not only improves the recognition accuracy, but also effectively reduces the computational complexity of output weights and improves the compactness of the model structure.

  • Research Article
  • Cite Count Icon 4
  • 10.4028/www.scientific.net/amr.765-767.2603
Weighted Modified Hu Moment in Human Behavior Recognition
  • Sep 1, 2013
  • Advanced Materials Research
  • Chen Hua Liang + 1 more

t has been shown that the traditional seven Hu invariant moment does not have scaling invariance with low recognition rate in human behavior recognition. In order to improve the recognition rate, a human behavior recognition method will be put forward in this paper based on weighted modified Hu moments. Firstly, the traditional seven Hu moments will be extended to ten Hu moments to get more image details. Then, the extended Hu moments will be modified to make the Hu moments has the feature of scaling invariance. Lastly, the weighted modified Hu moment will be obtained through least squares method based on minimum variance criterion. The simulation of the sequence images shows that the weighted modified Hu moment can improve the recognition rate effectively.

  • Book Chapter
  • Cite Count Icon 3
  • 10.1007/978-1-4615-1003-1_4
Recognizing Human Behavior Using Hidden Markov Models
  • Jan 1, 2002
  • Junji Yamato

This chapter describes a new human behavior recognition method based on Hidden Markov Models (HMM). We use a feature-based bottom-up approach using HMM, which can provide a learning capability and time-scale invariability. To apply HMM to our aim, time sequential images are transformed to an image feature vector sequence, and the sequence is converted to a symbol sequence by Vector Quantization. In learning human behavior categories, the model parameters of HMM are optimized so as to best describe training sequences. For recognition, the model that best matches the observed sequence is chosen. A new VQ method and HMM configuration for behavior recognition is proposed and evaluated. Experimental results of real time-sequential images of sports scenes show a higher than 90% recognition rate. We also describe an example application of this behavior recognition method to content-based video database retrieval.

  • Research Article
  • Cite Count Icon 27
  • 10.1016/j.comcom.2020.01.012
Improved 1D-CNNs for behavior recognition using wearable sensor network
  • Jan 8, 2020
  • Computer Communications
  • Zhiou Xu + 3 more

Improved 1D-CNNs for behavior recognition using wearable sensor network

More from: International Journal of Crowd Science
  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2024.9100025
Monitoring Changes to Small-Sized Lakes Using Medium Spatial and Temporal Satellite Imagery in the Badain Jaran Desert from 2015 to 2020
  • May 1, 2025
  • International Journal of Crowd Science
  • Qinyu Zhao + 4 more

  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2024.9100007
B5G Applications and Emerging Services in Smart IoT Environments
  • May 1, 2025
  • International Journal of Crowd Science
  • Hani Attar + 5 more

  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2023.9100004
Predictive Emojis for User Cognitive Response to Dynamic Behaviour of Webpages Using Pupil and Click Reinforcement
  • May 1, 2025
  • International Journal of Crowd Science
  • Fatima Isiaka

  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2024.9100014
Lightweight Human Behavior Recognition Method for Visual Communication AGV Based on CNN-LSTM
  • May 1, 2025
  • International Journal of Crowd Science
  • Shuhua Zhao + 4 more

  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2023.9100019
Remote Monitoring System of Patient Status in Social IoT Environments Using Amazon Web Services Technologies and Smart Health Care
  • May 1, 2025
  • International Journal of Crowd Science
  • Amer Tahseen Abu-Jassar + 5 more

  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2024.9100016
Survey on AI-Assisted Power Transmission Line Fault Detection
  • May 1, 2025
  • International Journal of Crowd Science
  • Yue Zhang + 2 more

  • Research Article
  • 10.26599/ijcs.2023.9100015
CGLS Method for Efficient Equalization of OFDM Systems Under Doubly Dispersive Fading Channels with an Application Into 6G Communications in Smart Overcrowded
  • Jan 1, 2025
  • International Journal of Crowd Science
  • Issa Haitham M Ata

  • Open Access Icon
  • Research Article
  • Cite Count Icon 2
  • 10.26599/ijcs.2023.9100022
Improving Energy Harvesting System from Ambient RF Sources in Social Systems with Overcrowding
  • Jan 1, 2025
  • International Journal of Crowd Science
  • Ramy Agieb + 5 more

  • Research Article
  • 10.26599/ijcs.2023.9100018
Development and Investigation of Vision System for a Small-Sized Mobile Humanoid Robot in a Smart Environment
  • Jan 1, 2025
  • International Journal of Crowd Science
  • Amer Tahseen Abu-Jassar + 5 more

  • Open Access Icon
  • Research Article
  • 10.26599/ijcs.2024.9100045
Editorial of Cyber-Physical Social Systems and Smart Environments
  • Jan 1, 2025
  • International Journal of Crowd Science
  • Mohamadreza Khosravi + 1 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon