Abstract

Facing direction detection plays a critical role in human–computer interaction and has a wide application in surveillance systems, driving awareness recognition, smart home appliances, computer games, and so on. Current detection methods are mainly focused on extracting specific patterns from user’s optical images, which raises concerns on privacy invasion and these detection techniques do not usually work in a dark environment. To address these concerns, this paper proposes an activity recognition system guided by an unobtrusive sensor (ARGUS). By using a low pixel infrared thermopile array sensor, ARGUS is capable of identifying five facing directions (left 45°/90°, right 45°/90°, and front) through the support vector machine classifier. Also two feature extraction methods are compared. One is manually-defined and the other is based on a pre-trained convolutional neural network (CNN) model. The facing direction detection accuracy resulting from manually-defined features reaches 85.3%, 90.6%, and 85.2% at detection distance of 0.6, 1.2, and 1.8 m, respectively. The level of accuracy resulted from using pre-trained CNN features demonstrates a much more reliable performance (89.1%, 95.3%, and 95.1% at distance of 0.6, 1.2, and 1.8 m, respectively). In addition, ARGUS has been successfully applied for occupancy tracking with a root mean square error of 0.19 m.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.