Abstract
Ambient Intelligence (AmI) encompasses technological infrastructures capable of sensing data from environments and extracting high-level knowledge to detect or recognize users’ features and actions, as well as entities or events in their surroundings. Visual perception, particularly object detection, has become one of the most relevant enabling factors for this context-aware user-centered intelligence, being the cornerstone of relevant but complex tasks, such as object tracking or human action recognition. In this context, convolutional neural networks have proven to achieve state-of-the-art accuracy levels. However, they typically result in large and highly complex models that typically demand computation offloading onto remote cloud platforms. Such an approach has security- and latency-related limitations and may not be appropriate for some AmI use cases where the system response time must be as short as possible, and data privacy must be guaranteed. In the last few years, the on-device paradigm has emerged in response to those limitations, yielding more compact and efficient neural networks able to address inference directly on client machines, thus providing users with a smoother and better-tailored experience, with no need of sharing their data with an outsourced service. Framed in that novel paradigm, this work presents a review of the recent advances made along those lines in object detection, providing a comprehensive study of the most relevant lightweight CNN-based detection frameworks, discussing the most paradigmatic AmI domains where such an approach has been successfully applied, the different challenges arisen, the key strategies and techniques adopted to create visual solutions for image-based object classification and localization, as well as the most relevant factors to bear in mind when assessing or comparing those techniques, such as the evaluation metrics or the hardware setups used.
Highlights
The analysis of on-device models goes beyond the mere evaluation of detector performance, covering costspecific aspects, which are fundamental to assessing the feasibility of such solutions in resource-limited devices
On the contrary, speed and accuracy are addressed in virtually all the ambient intelligence (AmI) studies analyzed [48,52,53,54,55,56,57,59,60,61,62,64,65,66,67,69,70,71,74,75,76,77,78,79,82,83,84,85,87,90,91,92,93,95,96,98,100,101,102,103,105,106,107,109], revealing a trend consistent with the experimentation traditionally conducted for evaluating conventional machine learning (ML) and deep learning (DL) techniques
As far as accuracy is concerned, the related metrics that emerge in Table 4 as the most popular options are average precision (AP) and its variant, the mean average precision [50,52,54,55,56,57,58,59,60,61,62,63,64,70,74,75,78,79,80,82,83,84,85,86,89,90,91,92,93,96,97,98,102,103,104,105,107,109,110], distantly followed by precision (P) [55,65,68,69,89,96,97,99,102,103,106,109], recall (R) [55,65,68,89, 96,97,99,102,103,106,109], and F1 score [55,66,76,80,91,96,97,107,109]
Summary
Today ambient intelligence (AmI) systems are experiencing an unprecedented growth momentum essentially due to the thrust and fast development of two of their main enabling technological forces: the so-called Internet of Things (IoT), focused on the exploitation of networked sensor infrastructures to remotely gather data and enable data exchange between several distributed ends and Artificial Intelligence (AI), more oriented to the use of gathered data and the subsequent distillation of knowledge necessary to make AmI systems adaptable and “aware” of their surroundings Both fields, IoT and AI have experienced outstanding progress almost side by side in the last two decades, a fact in no case fortunate, but rather the result of a continuous interaction that is still going on nowadays
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.