Constructing an accurate representation model of phenomena with fewer measurements is a fundamental challenge in the Internet of Things. Leveraging sparse sensing policies to select the most informative measurements is a prominent technique for addressing resource constraints. However, designing such sensing policies requires significant domain knowledge and involves manually fine-tuned heuristics that are task-specific and often non-adaptive. In this work, we propose reducing manual-engineering efforts in designing sensing policies by using an automated approach based on deep reinforcement learning. Guided by an uncertainty-aware prediction model, the sensors learn sensing behaviors autonomously by optimizing an application goal formulated in the reward function based on the measured peaks-over-threshold. We apply the proposed approach in two use cases of monitoring air quality and indoor noise and show the adaptability and transferability of the learned policies. Compared to conventional periodic sensing methods, our results achieve, on average, an increased detection in periods of interest by 78.5% and 357.3% while reducing energy expenditure by 14.3% and 7.6% for air quality and noise monitoring, respectively. Additionally, the resulting representation models are more credible, as measured by various metrics of probabilistic modeling.
Read full abstract