Forest fires pose a significant threat to ecosystems, property, and human life, making their early and accurate detection crucial for effective intervention. This study presents a novel, lightweight approach to real-time forest fire detection that is optimized for resource-constrained devices like drones. The method integrates multi-task knowledge distillation, transferring knowledge from a high-performance DenseNet201 teacher model that was trained on a hierarchically structured wildfire dataset. The dataset comprised primary classes (fire vs. non-fire) and detailed subclasses that account for confounding elements such as smoke, fog, and reflections. The novelty of this approach lies in leveraging knowledge distillation to transfer the deeper insights learned by the DenseNet201 teacher model—specifically, the auxiliary task of recognizing the confounding elements responsible for false positives—into a lightweight student model, enabling it to achieve a similar robustness without the need for complex architectures. Using this distilled knowledge, we trained a MobileNetV3-based student model, which was designed to operate efficiently in real-time while maintaining a low computational overhead. To address the challenge of false positives caused by visually similar non-fire elements, we introduced the Confounding Element Specificity (CES) metric. This novel metric, made possible by the hierarchical structure of the wildfire dataset, is unique in its focus on evaluating how well the model distinguishes actual fires from the confounding elements that typically result in false positives within the negative class. The proposed approach outperformed the baseline methods—including single-task learning and direct multi-task learning—achieving a primary accuracy of 93.36%, an F1-score of 91.57%, and a higher MES score, demonstrating its enhanced robustness and reliability in diverse environmental conditions. This work bridges the gap between advanced deep learning techniques and practical, scalable solutions for environmental monitoring. Future research will focus on integrating multi-modal data and developing adaptive distillation techniques to further enhance the model’s performance in real-time applications.
Read full abstract