Abstract

The Internet of Things (IoT) is constantly growing, generating an uninterrupted data stream pipeline to monitor physical world information. Hence, Artificial Intelligence (AI) continuously evolves, improving life quality and business and academic activities. Kafka-ML is an open-source framework that focuses on managing Machine Learning (ML) and AI pipelines through data streams in production scenarios. Consequently, it facilitates Deep Neural Network (DNN) deployments in real-world applications. However, this framework does not consider the distribution of DNN models on the Cloud-to-Things Continuum. Distributed DNN significantly reduces latency, allocating the computational and network load between different infrastructures. In this work, we have extended our Kafka-ML framework to support the management and deployment of Distributed DNN throughout the Cloud-to-Things Continuum. Moreover, we have considered the possibility of including early exits in the Cloud-to-Things layers to provide immediate responses upon predictions. We have evaluated these new features by adapting and deploying the DNN model AlexNet in three different Cloud-to-Things scenarios. Experiments demonstrate that Kafka-ML can significantly improve response time and throughput by distributing DNN models throughout the Cloud-to-Things Continuum, compared to a Cloud-only deployment.

Highlights

  • O Ver the last years, a great number of sources and fields have persistently procured tons of data [1]

  • Machine Learning (ML) and Artificial Intelligence (AI) [2] algorithms have assumed an essential role in their processing and allowed us to rely on practical tools in this computerized era

  • We developed KafkaML [15], an open-source1 framework to manage ML/AI pipelines through data streams

Read more

Summary

INTRODUCTION

O Ver the last years, a great number of sources and fields have persistently procured tons of data [1]. ML/AI algorithms, especially Deep Neural Networks (DNN), usually need a large number of computational resources and specialized hardware (e.g., GPUs), which may not satisfy the requirements of resource-constrained systems For this reason, DNN are partitioned along the Cloud-to-Things Continuum in order to have smaller processing units (and sub-models) distributed. A possible deployment in this context could place IoT devices generating data stream connected to edge devices, edge devices connected to fog servers for intermediary processing before sending the information generated to the Cloud These processing units can allocate DNN, whose sub-models may present less accuracy at the lowest layers, but they do reduce the response time [18].

RELATED WORK
TRAINING DISTRIBUTED MODELS
Result
INFERENCE OF DISTRIBUTED MODELS
D4 D3 D2 D1
EVALUATION
Brokers down
VIII. CONCLUSION

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.