Abstract

The capability of efficiently processing the data streams emitted by nowadays ubiquitous sensing devices enables the development of new intelligent services. Data Stream Processing (DSP) applications allow for processing huge volumes of data in near real-time. To keep up with the high volume and velocity of data, these applications can elastically scale their execution on multiple computing resources to process the incoming data flow in parallel. Being that data sources and consumers are usually located at the network edges, nowadays the presence of geo-distributed computing resources represents an attractive environment for DSP. However, controlling the applications and the processing infrastructure in such wide-area environments represents a significant challenge. In this paper, we present a hierarchical solution for the autonomous control of elastic DSP applications and infrastructures. It consists of a two-layered hierarchical solution, where centralized components coordinate subordinated distributed managers, which, in turn, locally control the elastic adaptation of the application components and deployment regions. Exploiting this framework, we design several self-adaptation policies, including reinforcement learning based solutions. We show the benefits of the presented self-adaptation policies with respect to static provisioning solutions, and discuss the strengths of reinforcement learning based approaches, which learn from experience how to optimize the application performance and resource allocation.

Highlights

  • Nowadays, sensing devices are widely diffused in our everyday environment, being embedded in human-carried mobile devices, Internet-of-Things (IoT), as well as monitoring systems for public utilities, transportation, and facilities

  • A Data Stream Processing (DSP) application can be regarded as a directed acyclic graph (DAG), where data sources, operators, and consumers are connected by streams An operator is a self-contained processing element that carries out a specific operation, whereas a stream is an unbounded sequence of data

  • In this set of experiment, we investigated the flexibility of our Reinforcement Learning (RL)-based policy in optimizing different trade-offs among resource usage cost and application requirements satisfaction

Read more

Summary

Introduction

Nowadays, sensing devices are widely diffused in our everyday environment, being embedded in human-carried mobile devices, Internet-of-Things (IoT), as well as monitoring systems for public utilities, transportation, and facilities. Since the operator replicas run on computing resources, modern DSP systems should be able to dynamically adjust the set of computing resources to satisfy the application resource demand while avoiding costly deployment It results a complex system which should quickly control elasticity at multiple levels, namely at the application and at the infrastructure level (e.g., [5]). E2DF consists of loosely coupled components that interact to realize the multi-level elasticity It includes the Application Control System (ACS), which manages the elasticity of DSP applications, and the Infrastructure Control System (ICS), which can dynamically acquire and release computing resource for the framework. Our simulation results show the benefits of having two separate control components that autonomously adapt the deployment of DSP applications on a dynamic set of computing resources They demonstrate the flexibility of the proposed infrastructure-level policy based on RL, which can be tuned to optimize different deployment objectives while still supporting the application-level elasticity.

Related Work
System Architectures
Elasticity Policies
DSP Application Model
Infrastructure Model
Problem Definition
Hierarchical System Architecture
Infrastructure Control System
Application Control System
Multi-Level Adaptation Policy
Infrastructure Control Policy
Local Policy
Learning the Optimal Policy
Model-Based Reinforcement Learning
Global Policy
Application Control Policy
Evaluation
Simulation Setup
Results with Different Combinations of Policies
Results with Different Objectives
Considerations on the Initial Learning Phase
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call