Conference Paper

A Distributed Autonomic Management Framework for Cloud Computing Orchestration

Authors:
To read the full-text of this research, you can request a copy directly from the authors.

Abstract

Abstract—Due to constant workload growth, the infrastructure used to support cloud computing (CC) environments increases in size and complexity. As a consequence of that, human administrators are not able to monitor, analyze, plan and execute actions upon the environment, seeking goals such as the environment optimization and service level agreements fulfillment. This proposal provides an autonomic framework to create virtual machines migrations and heuristics to select hosts to be activated or deactivated when needed. Moreover, the framework proposed in this paper works in a distributed way using multi-agent systems concepts. We provide an architecture to deal with the size, heterogeneity and dynamism of CC environments. Further, our proposal was added to the CloudStack platform as a plug-in for validation and experimentation. Keywords-Cloud computing orchestration; autonomic management framework.

No full-text available

Request Full-text Paper PDF

To read the full-text of this research,
you can request a copy directly from the authors.

... Multitask applications utilizing dynamic resource orchestration in varied mobile cloud computing, which realizes a processing resource provision for varied computing infrastructures, was conducted as a multi-purpose optimization problem depending on different metrics [39]. A distributed virtual machines (VMs) cloud computing orchestration framework that provides a self-adaptive framework to create VMs migrations to select hosts to be running or stopped, when necessary was implemented in [40]. Additionally, polyphony was presented in [41]; polyphony is a robust, scalable, and flexible framework that efficiently controls a large set of computing resources to perform parallel computations which can employ resources on the cloud and additional capacity on local machines, as well as saving resources on the supercomputing center. ...
... Multitask applications utilizing dynamic resource orchestration in varied mobile cloud computing, which realizes a processing resource provision for varied computing infrastructures, was conducted as a multi-purpose optimization problem depending on different metrics [39]. A distributed virtual machines (VMs) cloud computing orchestration framework that provides a self-adaptive framework to create VMs migrations to select hosts to be running or stopped, when necessary, was implemented in [40]. Additionally, polyphony was presented in [41]; polyphony is a robust, scalable, and flexible framework that efficiently controls a large set of computing resources to perform parallel computations which can employ resources on the cloud and additional capacity on local machines, as well as saving resources on the supercomputing center. ...
Article
Full-text available
IoT is a trending computational concept that converts almost everything in modern life into a smart thing in various innovational and outstanding approaches. Smart homes, connected cities, autonomous vehicles, industrial automation, and smart healthcare that allows doctors to perform a patient examination and enable executing a remote surgery are now applicable through the smart connected things. Moreover, the recent IoT analytics report expects the universal number of connected IoT things to grow by 9%, to 12.3 billion operating terminals. It is expected that there will be more than 27 billion IoT live connections by 2025. In this paper, we present osmotic message-oriented middleware, introducing an end-to-end IoT platform to federate the dynamic orchestration process of resources across different heterogeneous types of devices belonging to physical and virtual infrastructures (e.g., edge, fog, and cloud layers); the orchestration process follows the osmotic computing concepts represented as the self-adaptive MAPE-K model, which maintains/adopts itself on the runtime through feedback loops from the provisioning engine, which collects the node’s hardware and software performance matrices. Accordingly, the orchestration process utilizes the optimized dynamic Hungarian algorithm to solve the MELs’ assignment problem based on the vibrant runtime provisioning data. The implemented middleware prototype is tested on both simulated and real-life environments to validate the architecture hypothesis of running an efficient, robust, elastic, and cost-efficient end-to-end osmotic IoT ecosystem, which unlocks a new implementation model for the IoT numerous domains.
... Here, the resource orchestration was formulated as multi-objective optimization problem considering energy consumption, cost, and availability metrics. A distributed framework for Cloud Computing orchestration able to manage the migration of Virtual Machines (VMs) was discussed in [28]. Platforms able to control high performance scientific workflow by means of Cloud applications were discussed in [29], [30] and [31]. ...
Article
Full-text available
Nowadays, the latest technological advancements have changed the centralized Cloud Computing model, going through Edge and Internet of Things (IoT), which are closer to end users. In particular, current Cloud Computing programming models deal with the recent evolution of the IoT phenomenon because smart devices are becoming more and more pervasive, powerful and inexpensive. Therefore, services need to be placed near such devices and resource orchestration techniques need to be redesigned. In this regard, Osmotic Computing aims at providing a new computing paradigm based on the deployment and migration strategies related to the infrastructures and applications requirements across Cloud, Edge, and IoT layers. The objective of this scientific work is to propose an Osmotic Computing architecture, based on a multi-agent system, according to a new software abstraction called MicroELement (MEL), that encapsulates resources, services and data necessary to run IoT applications. A Body Area Network (BAN) scenario is proposed in order to explore the Osmotic Computing potentiality and explain the reasons behind this new paradigm.
... Here, the resource orchestration was formulated as multi-objective optimal problem considering energy consumption, cost and availability metrics. A Distributed pieces of framework for cloud computing orchestration able to manage the migration of Virtual Machines (VMs) were discussed in [20]. Platforms able to control high performance scientific workflow by means of computing Cloud applications were discussed in [21], [22] and [23]. ...
... The work presented in [20] proposed a distributed autonomic management framework for cloud computing orchestration. The targets of the research work are: 1) Dealing with the heterogeneity, dynamism and stochastic nature of the cloud. ...
... The constant growth of networks, the management of hidden traffic in NFV and the need for scalable architectures are challenges faced by system administrators and operators (Westphall, 2016). The implementation of autonomic management within the systems helps to address these challenges, automating and distributing processes within the network. ...
Article
Full-text available
The implementation of NFV allows improving the flexibility, efficiency, and manageability of networks by leveraging virtualization and cloud computing technologies to deploy computer networks. The implementation of autonomic management and supervised algorithms from Machine Learning (ML) become a key strategy to manage this hidden traffic. In this work, we focus on analyzing the traffic features of NFV-based networks while performing a benchmarking of the behavior of supervised ML algorithms, namely J48, Naïve Bayes, and Bayes Net, in the IP traffic classification regarding their efficiency; considering that such an efficiency is related to the trade-off between time-response and precision. We used two test scenarios (an NFV-based SDN and an NFV-based LTE EPC). The benchmarking results reveal that the Naïve Bayes and Bayes Net algorithms achieve the best performance in traffic classification. In particular, their performance corroborates a good trade-off between precision and time-response, with precision values higher than 80 % and 96 %, respectively, in times less than 1,5 sec.
Book
Full-text available
Os principais problemas associados à implementação e uso da gerência de redes e serviços ocorrem devido à grande quantidade de proposições, padrões e de diferentes produtos oferecidos no mercado, dificultando consideravelmente a tomada de decisão no que se refere a utilização da abordagem de gerência de redes e serviços mais adequada. Além disso, novas tendências na área de gerência de redes e serviços vêm sendo pesquisadas, entre estas destacam-se atualmente: gerência de redes sem fio, de sensores, óticas, futura internet, internet das coisas, internet espacial...; áreas funcionais de segurança, configuração, desempenho, contabilidade...; gerência de serviços de multimídia, data centers, grid, cloud, fog, edge virtualização...; e gerência centralizada, autonômica, distribuída, auto-gerência, baseada em políticas... Estas novas tendências vêm sendo pesquisadas no Laboratório de Redes e Gerência (LRG) da UFSC e a partir deste projeto as mesmas poderão ser aperfeiçoadas através das seguintes atividades deste projeto: A - Aperfeiçoamentos na Gerência Autonômica para Fog e IoT; B - Aperfeiçoamentos na Qualidade de Serviço para Aplicações de Tempo Real em IoT e Fog; C Aperfeiçoamentos na Segurança para Fog e IoT; D - Aperfeiçoamentos no Sistema de Resposta de Intrusão Autonômica em Cloud e IoT; E - Aperfeiçoamentos na Privacidade em Gerência de Identidade para Federações Dinâmicas em Cloud e IoT; e F - Aperfeiçoamentos no Controle de Acesso Dinâmico Baseado em Risco para uma Federação de Nuvem e IoT..
Conference Paper
Full-text available
Quality of service is one of the major concerns in cloud computing. Virtual machines (VMs) balancing techniques can help reduce service degradation in cloud computing environments. Several works have presented cloud computing balance techniques; however, only a few used the similarity between VMs and physical hosts to map VMs migrations. In addition, most proposals do not consider the size, dynamism, and heterogeneity of the cloud when developing a management technique. We present a cloud computing VMs balancing algorithm that uses the similarity between VMs and physical hosts to create the map of migrations. Furthermore, the proposal takes into account the size, dynamism, and heterogeneity of the cloud when mapping VMs migrations; thus the proposal is developed in a distributed fashion, enabling the processing of each cluster at a time. To evaluate the proposal, we used the Google cluster data set. Experiments demonstrate that the proposed technique can improve the balance of allocated resources; thus helping reduce service degradation. Moreover, the runtime of the algorithm indicates that it is feasible to be used in a real cloud computing environment with hundreds of physical servers and virtual machines.
Article
Full-text available
A promising solution for the management of services in clouds, as fostered by autonomic computing, is to resort to self-management. However, the obfuscation of underlying details of services in cloud computing, also due to privacy requirements, affects the effectiveness of autonomic managers. Data-driven approaches, in particular those relying on service clustering based on machine learning techniques, can assist the autonomic management and support decisions concerning, e.g. the scheduling and deployment of services. Unfortunately, applying such approaches is further complicated by the coexistence of different types of data within the information provided by the monitoring of cloud systems: both continuous (e.g. CPU load) and categorical (e.g. VM instance type) data are available. Current approaches deal with this problem in a heuristic fashion. In this paper, instead, we propose an approach that uses all types of data, and learns in a data-driven fashion the similarities and patterns among the services. More specifically, we design an unsupervised formulation of Random Forest to calculate service similarities and provide them as input to a clustering algorithm. For the sake of efficiency and to meet the dynamism requirement of autonomic clouds, our methodology consists of two steps: off-line clustering and on-line prediction. Using datasets from real-world clouds, we demonstrate the superiority of our solution with respect to others and validate the accuracy of the on-line prediction. Moreover, to show applicability of our approach, we devise a service scheduler that uses similarity among services, and evaluate its performance in a cloud test-bed using realistic data.
Conference Paper
Full-text available
Virtual machine (VM) consolidation is an emerging solution for energy saving in cloud data centers. An appropriate VM consolidation policy must change accordingly, depending on the situation. To easily deploy a VM consolidation solution, this paper proposes a VM consolidation framework for Apache CloudStack, which is a popular open source cloud platform software suite. We separate the VM consolidation mechanism from the policy used to decide which VMs should be moved to somewhere more appropriate to run. This framework is a modular architecture that combines various resource monitoring systems, power control mechanisms, and VM packing algorithms. Using the proposed framework, we have demonstrated that a web application can consolidate and deconsolidate VMs so as to balance the load of CPU utilization among physical machines in response to the number of physical machines in operation. Experimental results reveal the proposed framework is preliminary, but provides the required functionality.
Article
Full-text available
Internet hosting centers serve multiple service sites from a common hardware base. This paper presents the design and implementation of an architecture for resource management in a hosting center operating system, with an emphasis on energy as a driving resource management issue for large server clusters. The goals are to provision server resources for co-hosted services in a way that automatically adapts to offered load, improve the energy efficiency of server clusters by dynamically resizing the active server set, and respond to power supply disruptions or thermal events by degrading service in accordance with negotiated Service Level Agreements (SLAs).Our system is based on an economic approach to managing shared server resources, in which services "bid" for resources as a function of delivered performance. The system continuously monitors load and plans resource allotments by estimating the value of their effects on service performance. A greedy resource allocation algorithm adjusts resource prices to balance supply and demand, allocating resources to their most efficient use. A reconfigurable server switching infrastructure directs request traffic to the servers assigned to each service. Experimental results from a prototype confirm that the system adapts to offered load and resource availability, and can reduce server energy usage by 29% or more for a typical Web workload.
Article
Full-text available
Many researchers have demonstrated that the organizational design employed by an agent system can have a significant, quantitative effect on its performance characteristics. A range of organizational strategies have emerged from this line of research, each with different strengths and weaknesses. In this article we present a survey of the major organizational paradigms used in multi-agent systems. These include hierarchies, holarchies, coalitions, teams, congregations, societies, federations, markets, and matrix organizations. We will provide a description of each, discuss their advantages and disadvantages, and provide examples of how they may be instantiated and maintained. This summary will facilitate the comparative evaluation of organizational styles, allowing designers to first recognize the spectrum of possibilities, and then guiding the selection of an appropriate organizational design for a particular domain and environment.
Conference Paper
Full-text available
Virtualization is being widely used in large-scale computing environments, such as clouds, data centers, and grids, to provide application portability and facilitate resource multiplexing while retaining application isolation. In many exist- ing virtualized platforms, it has been found that the network bandwidth often becomes the bottleneck resource due to the hierarchical topology of the underlying network, causing both high network contention and reduced performance for com- munication and data-intensive applications. In this paper, we present a decentralized affin ity-aware migration technique that incorporates heterogeneity and dynamism in network topology and job communication patterns to allocate virtual machines on the available physical resources. Our technique monitors network affin ity between pairs of VMs and uses a distributed bartering algorithm, coupled with migration, to dynamically adjust VM placement such that communication overhead is minimized. Our experimental results running the Intel MPI benchmark and a scientifi c application on an 8-node Xen cluster show that we can get up to 42% improvement in the runtime of the application over a no-migration technique, while achieving up to 85% reduction in network communication cost. In addition, our technique is able to adjust to dynamic variations in communication patterns and provides both good performance and low network contention with minimal overhead.
Article
Full-text available
CLOUD COMPUTING, the long-held dream of computing as a utility, has the potential to transform a large part of the IT industry, making software even more attractive as a service and shaping the way IT hardware is designed and purchased. Developers with innovative ideas for new Internet services no longer require the large capital outlays in hardware to deploy their service or the human expense to operate it. They need not be concerned about overprovisioning for a service whose popularity does not meet their predictions, thus wasting costly resources, or underprovisioning for one that becomes wildly popular, thus missing potential customers and revenue. Moreover, companies with large batch-oriented tasks can get results as quickly as their programs can scale, since using 1,000 servers for one hour costs no more than using one server for 1,000.
Article
Full-text available
A 2001 IBM manifesto observed that a looming software complexity crisis -caused by applications and environments that number into the tens of millions of lines of code - threatened to halt progress in computing. The manifesto noted the almost impossible difficulty of managing current and planned computing systems, which require integrating several heterogeneous environments into corporate-wide computing systems that extend into the Internet. Autonomic computing, perhaps the most attractive approach to solving this problem, creates systems that can manage themselves when given high-level objectives from administrators. Systems manage themselves according to an administrator's goals. New components integrate as effortlessly as a new cell establishes itself in the human body. These ideas are not science fiction, but elements of the grand challenge to create self-managing computing systems.
Article
Virtualization technologies changed the way data centers of enterprises utilize their server resources. Instead of using dedicated servers for each type of application, virtualization allows viewing resources as a pool of unified resources, thereby reducing complexity and easing manageability. Server consolidation technique, which deals with reducing the number of servers used by consolidating applications, is one of the main applications of virtualization in data centers. The latter technique helps to use computing resources more effectively and has many benefits, such as reducing costs of power, cooling and, hence, contributes to the Green IT initiative. In a dynamic data center environment, where applications encapsulated as virtual machines are mapped to and released from the nodes frequently, reducing the number of server nodes used can be achieved by migrating applications without stopping their services, the technology known as live migration. However, live migration is a costly operation; hence, how to perform periodic server consolidation operation in a migration-aware way is a challenging task. We propose server consolidation algorithm - Sercon, which not only minimizes the overall number of used servers, but also minimizes the number of migrations. We verify the feasibility of our algorithm along with showing its scalability by conducting experiments with eight different test cases.
Article
With the rise of cloud computing, a huge complexity growth of the structure that is the base of the Cloud happens. Thus, to effectively manage applications and resources it is crucial the use of models and tools that create an application profile which is used to apply forecasting models to determine the most suitable amount of resource for each workload. There are models and tools that address the creation of an application profile to later apply some forecasting technique and estimate the amount of resource needed for a workload. Therefore, this article aims to present a taxonomy for application profiling models and tools, presenting its main characteristics, challenges, describing and comparing such models and tool. At the end this work presents a discussion about the use of application profiling and its future research trends.
Article
In this paper, we present techniques for provisioning CPU and network resources in shared hosting platforms running potentially antagonistic third-party applications. The primary contribution of our work is to demonstrate the feasibility and benefits of overbooking resources in shared platforms. Since an accurate estimate of an application's resource needs is necessary when overbooking resources, we present techniques to profile applications on dedicated nodes, possibly while in service, and use these profiles to guide the placement of application components onto shared nodes. We then propose techniques to overbook cluster resources in a controlled fashion such that the platform can provide performance guarantees to applications even when overbooked. We show how our techniques can be combined with commonly used QoS resource allocation mechanisms to provide application isolation and performance guarantees at run-time. We implement our techniques in a Linux cluster and evaluate them using common server applications. We find that the efficiency (and consequently revenue) benefits from controlled overbooking of resources can be dramatic. Specifically, we find that overbooking resources by as little as 1% we can increase the utilization of the cluster by a factor of two, and a 5% overbooking yields a 300-500% improvement, while still providing useful resource guarantees to applications.
Opportunities for csps in enterprise-grade public cloud computing
  • P Hall
System x virtualization strategies
  • B Archer
  • M Ausilio
  • W Bauman
  • F Ferri
  • C Kuhlmann
  • M Madduri
  • M Mariani
  • R Mente
  • J Patterson
What is apache cloudstack?
  • A Cloudstack
Xenserver 6.5: Configuration limits
  • Citrix
Energy-efficient management of virtual machines in data centers for cloud computing
  • A Beloglazov
Hp helion eucalyptus overview
  • H H Eucalyptus
Hypervisors, virtualization, and the cloud: Learn about hypervisors, system virtualization, and how it works in a cloud environment
  • B P Tholeti
Xen project software overview
  • X Project