Transiency Driven Resource Management For Cloud Computing Platforms PDF Download

Are you looking for read ebook online? Search for your book and save it on your Kindle device, PC, phones or tablets. Download Transiency Driven Resource Management For Cloud Computing Platforms PDF full book. Access full book title Transiency Driven Resource Management For Cloud Computing Platforms.

Transiency-driven Resource Management for Cloud Computing Platforms

Transiency-driven Resource Management for Cloud Computing Platforms
Author: Prateek Sharma
Publisher:
Total Pages:
Release: 2018
Genre:
ISBN:

Download Transiency-driven Resource Management for Cloud Computing Platforms Book in PDF, ePub and Kindle

Modern distributed server applications are hosted on enterprise or cloud data centers that provide computing, storage, and networking capabilities to these applications. These applications are built using the implicit assumption that the underlying servers will be stable and normally available, barring for occasional faults. In many emerging scenarios, however, data centers and clouds only provide transient, rather than continuous, availability of their servers. Transiency in modern distributed systems arises in many contexts, such as green data centers powered using renewable intermittent sources, and cloud platforms that provide lower-cost transient servers which can be unilaterally revoked by the cloud operator. Transient computing resources are increasingly important, and existing fault-tolerance and resource management techniques are inadequate for transient servers because applications typically assume continuous resource availability. This thesis presents research in distributed systems design that treats transiency as a first-class design principle. I show that combining transiency-specific fault-tolerance mechanisms with resource management policies to suit application characteristics and requirements, can yield significant cost and performance benefits. These mechanisms and policies have been implemented and prototyped as part of software systems, which allow a wide range of applications, such as interactive services and distributed data processing, to be deployed on transient servers, and can reduce cloud computing costs by up to 90\%. This thesis makes contributions to four areas of computer systems research: transiency-specific fault-tolerance, resource allocation, abstractions, and resource reclamation. For reducing the impact of transient server revocations, I develop two fault-tolerance techniques that are tailored to transient server characteristics and application requirements. For interactive applications, I build a derivative cloud platform that masks revocations by transparently moving application-state between servers of different types. Similarly, for distributed data processing applications, I investigate the use of application level periodic checkpointing to reduce the performance impact of server revocations. For managing and reducing the risk of server revocations, I investigate the use of server portfolios that allow transient resource allocation to be tailored to application requirements. Finally, I investigate how resource providers (such as cloud platforms) can provide transient resource availability without revocation, by looking into alternative resource reclamation techniques. I develop resource deflation, wherein a server's resources are fractionally reclaimed, allowing the application to continue execution albeit with fewer resources. Resource deflation generalizes revocation, and the deflation mechanisms and cluster-wide policies can yield both high cluster utilization and low application performance degradation.


Resource Management in Utility and Cloud Computing

Resource Management in Utility and Cloud Computing
Author: Han Zhao
Publisher: Springer Science & Business Media
Total Pages: 94
Release: 2013-10-17
Genre: Computers
ISBN: 1461489709

Download Resource Management in Utility and Cloud Computing Book in PDF, ePub and Kindle

This SpringerBrief reviews the existing market-oriented strategies for economically managing resource allocation in distributed systems. It describes three new schemes that address cost-efficiency, user incentives, and allocation fairness with regard to different scheduling contexts. The first scheme, taking the Amazon EC2TM market as a case of study, investigates the optimal resource rental planning models based on linear integer programming and stochastic optimization techniques. This model is useful to explore the interaction between the cloud infrastructure provider and the cloud resource customers. The second scheme targets a free-trade resource market, studying the interactions amongst multiple rational resource traders. Leveraging an optimization framework from AI, this scheme examines the spontaneous exchange of resources among multiple resource owners. Finally, the third scheme describes an experimental market-oriented resource sharing platform inspired by eBay's transaction model. The study presented in this book sheds light on economic models and their implication to the utility-oriented scheduling problems.


Application-aware Resource Management for Cloud Platforms

Application-aware Resource Management for Cloud Platforms
Author: Xin He
Publisher:
Total Pages:
Release: 2016
Genre:
ISBN:

Download Application-aware Resource Management for Cloud Platforms Book in PDF, ePub and Kindle

Cloud computing has become increasingly popular in recent years. The benefits of cloud platforms include ease of application deployment, a pay-as-you-go model, and the ability to scale resources up or down based on an application's workload. Today's cloud platforms are being used to host increasingly complex distributed and parallel applications. The main premise of this thesis is that application-aware resource management techniques are better suited for distributed cloud applications over a systems-level one-size-fits-all approach. In this thesis, I study the cloud-based resource management techniques with a particular emphasis on how application-aware approaches can be used to improve system resource utilization and enhance applications' performance and cost. I first study always-on interactive applications that run on transient cloud servers such as Amazon spot instances. I show that by combining techniques like nested virtualization, live migration and lazy restoration together with intelligent bidding strategies, it is feasible to provide high availability to such applications while significantly reducing cost. I next study how to improve performance of parallel data processing applications like Hadoop and Spark that run in the cloud. I argue that network I/O contention in Hadoop can impact application throughput and implement a collaborative application-aware network and task scheduler using software-defined networking. By combining flow scheduling with task scheduling, our system can effectively avoid network contention and improve Hadoop's performance. I then investigate similar issues in Spark and find that task scheduling is more important for Spark jobs. I propose a network-aware task scheduling method that can adaptively schedule tasks for different types of jobs without system tuning and improve Spark's performance significantly. Finally, I study how to deploy network functions in the cloud. Specifically, I focus on comparing different methods of chaining network functions. By carrying out empirical evaluation of two different deployment methods, we figure out the advantages and disadvantages of each method. Our results suggest that the tenant-centric placement provides lower latencies while service-centric approach is more flexible for reconfiguration and capacity scaling.


Adaptive Resource Management and Scheduling for Cloud Computing

Adaptive Resource Management and Scheduling for Cloud Computing
Author: Florin Pop
Publisher: Springer
Total Pages: 197
Release: 2016-01-07
Genre: Computers
ISBN: 3319284487

Download Adaptive Resource Management and Scheduling for Cloud Computing Book in PDF, ePub and Kindle

This book constitutes the thoroughly refereed post-conference proceedings of the Second International Workshop on Adaptive Resource Management and Scheduling for Cloud Computing, ARMS-CC 2015, held in Conjunction with ACM Symposium on Principles of Distributed Computing, PODC 2015, in Donostia-San Sebastián, Spain, in July 2015. The 12 revised full papers, including 1 invited paper, were carefully reviewed and selected from 24 submissions. The papers have identified several important aspects of the problem addressed by ARMS-CC: self-* and autonomous cloud systems, cloud quality management and service level agreement (SLA), scalable computing, mobile cloud computing, cloud computing techniques for big data, high performance cloud computing, resource management in big data platforms, scheduling algorithms for big data processing, cloud composition, federation, bridging, and bursting, cloud resource virtualization and composition, load-balancing and co-allocation, fault tolerance, reliability, and availability of cloud systems.


Adaptive Resource Management and Scheduling for Cloud Computing

Adaptive Resource Management and Scheduling for Cloud Computing
Author: Florin Pop
Publisher: Springer
Total Pages: 223
Release: 2014-11-25
Genre: Computers
ISBN: 3319134647

Download Adaptive Resource Management and Scheduling for Cloud Computing Book in PDF, ePub and Kindle

This book constitutes the thoroughly refereed post-conference proceedings of the First International Workshop on Adaptive Resource Management and Scheduling for Cloud Computing, ARMS-CC 2014, held in Conjunction with ACM Symposium on Principles of Distributed Computing, PODC 2014, in Paris, France, in July 2014. The 14 revised full papers (including 2 invited talks) were carefully reviewed and selected from 29 submissions and cover topics such as scheduling methods and algorithms, services and applications, fundamental models for resource management in the cloud.


Elastic Resource Management in Cloud Computing Platforms

Elastic Resource Management in Cloud Computing Platforms
Author: Upendra Sharma
Publisher:
Total Pages: 156
Release: 2013
Genre: Cloud computing
ISBN:

Download Elastic Resource Management in Cloud Computing Platforms Book in PDF, ePub and Kindle

Large scale enterprise applications are known to observe dynamic workload; provisioning correct capacity for these applications remains an important and challenging problem. Predicting high variability fluctuations in workload or the peak workload is difficult; erroneous predictions often lead to under-utilized systems or in some situations cause temporarily outage of an otherwise well provisioned web-site. Consequently, rather than provisioning server capacity to handle infrequent peak workloads, an alternate approach of dynamically provisioning capacity on-the-fly in response to workload fluctuations has become popular. Cloud platforms are particularly suited for such applications due to their ability to provision capacity when needed and charge for usage on pay-per-use basis. Cloud environments enable elastic provisioning by providing a variety of hardware configurations as well as mechanisms to add or remove server capacity. The first part of this thesis presents Kingfisher, a cost-aware system that provides a generalized provisioning framework for supporting elasticity in the cloud by (i) leveraging multiple mechanisms to reduce the time to transition to new configurations, and (ii) optimizing the selection of a virtual server configuration that minimize cost. Majority of these enterprise applications, deployed as web applications, are distributed or replicated with a multi-tier architecture. SLAs for such applications are often expressed as a high percentile of a performance metric, for e.g. 99 percentile of end to end response time is less than 1 sec. In the second part of this thesis I present a model driven technique which provisions a multi-tier application for such an SLA and is targeted for cloud platforms. Enterprises critically depend on these applications and often own large IT infrastructure to support the regular operation of these applications. However, provisioning for a peak load or for high percentile of response time could be prohibitively expensive. Thus there is a need of hybrid cloud model, where the enterprise uses its own private resources for the majority of its computing, but then "bursts" into the cloud when local resources are insufficient. I discuss a new system, namely Seagull, which performs dynamic provisioning over a hybrid cloud model by enabling cloud bursting. Finally, I describe a methodology to model the configuration patterns (i.e deployment topologies) of different control plane services of a cloud management system itself. I present a generic methodology, based on empirical profiling, which provides initial deployment configuration of a control plane service and also a mechanism which iteratively adjusts the configuration to avoid violation of control plane's Service Level Objective (SLO).


Resource Management in Distributed Systems

Resource Management in Distributed Systems
Author: Anwesha Mukherjee
Publisher: Springer
Total Pages: 0
Release: 2024-07-10
Genre: Computers
ISBN: 9789819726431

Download Resource Management in Distributed Systems Book in PDF, ePub and Kindle

This book focuses on resource management in distributed computing systems. The book presents a collection of original, unpublished, and high-quality research works, which report the latest research advances on resource discovery, allocation, scheduling, etc., in cloud, fog, and edge computing. The topics covered in the book are resource management in cloud computing/edge computing/fog computing/dew computing, resource management in Internet of things, resource allocation, scheduling, monitoring, and orchestration in distributed computing systems, resource management in 5G network and beyond, latency-aware resource management, energy-efficient resource management, interoperability and portability, security and privacy in resource management, reliable resource management, trustworthiness in resource management, fault tolerance in resource management, and simulation related to resource management.


Elastic Resource Management in Distributed Clouds

Elastic Resource Management in Distributed Clouds
Author: Tian Guo
Publisher:
Total Pages:
Release: 2016
Genre:
ISBN:

Download Elastic Resource Management in Distributed Clouds Book in PDF, ePub and Kindle

The ubiquitous nature of computing devices and their increasing reliance on remote resources have driven and shaped public cloud platforms into unprecedented large-scale, distributed data centers. Concurrently, a plethora of cloud-based applications are experiencing multi-dimensional workload dynamics---workload volumes that vary along both time and space axes and with higher frequency. The interplay of diverse workload characteristics and distributed clouds raises several key challenges for efficiently and dynamically managing server resources. First, current cloud platforms impose certain restrictions that might hinder some resource management tasks. Second, an application-agnostic approach might not entail appropriate performance goals, therefore, requires numerous specific methods. Third, provisioning resources outside LAN boundary might incur huge delay which would impact the desired agility. In this dissertation, I investigate the above challenges and present the design of automated systems that manage resources for various applications in distributed clouds. The intermediate goal of these automated systems is to fully exploit potential benefits such as reduced network latency offered by increasingly distributed server resources. The ultimate goal is to improve end-to-end user response time with novel resource management approaches, within a certain cost budget. Centered around these two goals, I first investigate how to optimize the location and performance of virtual machines in distributed clouds. I use virtual desktops, mostly serving a single user, as an example use case for developing a black-box approach that ranks virtual machines based on their dynamic latency requirements. Those with high latency sensitivities have a higher priority of being placed or migrated to a cloud location closest to their users. Next, I relax the assumption of well-provisioned virtual machines and look at how to provision enough resources for applications that exhibit both temporal and spatial workload fluctuations. I propose an application-agnostic queueing model that captures the resource utilization and server response time. Building upon this model, I present a geo-elastic provisioning approach---referred as geo-elasticity---for replicable multi-tier applications that can spin up an appropriate amount of server resources in any cloud locations. Last, I explore the benefits of providing geo-elasticity for database clouds, a popular platform for hosting application backends. Performing geo-elastic provisioning for backend database servers entails several challenges that are specific to database workload, and therefore requires tailored solutions. In addition, cloud platforms offer resources at various prices for different locations. Towards this end, I propose a cost-aware geo-elasticity that combines a regression-based workload model and a queueing network capacity model for database clouds. In summary, hosting a diverse set of applications in an increasingly distributed cloud makes it interesting and necessary to develop new, efficient and dynamic resource management approaches.


Service Level Agreement Driven Adaptive Resource Management For Web Applications on Heterogeneous Compute Clouds

Service Level Agreement Driven Adaptive Resource Management For Web Applications on Heterogeneous Compute Clouds
Author: Waheed Iqbal
Publisher:
Total Pages:
Release: 2009
Genre:
ISBN:

Download Service Level Agreement Driven Adaptive Resource Management For Web Applications on Heterogeneous Compute Clouds Book in PDF, ePub and Kindle

Cloud computing is an emerging topic in the field of parallel and distributed computing. Many IT giants such as IBM, Sun, Amazon, Google, and Microsoft are promoting and offering various storage and compute clouds. Clouds provide services such as high performance computing, storage, and application hosting. Cloud providers are expected to ensure Quality of Service (QoS) through a Service Level Agreement (SLA) between the provider and the consumer. In this research, I develop a heterogeneous testbed compute cloud and investigate adaptive management of resources for Web applications to satisfy a SLA that enforces specific response time requirements. I develop a system on top of EUCALYTPUS framework that actively monitors the response time of the compute resources assign to a Web application and dynamically allocates the resources required by the application to satisfy the specific response time requirements.