SlideShare a Scribd company logo
Module 4
Cloud Resource Management
and Scheduling
Policies and mechanisms for resource management
Cloud resource management policies:
Policies
Admission control. The explicit goal of an admission control policy is to prevent
the system from accepting workloads in violation of high-level
system policies
Capacity allocation. To allocate resources for individual instances; an instance is an
activation of a service.
Load balancing. Distribute the workload evenly among the servers.
Energy optimization. Minimization of energy consumption.
Quality-of-service (QoS)
guarantees.
Ability to satisfy timing or other conditions specified by a
Service Level Agreement.
CLOUD RESOURCE MANAGEMENT AND SCHEDULING
Mechanisms for the implementation of resource management policies
• Control theory : uses the feedback to guarantee system stability and
predict transient behavior.
• Machine learning : does not need a performance model of the
system.
• Utility-based : require a performance model and a mechanism to
correlate user-level performance with cost.
• Market-oriented/economic : do not require a model of the system,
e.g., combinatorial auctions for bundles of resources.
Applications of control theory to task
scheduling on a cloud
The controller uses the feedback regarding the current state as well
as the estimation of the future disturbance due to environment to
compute the optimal inputs over a finite horizon. The two
parameters r and s are the weighting factors of the performance
index
Predictive Filter:
• The system starts by analyzing the "external traffic," which might represent
the amount of incoming tasks or requests.
• The predictive filter uses this traffic information to create a "forecast." This
forecast predicts future traffic, helping the system prepare for what's
coming.
Optimal Controller:
• The forecast from the predictive filter, along with some other inputs labeled
r (reference or target) and s (settings or constraints), goes into the optimal
controller.
• The optimal controller's job is to make the best decisions on how to
manage the incoming traffic based on the forecast, target, and constraints.
It outputs u (k)u^*(k)u (k), which could represent the optimal actions or
∗ ∗
settings to keep the system running smoothly.
• The controller also receives "state feedback," q(k) which tells it the current
state of the queue. This feedback helps it adjust its decisions in real-time.
Queuing Dynamics:
• The optimal actions, u (k), from the controller go to the queuing
∗
dynamics block. This part models the actual queue, handling how
tasks are processed and managed.
• The queuing dynamics block has two outputs:
λ(k): the rate at which tasks are being processed or entering the
system.
ω(k): some other state or measurement related to the queue (e.g.,
waiting time or queue length).
Stability of a two-level resource allocation
architecture
Key Components of the Control System:
Control System Components:
• The system has two levels of control:
• Application Level (Local Controllers): Each application has its own controller
(Application Controller) that manages resources for that specific application.
These controllers work based on the application’s Service Level Agreement
(SLA) to ensure the application meets its performance requirements.
• Service Provider Level (Cloud Controller): This is the higher-level controller
that oversees all applications on the cloud platform. It manages resources
across applications, coordinating and balancing resources to ensure efficient
use of the cloud platform.
Each application-level controller is equipped with:
• Monitor: Tracks performance metrics like CPU usage, memory
utilization, etc., for its respective application.
• Decision Maker: Analyzes the data from the monitor and decides
whether to allocate or deallocate resources.
• Actuator: Executes the resource adjustments, like adding or removing
virtual machines (VMs) for the application.
Feedback Loop and Stability
• This control system uses feedback from the Monitors to keep the
system stable. For example:
• If an application’s workload increases, the monitor detects this, and the
application controller might decide to allocate more VMs to handle the load.
• The cloud controller receives feedback from all application controllers and
ensures the cloud platform remains balanced overall.
• Stability is crucial. If adjustments (like adding or removing VMs) are
made too quickly or too frequently, the system can become unstable,
causing problems like: Thrashing, Instability Sources.
Lessons Learned from the two-level experiment
• Controllers should wait for the system to stabilize before making
adjustments, preventing rapid changes that could lead to instability.
• If upper and lower thresholds are set, they should be spaced
sufficiently apart to prevent frequent oscillations. Adjustments like
adding or removing VMs must be done carefully to avoid crossing
thresholds repeatedly, which could destabilize the system.
Feedback control based on dynamic
thresholds
• The elements involved in a control system are sensors, monitors, and
actuators.
• The sensors measure the parameter(s) of interest, then transmit the
measured values to a monitor, which determines whether the system
behavior must be changed, and, if so, it requests that the actuators
carry out the necessary actions.
• Thresholds.
A threshold is the value of a parameter related to the state of a system
that triggers a change in the system behavior.
The two thresholds determine different actions; for example, a high
threshold could force the system to limit its activities and a low
threshold could encourage additional activities.
• Control granularity refers to the level of detail of the information used
to control the system.
• Fine control means that very detailed information about the
parameters controlling the system state is used, whereas coarse
control means that the accuracy of these parameters is traded for the
efficiency of implementation.
• Proportional Thresholding
In Proportional Thresholding The questions addressed are:
1. Is it beneficial to have two types of controllers, (1) application
controllers that determine whether additional resources are needed
and (2) cloud controllers that arbitrate requests for resources and
allocate the physical resources?
2. Is it feasible to consider fine control? Is course control more
adequate in a cloud computing environment?
3. Are dynamic thresholds based on time averages better than static
ones?
4. Is it better to have a high and a low threshold, or it is sufficient to
define only a high threshold?
• The essence of the proportional thresholding is captured by the
following algorithm:
1. Compute the integral value of the high and the low thresholds as
averages of the maximum and, respectively, the minimum of the
processor utilization over the process history.
2. Request additional VMs when the average value of the CPU
utilization over the current time slice exceeds the high threshold.
3. Release a VM when the average value of the CPU utilization over the
current time slice falls below the low threshold.
Conclusions
• Dynamic thresholds perform better than the static ones.
• Two thresholds are better than one.
Coordination of specialized autonomic
performance managers
• Autonomous performance and
power managers cooperate to
ensure SLA prescribed
performance and energy
optimization.
• They are fed with performance
and power data and implement
the performance and power
management policies,
respectively.
• Use separate controllers/managers for the two objectives.
• Identify a minimal set of parameters to be exchanged between the
two managers.
• Use a joint utility function for power and performance.
• Set up a power cap for individual systems based on the utility-
optimized power management policy.
• Use a standard performance manager modified only to accept input
from the power manager regarding the frequency determined
according to the power management policy.
• Use standard software systems.
A utility-based model for cloud-based Web
services
The utility function U(R) is a series of
step functions with jumps
corresponding to the response time,
R = R0|R1|R2, when the reward and
the penalty levels change according to
the SLA. The dotted line shows a
quadratic approximation of the utility
function.
(a) The utility function: vk the revenue
(or the penalty) function of the
response time rk for a request of class
k.
(b) A network of multiqueues.
• A service level agreement (SLA) : specifies the rewards as well as
penalties associated with specific performance metrics.
• The SLA for cloud-based web services uses the average response time
to reflect the Quality of Service.
• We assume a cloud providing K different classes of service, each class
k involving Nk applications.
• The system is modeled as a network of queues with multi-queues for
each server.
• A delay center models the think time of the user after the completion
of service at one server and the start of processing at the next server.
Resource bundling: Combinatorial auctions
for cloud resources
• Resources in a cloud are allocated in bundles, allowing users get
maximum benefit from a specific combination of resources. Indeed,
along with CPU cycles, an application needs specific amounts of main
memory, disk space, network bandwidth, and so on.
Combinatorial Auctions.
In combinatorial auctions, users can bid on combinations (or bundles)
of resources rather than individual items.
Types of Combinatorial Auctions
• Some recent combinatorial auction algorithms include:
• Simultaneous Clock Auction: All resource prices are visible and
updated in real time like a "clock."
• Clock Proxy Auction: Users bid on bundles via proxies based on
current clock prices.
• Ascending Clock Auction (ASCA): Prices start low and gradually
increase as users place bids.
• In these auctions:
• The current price of each resource is displayed on a "clock" that
everyone can see.
• Users compete by placing bids as prices increase.
The schematics of the ASCA algorithm.
To allow for a single round, auction
users are represented by proxies that
place the bids xu(t).
The auctioneer determines whether
there is an excess demand and, in that
case, raises the price of resources for
which the demand exceeds the supply
and requests new bids.
• Resource Bundling solves the problem of allocating multiple resources
at once.
• Combinatorial Auctions are a flexible and scalable way to allocate
bundles of resources.
• The process ensures fairness, scalability, and efficient pricing, with
clear winners and losers.
Scheduling algorithms for computing clouds
Scheduling in Cloud Systems
In Resource Sharing Layers: A server supports multiple virtual machines
(VMs). Each VM supports multiple applications. Each application comprises
threads, with resources dynamically allocated across these layers.
Quality-of-Service (QoS) Requirements
The QoS needs vary by application type, leading to different scheduling
approaches:
Best-effort applications, Soft real-time applications, Hard real-time
applications:
Objectives of Schedulers
Schedulers have different objectives depending on the type of system:
Best-effort applications:
Examples: Batch jobs, data
analytics.
Maximize throughput: Complete as many
jobs as possible within a given time.
Minimize turnaround time: Reduce the
time between job submission and
completion.
Scheduling policies: Algorithms like Round-
robin, FCFS (First Come First Serve), and
SJF (Shortest Job First) are used here.
Soft real-time applications:
Examples: Multimedia
streaming (audio/video)
QoS requirements are less rigid than hard
real-time tasks but must still maintain
performance.
Real-time algorithms like EDF (Earliest
Deadline First) or RMA (Rate Monotonic
Algorithm) are often used to prioritize
tasks.
Hard real-time
applications:
Examples: Critical systems
(medical, industrial
automation).
Strict deadlines and resource guarantees.
Tasks must be executed precisely within
their timing constraints, or failure occurs.
Advanced algorithms like RAD (Resource
Allocation/Dispatching) and RBED (Rate-
Based Earliest Deadline) help integrate
hard real-time scheduling in mixed
workloads.
Fairness in Scheduling
A scheduler must ensure fair resource allocation, especially when
multiple users or threads share the same resource. Two fairness criteria
are discussed:
Max-min Fairness:
• Ensures no user gets more resources than they request.
• Maximizes the smallest allocation (Bmin) while satisfying constraints
recursively.
Weighted Fairness:
• Allocates resources proportionally based on task weights (e.g.,
priority or importance).
• Ensures tasks with higher weights receive more resources.

More Related Content

PPTX
NIST Model of Cloud Computing by Piyush Bujade.pptx
PPT
Communications is distributed systems
PPTX
Task scheduling Survey in Cloud Computing
DOC
Naming in Distributed System
PPTX
Introduction to Aneka, Aneka Model is explained
PPTX
CLOUD ENABLING TECHNOLOGIES.pptx
PPTX
Foundations of cloud security monitoring
PDF
Requirements Engineering - Goals
NIST Model of Cloud Computing by Piyush Bujade.pptx
Communications is distributed systems
Task scheduling Survey in Cloud Computing
Naming in Distributed System
Introduction to Aneka, Aneka Model is explained
CLOUD ENABLING TECHNOLOGIES.pptx
Foundations of cloud security monitoring
Requirements Engineering - Goals

What's hot (20)

PPT
Cloud service management
PDF
BUSINESS CONSIDERATIONS FOR CLOUD COMPUTING
PDF
Resource Cluster and Multi-Device Broker.pdf
PDF
Virtualization Basics
PPTX
Introduction to simulation and modeling
PDF
Cloud service lifecycle management
PPTX
Cloud computing stack
PPT
Simulation in terminated system
PPTX
Load balancing in cloud computing.pptx
PPTX
The principles of simulation system design.pptx
PPTX
Challenges in Cloud Computing – VM Migration
PDF
Xml databases
PPT
Unit 2 -Cloud Computing Architecture
PPTX
VTU 6th Sem Elective CSE - Module 3 cloud computing
PPTX
Vm consolidation for energy efficient cloud computing
PPTX
Cloud Management Mechanisms
PDF
Web Services / Technology in Cloud Computing
PPTX
Cloud security Presentation
PPTX
System Modeling & Simulation Introduction
PPTX
Genetic Algorithm for task scheduling in Cloud Computing Environment
Cloud service management
BUSINESS CONSIDERATIONS FOR CLOUD COMPUTING
Resource Cluster and Multi-Device Broker.pdf
Virtualization Basics
Introduction to simulation and modeling
Cloud service lifecycle management
Cloud computing stack
Simulation in terminated system
Load balancing in cloud computing.pptx
The principles of simulation system design.pptx
Challenges in Cloud Computing – VM Migration
Xml databases
Unit 2 -Cloud Computing Architecture
VTU 6th Sem Elective CSE - Module 3 cloud computing
Vm consolidation for energy efficient cloud computing
Cloud Management Mechanisms
Web Services / Technology in Cloud Computing
Cloud security Presentation
System Modeling & Simulation Introduction
Genetic Algorithm for task scheduling in Cloud Computing Environment
Ad

Similar to CLOUD RESOURCE MANAGEMENT AND SCHEDULING (20)

PPTX
Cloud computing unit4 ppt for btech students
PDF
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...
PDF
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...
PDF
Resource usage optimization in cloud based networks
PDF
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
PDF
Impact of Resource Management and Scalability on performance of Cloud Applica...
PDF
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
PDF
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
PDF
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
PDF
T04503113118
PDF
International Journal of Engineering Research and Development
PDF
Ijariie1161
PDF
A Survey on Heuristic Based Techniques in Cloud Computing
PPT
Scheduling in CCE
PDF
A survey on various resource allocation policies in cloud computing environment
PDF
A survey on various resource allocation policies in cloud computing environment
PPTX
Self-adaptation Challenges for Cloud-based Applications (Feedback Computing 2...
PDF
A Survey on Task Scheduling and Load Balanced Algorithms in Cloud Computing
PDF
A Survey on Service Request Scheduling in Cloud Based Architecture
PDF
A Survey on Service Request Scheduling in Cloud Based Architecture
Cloud computing unit4 ppt for btech students
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...
Resource usage optimization in cloud based networks
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
Impact of Resource Management and Scalability on performance of Cloud Applica...
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
IMPACT OF RESOURCE MANAGEMENT AND SCALABILITY ON PERFORMANCE OF CLOUD APPLICA...
T04503113118
International Journal of Engineering Research and Development
Ijariie1161
A Survey on Heuristic Based Techniques in Cloud Computing
Scheduling in CCE
A survey on various resource allocation policies in cloud computing environment
A survey on various resource allocation policies in cloud computing environment
Self-adaptation Challenges for Cloud-based Applications (Feedback Computing 2...
A Survey on Task Scheduling and Load Balanced Algorithms in Cloud Computing
A Survey on Service Request Scheduling in Cloud Based Architecture
A Survey on Service Request Scheduling in Cloud Based Architecture
Ad

More from 2021ismadhuprasadrna (9)

PPTX
1.introduction analysis and design of algorithms(1).pptx
PPTX
Interpretation and report writing module 3
PPTX
Database relational model_unit3_2023 (1).pptx
PPTX
chapter 7- Relations The Second Time Around.pptx
PPTX
Data communications and networking by behrouz
PPTX
Chapter 10- Recurrence Relations [Autosaved] by BMN.pptx
PPT
natural language processing by Christopher
PPTX
Customer-Success-Management- by MPRN.pptx
PPTX
CC CLOUD RESOURCE VIRTUALIZATION PPT TO REFER
1.introduction analysis and design of algorithms(1).pptx
Interpretation and report writing module 3
Database relational model_unit3_2023 (1).pptx
chapter 7- Relations The Second Time Around.pptx
Data communications and networking by behrouz
Chapter 10- Recurrence Relations [Autosaved] by BMN.pptx
natural language processing by Christopher
Customer-Success-Management- by MPRN.pptx
CC CLOUD RESOURCE VIRTUALIZATION PPT TO REFER

Recently uploaded (20)

DOCX
573137875-Attendance-Management-System-original
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPT
Introduction, IoT Design Methodology, Case Study on IoT System for Weather Mo...
PPTX
web development for engineering and engineering
PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPTX
UNIT 4 Total Quality Management .pptx
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
PDF
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PDF
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PDF
Unit I ESSENTIAL OF DIGITAL MARKETING.pdf
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PPTX
OOP with Java - Java Introduction (Basics)
PPTX
Current and future trends in Computer Vision.pptx
PPTX
CH1 Production IntroductoryConcepts.pptx
573137875-Attendance-Management-System-original
bas. eng. economics group 4 presentation 1.pptx
Introduction, IoT Design Methodology, Case Study on IoT System for Weather Mo...
web development for engineering and engineering
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
R24 SURVEYING LAB MANUAL for civil enggi
UNIT 4 Total Quality Management .pptx
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
Model Code of Practice - Construction Work - 21102022 .pdf
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
Unit I ESSENTIAL OF DIGITAL MARKETING.pdf
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
OOP with Java - Java Introduction (Basics)
Current and future trends in Computer Vision.pptx
CH1 Production IntroductoryConcepts.pptx

CLOUD RESOURCE MANAGEMENT AND SCHEDULING

  • 1. Module 4 Cloud Resource Management and Scheduling
  • 2. Policies and mechanisms for resource management Cloud resource management policies: Policies Admission control. The explicit goal of an admission control policy is to prevent the system from accepting workloads in violation of high-level system policies Capacity allocation. To allocate resources for individual instances; an instance is an activation of a service. Load balancing. Distribute the workload evenly among the servers. Energy optimization. Minimization of energy consumption. Quality-of-service (QoS) guarantees. Ability to satisfy timing or other conditions specified by a Service Level Agreement.
  • 4. Mechanisms for the implementation of resource management policies • Control theory : uses the feedback to guarantee system stability and predict transient behavior. • Machine learning : does not need a performance model of the system. • Utility-based : require a performance model and a mechanism to correlate user-level performance with cost. • Market-oriented/economic : do not require a model of the system, e.g., combinatorial auctions for bundles of resources.
  • 5. Applications of control theory to task scheduling on a cloud The controller uses the feedback regarding the current state as well as the estimation of the future disturbance due to environment to compute the optimal inputs over a finite horizon. The two parameters r and s are the weighting factors of the performance index
  • 6. Predictive Filter: • The system starts by analyzing the "external traffic," which might represent the amount of incoming tasks or requests. • The predictive filter uses this traffic information to create a "forecast." This forecast predicts future traffic, helping the system prepare for what's coming. Optimal Controller: • The forecast from the predictive filter, along with some other inputs labeled r (reference or target) and s (settings or constraints), goes into the optimal controller. • The optimal controller's job is to make the best decisions on how to manage the incoming traffic based on the forecast, target, and constraints. It outputs u (k)u^*(k)u (k), which could represent the optimal actions or ∗ ∗ settings to keep the system running smoothly. • The controller also receives "state feedback," q(k) which tells it the current state of the queue. This feedback helps it adjust its decisions in real-time.
  • 7. Queuing Dynamics: • The optimal actions, u (k), from the controller go to the queuing ∗ dynamics block. This part models the actual queue, handling how tasks are processed and managed. • The queuing dynamics block has two outputs: λ(k): the rate at which tasks are being processed or entering the system. ω(k): some other state or measurement related to the queue (e.g., waiting time or queue length).
  • 8. Stability of a two-level resource allocation architecture
  • 9. Key Components of the Control System: Control System Components: • The system has two levels of control: • Application Level (Local Controllers): Each application has its own controller (Application Controller) that manages resources for that specific application. These controllers work based on the application’s Service Level Agreement (SLA) to ensure the application meets its performance requirements. • Service Provider Level (Cloud Controller): This is the higher-level controller that oversees all applications on the cloud platform. It manages resources across applications, coordinating and balancing resources to ensure efficient use of the cloud platform.
  • 10. Each application-level controller is equipped with: • Monitor: Tracks performance metrics like CPU usage, memory utilization, etc., for its respective application. • Decision Maker: Analyzes the data from the monitor and decides whether to allocate or deallocate resources. • Actuator: Executes the resource adjustments, like adding or removing virtual machines (VMs) for the application.
  • 11. Feedback Loop and Stability • This control system uses feedback from the Monitors to keep the system stable. For example: • If an application’s workload increases, the monitor detects this, and the application controller might decide to allocate more VMs to handle the load. • The cloud controller receives feedback from all application controllers and ensures the cloud platform remains balanced overall. • Stability is crucial. If adjustments (like adding or removing VMs) are made too quickly or too frequently, the system can become unstable, causing problems like: Thrashing, Instability Sources.
  • 12. Lessons Learned from the two-level experiment • Controllers should wait for the system to stabilize before making adjustments, preventing rapid changes that could lead to instability. • If upper and lower thresholds are set, they should be spaced sufficiently apart to prevent frequent oscillations. Adjustments like adding or removing VMs must be done carefully to avoid crossing thresholds repeatedly, which could destabilize the system.
  • 13. Feedback control based on dynamic thresholds • The elements involved in a control system are sensors, monitors, and actuators. • The sensors measure the parameter(s) of interest, then transmit the measured values to a monitor, which determines whether the system behavior must be changed, and, if so, it requests that the actuators carry out the necessary actions.
  • 14. • Thresholds. A threshold is the value of a parameter related to the state of a system that triggers a change in the system behavior. The two thresholds determine different actions; for example, a high threshold could force the system to limit its activities and a low threshold could encourage additional activities. • Control granularity refers to the level of detail of the information used to control the system. • Fine control means that very detailed information about the parameters controlling the system state is used, whereas coarse control means that the accuracy of these parameters is traded for the efficiency of implementation.
  • 15. • Proportional Thresholding In Proportional Thresholding The questions addressed are: 1. Is it beneficial to have two types of controllers, (1) application controllers that determine whether additional resources are needed and (2) cloud controllers that arbitrate requests for resources and allocate the physical resources? 2. Is it feasible to consider fine control? Is course control more adequate in a cloud computing environment? 3. Are dynamic thresholds based on time averages better than static ones? 4. Is it better to have a high and a low threshold, or it is sufficient to define only a high threshold?
  • 16. • The essence of the proportional thresholding is captured by the following algorithm: 1. Compute the integral value of the high and the low thresholds as averages of the maximum and, respectively, the minimum of the processor utilization over the process history. 2. Request additional VMs when the average value of the CPU utilization over the current time slice exceeds the high threshold. 3. Release a VM when the average value of the CPU utilization over the current time slice falls below the low threshold. Conclusions • Dynamic thresholds perform better than the static ones. • Two thresholds are better than one.
  • 17. Coordination of specialized autonomic performance managers • Autonomous performance and power managers cooperate to ensure SLA prescribed performance and energy optimization. • They are fed with performance and power data and implement the performance and power management policies, respectively.
  • 18. • Use separate controllers/managers for the two objectives. • Identify a minimal set of parameters to be exchanged between the two managers. • Use a joint utility function for power and performance. • Set up a power cap for individual systems based on the utility- optimized power management policy. • Use a standard performance manager modified only to accept input from the power manager regarding the frequency determined according to the power management policy. • Use standard software systems.
  • 19. A utility-based model for cloud-based Web services The utility function U(R) is a series of step functions with jumps corresponding to the response time, R = R0|R1|R2, when the reward and the penalty levels change according to the SLA. The dotted line shows a quadratic approximation of the utility function.
  • 20. (a) The utility function: vk the revenue (or the penalty) function of the response time rk for a request of class k. (b) A network of multiqueues.
  • 21. • A service level agreement (SLA) : specifies the rewards as well as penalties associated with specific performance metrics. • The SLA for cloud-based web services uses the average response time to reflect the Quality of Service. • We assume a cloud providing K different classes of service, each class k involving Nk applications. • The system is modeled as a network of queues with multi-queues for each server. • A delay center models the think time of the user after the completion of service at one server and the start of processing at the next server.
  • 22. Resource bundling: Combinatorial auctions for cloud resources • Resources in a cloud are allocated in bundles, allowing users get maximum benefit from a specific combination of resources. Indeed, along with CPU cycles, an application needs specific amounts of main memory, disk space, network bandwidth, and so on. Combinatorial Auctions. In combinatorial auctions, users can bid on combinations (or bundles) of resources rather than individual items.
  • 23. Types of Combinatorial Auctions • Some recent combinatorial auction algorithms include: • Simultaneous Clock Auction: All resource prices are visible and updated in real time like a "clock." • Clock Proxy Auction: Users bid on bundles via proxies based on current clock prices. • Ascending Clock Auction (ASCA): Prices start low and gradually increase as users place bids. • In these auctions: • The current price of each resource is displayed on a "clock" that everyone can see. • Users compete by placing bids as prices increase.
  • 24. The schematics of the ASCA algorithm. To allow for a single round, auction users are represented by proxies that place the bids xu(t). The auctioneer determines whether there is an excess demand and, in that case, raises the price of resources for which the demand exceeds the supply and requests new bids.
  • 25. • Resource Bundling solves the problem of allocating multiple resources at once. • Combinatorial Auctions are a flexible and scalable way to allocate bundles of resources. • The process ensures fairness, scalability, and efficient pricing, with clear winners and losers.
  • 26. Scheduling algorithms for computing clouds
  • 27. Scheduling in Cloud Systems In Resource Sharing Layers: A server supports multiple virtual machines (VMs). Each VM supports multiple applications. Each application comprises threads, with resources dynamically allocated across these layers. Quality-of-Service (QoS) Requirements The QoS needs vary by application type, leading to different scheduling approaches: Best-effort applications, Soft real-time applications, Hard real-time applications:
  • 28. Objectives of Schedulers Schedulers have different objectives depending on the type of system: Best-effort applications: Examples: Batch jobs, data analytics. Maximize throughput: Complete as many jobs as possible within a given time. Minimize turnaround time: Reduce the time between job submission and completion. Scheduling policies: Algorithms like Round- robin, FCFS (First Come First Serve), and SJF (Shortest Job First) are used here. Soft real-time applications: Examples: Multimedia streaming (audio/video) QoS requirements are less rigid than hard real-time tasks but must still maintain performance. Real-time algorithms like EDF (Earliest Deadline First) or RMA (Rate Monotonic Algorithm) are often used to prioritize tasks. Hard real-time applications: Examples: Critical systems (medical, industrial automation). Strict deadlines and resource guarantees. Tasks must be executed precisely within their timing constraints, or failure occurs. Advanced algorithms like RAD (Resource Allocation/Dispatching) and RBED (Rate- Based Earliest Deadline) help integrate hard real-time scheduling in mixed workloads.
  • 29. Fairness in Scheduling A scheduler must ensure fair resource allocation, especially when multiple users or threads share the same resource. Two fairness criteria are discussed: Max-min Fairness: • Ensures no user gets more resources than they request. • Maximizes the smallest allocation (Bmin) while satisfying constraints recursively. Weighted Fairness: • Allocates resources proportionally based on task weights (e.g., priority or importance). • Ensures tasks with higher weights receive more resources.