Abstract
Public Infrastructure as a Service (IaaS) clouds such as Amazon, GoGrid
and Rackspace deliver computational resources by means of virtualisation
technologies. These technologies allow multiple independent virtual
machines to reside in apparent isolation on the same physical host.
Dynamically scaling applications running on IaaS clouds can lead to
varied and unpredictable results because of the performance interference
effects associated with co-located virtual machines. Determining
appropriate scaling policies in a dynamic non-stationary environment is
non-trivial. One principle advantage exhibited by IaaS clouds over their
traditional hosting counterparts is the ability to scale resources
on-demand. However, a problem arises concerning resource allocation as
to which resources should be added and removed when the underlying
performance of the resource is in a constant state of flux. Decision
theoretic frameworks such as Markov Decision Processes are particularly
suited to decision making under uncertainty. By applying a temporal
difference, reinforcement learning algorithm known as Q-learning,
optimal scaling policies can be determined. Additionally, reinforcement
learning techniques typically suffer from curse of dimensionality
problems, where the state space grows exponentially with each additional
state variable. To address this challenge, we also present a novel
parallel Q-learning approach aimed at reducing the time taken to
determine optimal policies whilst learning online.
| Original language | English (Ireland) |
|---|---|
| Journal | Concurrency And Computation: Practice And Experience (To Appear) |
| DOIs | |
| Publication status | Published - 1 May 2012 |
Authors (Note for portal: view the doc link for the full list of authors)
- Authors
- Barrett, E; Howley, E; Duggan, J