You are required to read and agree to the below before accessing a full-text version of an article in the IDE article repository.

The full-text document you are about to access is subject to national and international copyright laws. In most cases (but not necessarily all) the consequence is that personal use is allowed given that the copyright owner is duly acknowledged and respected. All other use (typically) require an explicit permission (often in writing) by the copyright owner.

For the reports in this repository we specifically note that

  • the use of articles under IEEE copyright is governed by the IEEE copyright policy (available at http://www.ieee.org/web/publications/rights/copyrightpolicy.html)
  • the use of articles under ACM copyright is governed by the ACM copyright policy (available at http://www.acm.org/pubs/copyright_policy/)
  • technical reports and other articles issued by M‰lardalen University is free for personal use. For other use, the explicit consent of the authors is required
  • in other cases, please contact the copyright owner for detailed information

By accepting I agree to acknowledge and respect the rights of the copyright owner of the document I am about to access.

If you are in doubt, feel free to contact webmaster@ide.mdh.se

Adaptive Service Performance Control using Cooperative Fuzzy Reinforcement Learning in Virtualized Environments

Authors:

Olumuyiwa Ibidunmoye , Mahshid Helali Moghadam, Ewnetu Bayuh Lakew , Erik Elmroth

Publication Type:

Conference/Workshop Paper

Venue:

10th IEEE/ACM International Conference on Utility and Cloud Computing


Abstract

Designing efficient control mechanisms to meet strict performance requirements with respect to changing workload demands without sacrificing resource efficiency remains a challenge in cloud infrastructures. A popular approach is fine-grained resource provisioning via auto-scaling mechanisms that rely on either threshold-based adaptation rules or sophisticated queuing/control-theoretic models. While it is difficult at design time to specify optimal threshold rules, it is even more challenging inferring precise performance models for the multitude of services. Recently, reinforcement learning has been applied to address this challenge. However, such approaches require many learning trials to stabilize at the beginning and when operational conditions vary thereby limiting their application under dynamic workloads. To this end, we extend the standard reinforcement learning approach in two ways: a) we formulate the system state as a fuzzy space and b) exploit a set of cooperative agents to explore multiple fuzzy states in parallel to speed up learning. Through multiple experiments on a real virtualized testbed, we demonstrate that our approach converges quickly, meets performance targets at high efficiency without explicit service models.

Bibtex

@inproceedings{Ibidunmoye5055,
author = {Olumuyiwa Ibidunmoye and Mahshid Helali Moghadam and Ewnetu Bayuh Lakew and Erik Elmroth},
title = {Adaptive Service Performance Control using Cooperative Fuzzy Reinforcement Learning in Virtualized Environments},
month = {December},
year = {2017},
booktitle = {10th IEEE/ACM International Conference on Utility and Cloud Computing},
url = {http://www.es.mdh.se/publications/5055-}
}