EconPapers    
Economics at your fingertips  
 

Holistic Utility Satisfaction in Cloud Data Centre Network Using Reinforcement Learning

Pejman Goudarzi (), Mehdi Hosseinpour, Roham Goudarzi and Jaime Lloret ()
Additional contact information
Pejman Goudarzi: ICT Research Institute (ITRC), Tehran 14155-3961, Iran
Mehdi Hosseinpour: ICT Research Institute (ITRC), Tehran 14155-3961, Iran
Roham Goudarzi: Faculty of Science, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
Jaime Lloret: Department of Communications, Polytechnic University of Valencia, Camino de Vera, 46022 Valencia, Spain

Future Internet, 2022, vol. 14, issue 12, 1-21

Abstract: Cloud computing leads to efficient resource allocation for network users. In order to achieve efficient allocation, many research activities have been conducted so far. Some researchers focus on classical optimisation theory techniques (such as multi-objective optimisation, evolutionary optimisation, game theory, etc.) to satisfy network providers and network users’ service-level agreement (SLA) requirements. Normally, in a cloud data centre network (CDCN), it is difficult to jointly satisfy both the cloud provider and cloud customer’ utilities, and this leads to complex combinatorial problems, which are usually NP-hard. Recently, machine learning and artificial intelligence techniques have received much attention from the networking community because of their capability to solve complicated networking problems. In the current work, at first, the holistic utility satisfaction for the cloud data centre provider and customers is formulated as a reinforcement learning (RL) problem with a specific reward function, which is a convex summation of users’ utility functions and cloud provider’s utility. The user utility functions are modelled as a function of cloud virtualised resources (such as storage, CPU, RAM), connection bandwidth, and also, the network-based expected packet loss and round-trip time factors associated with the cloud users. The cloud provider utility function is modelled as a function of resource prices and energy dissipation costs. Afterwards, a Q-learning implementation of the mentioned RL algorithm is introduced, which is able to converge to the optimal solution in an online and fast manner. The simulation results exhibit the enhanced convergence speed and computational complexity properties of the proposed method in comparison with similar approaches from the joint cloud customer/provider utility satisfaction perspective. To evaluate the scalability property of the proposed method, the results are also repeated for different cloud user population scenarios (small, medium, and large).

Keywords: CDCN; QoS; VM; reinforcement learning; resource assignment (search for similar items in EconPapers)
JEL-codes: O3 (search for similar items in EconPapers)
Date: 2022
References: View references in EconPapers View complete reference list from CitEc
Citations:

Downloads: (external link)
https://www.mdpi.com/1999-5903/14/12/368/pdf (application/pdf)
https://www.mdpi.com/1999-5903/14/12/368/ (text/html)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:gam:jftint:v:14:y:2022:i:12:p:368-:d:996917

Access Statistics for this article

Future Internet is currently edited by Ms. Grace You

More articles in Future Internet from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().

 
Page updated 2025-03-19
Handle: RePEc:gam:jftint:v:14:y:2022:i:12:p:368-:d:996917