Developing Deep RNN-Based Client Resource Utilization Prediction Model for an Improved Cloud Resource Provisioning

Loading...
Thumbnail Image

Journal Title

Journal ISSN

Volume Title

Publisher

ASTU

Abstract

One of the key problems facing cloud computing is to estimate accurately the use of resources for future requirements. Cloud computing resource provisioning requires an adaptive and reliable forecast of cloud workloads. The consumption of cloud resources is continuously evolving, which makes it difficult to provide reliable forecasts for predicting algorithms. However, the prevalent approaches cannot effectively forecast high dimensional and highly variable cloud workloads. This leads to a waste of resources and thus, failure to meet SLAs. This motivates the research, which aims to predict the cloud resource utilization for better provisioning of cloud resources using a Deep RNN-Based architecture to improve the accuracy in the prediction of the workload. Since the RNN is sequential data analysis, the problem of workload prediction has recently not been tackled. RNN, however, frequently performs poorly on LTM dependencies in learning and it also does not reliably forecast workloads. To overcome these important challenges, stacked LSTM RNNs have been proposed for predicting CPU utilization, memory, and disk I/O due to the ability to maintain information and accurately predict time series issues and this makes it a promising candidate to forecast cloud resource utilization more accurately compared to traditional approaches. First, the method uses multivariate analysis to derive the sequential and contextual characteristics of the historical workload data via the LSTM network from the original high-dimensional workload data to compare between metrics and increase the applicability of the model. Next, the model integrates the different callback API, hyperparameters (optimizers, metrics, and losses), and layers (Regularization, Normalization, recurrent, and activation layers) to achieve the adaptive and accurate prediction for highly variable workloads. The experiment carried out on real-world Alibaba, Google, and Intel Netbatch workload trace to demonstrate that the proposed method achieves superior prediction accuracy and state-of-the-art performance compared to other workload prediction methods for high-dimensional and highly variable real-world cloud workloads. The maximum mean squared prediction error of the proposed model is reduced up to 0.000403.

Description

Citation

Collections

Endorsement

Review

Supplemented By

Referenced By