Closed wtysos11 closed 4 years ago
没看过(标题)
以下为2019年以来的文章
摘要: With the development in the Cloud datacenters, the purpose of the efficient resource allocation is to meet the demand of the users instantly with the minimum rent cost. Thus, the elastic resource allocation strategy is usually combined with the prediction technology. This article proposes a novel predict method combination forecast technique, including both exponential smoothing (ES) and auto-regressive and polynomial fitting (PF) model. The aim of combination prediction is to achieve an efficient forecast technique according to the periodic and random feature of the workload and meet the application service level agreement (SLA) with the minimum cost. Moreover, the ES prediction with PSO algorithm gives a fine-grained scaling up and down the resources combining the heuristic algorithm in the future. APWP would solve the periodical or hybrid fluctuation of the workload in the cloud data centers. Finally, experiments improve that the combined prediction model meets the SLA with the better precision accuracy with the minimum renting cost. 预测式策略,使用功能了exponential smoothing and auto-regressive and polynomical fitting model,组合预测模型的目的是满足不同流量的需要同时满足服务SLA的要求使用PSO算法来进行一个细粒度的调度。用更低的租用成本实现更高的预测精度。
摘要: Most of the businesses nowadays have started using cloud platforms to host their software applications. A cloud platform is a shared resource that provides various services like software as a service (SAAS), infrastructure as a service (IAAS) or anything as a service (XAAS) that is required to develop and deploy any business application. These cloud services are provided as virtual machines (VM) that can handle the end-user’s requirements. The cloud providers have to ensure efficient resource handling mechanisms for different time intervals to avoid wastage of resources. Auto-scaling mechanisms would take care of using these resources appropriately along with providing an excellent quality of service. The researchers have used various approaches to perform auto-scaling. In this paper, a framework based on dynamic provisioning of cloud resources using workload prediction is discussed. 使用了预测式算法指导调度,但是并没有提及用了什么样的方法。
摘要: A distributed data stream processing system handles real-time, changeable and sudden streaming data load. Its elastic resource allocation has become a fundamental and challenging problem with a fixed strategy that will result in waste of resources or a reduction in QoS (quality of service). Spark Streaming as an emerging system has been developed to process real time stream data analytics by using micro-batch approach. In this paper, first, we propose an improved SVR (support vector regression) based stream data load prediction scheme. Then, we design a spark-based maximum sustainable throughput of time window (MSTW) performance model to find the optimized number of virtual machines. Finally, we present a resource scaling algorithm TWRES (time window resource elasticity scaling algorithm) with MSTW constraint and streaming data load prediction. The evaluation results show that TWRES could improve resource utilization and mitigate SLA (service level agreement) violation. 使用了SVR方法对流量数据进行预测,然后使用了spark-based maximum sustainable throughput of time window(MSTW)性能模型来找到最合适的VM数,然后使用TWRES(time window resource elasticity scaling algorithm)来进行调度。
摘要: Workload patterns of cloud applications are changing regularly. The workload prediction model is key for auto-scaling of resources in a cloud environment. It is helping with cost reduction and efficient resource utilization. The workload for the web applications is usually mixed for different application at different time span. The single prediction model is not able to predict different kinds of workload pattern of cloud applications. In this paper, an adaptive prediction model has been proposed using linear regression, ARIMA, and support vector regression for web applications. Workload classifier has been proposed to select the model as per workload features. Further the model parameters are selected through a heuristic approach. We have used real trace files to evaluate the proposed model with existing state-of-the-art models. The experiment results describe the significant improvement in root-mean-squared error and mean absolute percentage error metrics, and improve the quality of service of web applications in a cloud environment. 单一的预测模型不能够预测多种不同的流量,使用线性回归、ARIMA和SVR方法一起预测,使用RMSE来作为衡量指标。
摘要: Dynamic resource allocation and auto scalability are important aspects in mobile cloud computing environment. Predicting the cloud workload is a crucial task for dynamic resource allocation and auto scaling. Accuracy of workload prediction algorithm has significant impact on cloud quality of service and total cost of provided service. Since, existing prediction algorithms have competition for better accuracy and faster run time, in this paper we proposed a hybrid prediction algorithm to address both of these concerns. First we apply three level wavelet transform to decompose the workload time series into different resolution of time–frequency scales. An approximate and three details components. Second, we use support vector regression (SVR) for prediction of approximate and two low frequency detail components. The SVR parameters are tuned by a novel chaotic particle swarm optimization algorithm. Since the last detail component of time series has high frequency and is more likely to noise, we used generalized autoregressive conditional heteroskedasticity (GARCH) model to predict it. Finally, an ensemble method is applied to recompose these predicted samples from four multi scale predictions to achieve workload prediction for the next time step. The proposed method named wavelet decomposed 3 PSO optimized SVR plus GARCH (W3PSG). We evaluate the proposed W3PSG method with three different real cloud workload traces. Based on the results, the proposed method has relatively better prediction accuracy in comparison with competitive methods. According to mean absolute percentage error metric, in best case W3PSG method achieves 29.93%, 29.91%, and 24.53% of improvement in accuracy over three rival methods: GARCH, artificial neural network, and SVR respectively. 预测性模型,使用了三层的小波变换来分解时间序列到不同的尺度,然后使用SVR方法,SVR方法的参数使用chaotic particle swarm optimization方法调参。
摘要: Accurate estimation of data center resource utilization is a challenging task due to multi-tenant co-hosted applications having dynamic and time-varying workloads. Accurate estimation of future resources utilization helps in better job scheduling, workload placement, capacity planning, proactive auto-scaling, and load balancing. The inaccurate estimation leads to either under or over-provisioning of data center resources. Most existing estimation methods are based on a single model that often does not appropriately estimate different workload scenarios. To address these problems, we propose a novel method to adaptively and automatically identify the most appropriate model to accurately estimate data center resources utilization. The proposed approach trains a classifier based on statistical features of historical resources usage to decide the appropriate prediction model to use for given resource utilization observations collected during a specific time interval. We evaluated our approach on real datasets and compared the results with multiple baseline methods. The experimental evaluation shows that the proposed approach outperforms the state-of-the-art approaches and delivers 6% to 27% improved resource utilization estimation accuracy compared to baseline methods. 预测式方法,针对的是资源使用率。
摘要: The existing elastic control planes (ECPs) suffer from the immediacy and the computing overhead issues in software-defined data center networks (SD-DCNs). In this letter, we propose T-DCORAL which is a new ECP to mitigate both issues in SD-DCNs. T-DCORAL accelerates/decelerates the control plane by allocating a virtual CPU to controllers in runtime, whereas the existing ECPs resize the controller pool. As a result, T-DCORAL maximally reduces the latency to adjust the control plane from 46 s to 38 ms, the average CPU load by 22%, and the average rule installation time by 64.28%. 阈值法
摘要: Measurement of the dynamic elasticity of resource allocation in cloud computing continues to be a relevant problem in the related literature. Yet, there is scant evidence on determining the dynamic scaling quotient in such operations. Elasticity is defined as the ability to adapt to the changing workloads by provisioning and de-provisioning of Cloud resources and scaling is essential for maintaining elasticity in resource allocation. We propose ALVEC, as a model of resource allocation in Cloud data centers (Sarkar et al. , 2016) [7], [16], to address dynamic allocation by auto-tuning the model parameters. The proposed model, governed by a coupled differential equation known as Lotka Volterra (LV), fares better for management of Service Level Agreement (SLA) and Quality of Services (QoS). We show evidence of true elasticity both in theoretical and numerical applications. Additionally, we show that ALVEC as an example of unsupervised resource allocation, is able to predict the future load and allocate virtual machines efficiently.
摘要: Auto-scalers for clouds promise stable service quality at low costs when facing changing workload intensity. The major public cloud providers provide trigger-based auto-scalers based on thresholds. However, trigger-based auto-scaling has reaction times in the order of minutes. Novel auto-scalers from literature try to overcome the limitations of reactive mechanisms by employing proactive prediction methods. However, the adoption of proactive auto-scalers in production is still very low due to the high risk of relying on a single proactive method. This paper tackles the challenge of reducing this risk by proposing a new hybrid auto-scaling mechanism, called Chameleon, combining multiple different proactive methods coupled with a reactive fallback mechanism. Chameleon employs on-demand, automated time series-based forecasting methods to predict the arriving load intensity in combination with run-time service demand estimation to calculate the required resource consumption per work unit without the need for application instrumentation. We benchmark Chameleon against five different state-of-the-art proactive and reactive auto-scalers one in three different private and public cloud environments. We generate five different representative workloads each taken from different real-world system traces. Overall, Chameleon achieves the best scaling behavior based on user and elasticity performance metrics, analyzing the results from 400 hours aggregated experiment time. 含预测式模型的混合模型。
阅读和微服务相关的论文
综述文献:
高价值文献:
A reliable and cost-efficient auto-scaling system for web applications using heterogeneous spot instances
Renewable-aware geographical load balancing of web applications for sustainable data centers
Autonomic Resource Provisioning for Cloud-Based Software,开创性
A Comparison of Reinforcement Learning Techniques for Fuzzy Cloud Auto-Scaling
Pattern Matching Based Forecast of Non-periodic Repetitive Behavior for Cloud Clients
Integrated and Autonomic Cloud Resource Scaling
Optimal Cloud Resource Auto-Scaling for Web Applications
A cost-aware auto-scaling approach using the workload prediction in service clouds
Autoflex: Service Agnostic Auto-scaling Framework for IaaS Deployment Models
Towards an Autonomic Auto-Scaling Prediction System for Cloud Resource Provisioning
Dependable Horizontal Scaling Based on Probabilistic Model Checking,使用KMeans进行聚类复用。