語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
FindBook
Google Book
Amazon
博客來
Cost-Efficient Dynamic Management of Cloud Resources Via Supervised Learning.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Cost-Efficient Dynamic Management of Cloud Resources Via Supervised Learning./
作者:
Wajahat, Muhammad.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2020,
面頁冊數:
200 p.
附註:
Source: Dissertations Abstracts International, Volume: 82-04, Section: B.
Contained By:
Dissertations Abstracts International82-04B.
標題:
Computer science. -
電子資源:
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=28091349
ISBN:
9798678140364
Cost-Efficient Dynamic Management of Cloud Resources Via Supervised Learning.
Wajahat, Muhammad.
Cost-Efficient Dynamic Management of Cloud Resources Via Supervised Learning.
- Ann Arbor : ProQuest Dissertations & Theses, 2020 - 200 p.
Source: Dissertations Abstracts International, Volume: 82-04, Section: B.
Thesis (Ph.D.)--State University of New York at Stony Brook, 2020.
This item must not be sold to any third party vendors.
In the past decade, cloud computing has disrupted the computing landscape by providing economical compute and storage resources as services to end-users. Many web-based internet applications, such as Airbnb, Pinterest, and Expedia, are now hosted on public clouds because of the cost-effectiveness and scalability that the cloud provides. However, for cloud providers such as Amazon and Microsoft, providing cloud computing resources on-demand with high availability and elasticity features to latency-sensitive clients requires careful management of these resources to avoid client abandonment. Specifically, the allocation and reallocation of shared computing resources among clients must be done carefully to meet their SLA performance requirements and to realize the full potential of cloud computing -- increasing resource utilization and lowering costs. Prior works either focus on empirically tuning the amount of resources for a given application or develop analytical performance models with the goal of determining the optimal resource allocation for applications. Unfortunately, both approaches have their shortcomings. While the former is application-specific and not generalizable, the latter requires expert application and system knowledge to build accurate models.To solve the resource management problem for different application classes, we need to understand the workload characteristics as well as the relationship between resource allocation and system performance. In this thesis, we use supervised learning techniques to accomplish these objectives. To accomplish the first objective, we use learning and optimization to analyze and model the request-level characteristics of production traces, and show how certain models can be developed to answer questions about system performance under dynamic conditions. These models can also be used to optimally provision resources when a change in workload is known (or can be predicted) ahead of time. However, changes in workload are rarely known in advance, which makes dynamic resource allocation more challenging. To accomplish the second objective of dynamic resource allocation under uncertain workload changes, we consider two scenarios: (1) The scenario where the underlying physical servers that host the cloud are always available. For this scenario, we solve the problem of autoscaling -- the dynamic provisioning of cloud resources for applications that experience time-varying load. (2) The scenario where the underlying servers themselves might become unavailable. For this scenario, we solve the problem of rehoming -- the reallocation of resources across servers for cloud-hosted applications. To analyze and understand workload characteristics, we employ distribution fitting to request-level characteristics of production traces. More specifically, we model the inter-arrival times and service times to well known distributions (or mixtures of them) that enables powerful predictive capabilities in the form of Markov Chain models. These models not only predict system performance, but are also capable of answering what-if questions with respect to changes in workload or the system state. We present our analysis of more than 1600 production traces from storage and internet applications, and show that the Hyperexponential distribution (H2) is well suited for such workloads and can be leveraged to build powerful models due to its properties (i.e. Markovian, heavy-tailed, flexible).To address the autoscaling challenge, we propose MLscale, a machine learning based application-agnostic autoscaler that dynamically allocates and deallocates computing resources for cloud applications in response to unpredictable changes in workload demand. MLscale provides near optimal resource usage while reducing SLA violations without requiring expert application knowledge or tuning. We systematically evaluate the autoscaling of multi-tiered applications from downstream to upstream tiers, starting with the application tier and ending with the persistent storage tier.To address rehoming, we propose MERIT, a model-driven rehoming solution for reallocating compute resources for existing services. In certain cloud environments, unpredictable events such as hotspots or system upgrades, require the evacuation of certain physical servers, thus necessitating the timely resource reallocation of hosted applications through various rehoming actions like live-migration or rebuilding. To optimize rehoming, we employ machine learning techniques, including regression and neural nets, to build cost models for various rehoming options. We propose to leverage these models, to find the optimal rehoming strategy for distributed applications that minimizes the rehoming resource cost and application downtime.It is our thesis that learning-based techniques provide the right balance between accurate but application-specific empirical approaches and generalized but possibly inaccurate analytical approaches towards resource management in the cloud. Such learning-based resource management techniques can greatly enhance the cost-savings of a cloud provider (by up to 40% in our implementation results) while providing the necessary performance SLA guarantees.
ISBN: 9798678140364Subjects--Topical Terms:
523869
Computer science.
Subjects--Index Terms:
Autoscaling
Cost-Efficient Dynamic Management of Cloud Resources Via Supervised Learning.
LDR
:06403nmm a2200373 4500
001
2344587
005
20220531064600.5
008
241004s2020 ||||||||||||||||| ||eng d
020
$a
9798678140364
035
$a
(MiAaPQ)AAI28091349
035
$a
AAI28091349
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Wajahat, Muhammad.
$3
3683373
245
1 0
$a
Cost-Efficient Dynamic Management of Cloud Resources Via Supervised Learning.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2020
300
$a
200 p.
500
$a
Source: Dissertations Abstracts International, Volume: 82-04, Section: B.
500
$a
Advisor: Gandhi, Anshul.
502
$a
Thesis (Ph.D.)--State University of New York at Stony Brook, 2020.
506
$a
This item must not be sold to any third party vendors.
520
$a
In the past decade, cloud computing has disrupted the computing landscape by providing economical compute and storage resources as services to end-users. Many web-based internet applications, such as Airbnb, Pinterest, and Expedia, are now hosted on public clouds because of the cost-effectiveness and scalability that the cloud provides. However, for cloud providers such as Amazon and Microsoft, providing cloud computing resources on-demand with high availability and elasticity features to latency-sensitive clients requires careful management of these resources to avoid client abandonment. Specifically, the allocation and reallocation of shared computing resources among clients must be done carefully to meet their SLA performance requirements and to realize the full potential of cloud computing -- increasing resource utilization and lowering costs. Prior works either focus on empirically tuning the amount of resources for a given application or develop analytical performance models with the goal of determining the optimal resource allocation for applications. Unfortunately, both approaches have their shortcomings. While the former is application-specific and not generalizable, the latter requires expert application and system knowledge to build accurate models.To solve the resource management problem for different application classes, we need to understand the workload characteristics as well as the relationship between resource allocation and system performance. In this thesis, we use supervised learning techniques to accomplish these objectives. To accomplish the first objective, we use learning and optimization to analyze and model the request-level characteristics of production traces, and show how certain models can be developed to answer questions about system performance under dynamic conditions. These models can also be used to optimally provision resources when a change in workload is known (or can be predicted) ahead of time. However, changes in workload are rarely known in advance, which makes dynamic resource allocation more challenging. To accomplish the second objective of dynamic resource allocation under uncertain workload changes, we consider two scenarios: (1) The scenario where the underlying physical servers that host the cloud are always available. For this scenario, we solve the problem of autoscaling -- the dynamic provisioning of cloud resources for applications that experience time-varying load. (2) The scenario where the underlying servers themselves might become unavailable. For this scenario, we solve the problem of rehoming -- the reallocation of resources across servers for cloud-hosted applications. To analyze and understand workload characteristics, we employ distribution fitting to request-level characteristics of production traces. More specifically, we model the inter-arrival times and service times to well known distributions (or mixtures of them) that enables powerful predictive capabilities in the form of Markov Chain models. These models not only predict system performance, but are also capable of answering what-if questions with respect to changes in workload or the system state. We present our analysis of more than 1600 production traces from storage and internet applications, and show that the Hyperexponential distribution (H2) is well suited for such workloads and can be leveraged to build powerful models due to its properties (i.e. Markovian, heavy-tailed, flexible).To address the autoscaling challenge, we propose MLscale, a machine learning based application-agnostic autoscaler that dynamically allocates and deallocates computing resources for cloud applications in response to unpredictable changes in workload demand. MLscale provides near optimal resource usage while reducing SLA violations without requiring expert application knowledge or tuning. We systematically evaluate the autoscaling of multi-tiered applications from downstream to upstream tiers, starting with the application tier and ending with the persistent storage tier.To address rehoming, we propose MERIT, a model-driven rehoming solution for reallocating compute resources for existing services. In certain cloud environments, unpredictable events such as hotspots or system upgrades, require the evacuation of certain physical servers, thus necessitating the timely resource reallocation of hosted applications through various rehoming actions like live-migration or rebuilding. To optimize rehoming, we employ machine learning techniques, including regression and neural nets, to build cost models for various rehoming options. We propose to leverage these models, to find the optimal rehoming strategy for distributed applications that minimizes the rehoming resource cost and application downtime.It is our thesis that learning-based techniques provide the right balance between accurate but application-specific empirical approaches and generalized but possibly inaccurate analytical approaches towards resource management in the cloud. Such learning-based resource management techniques can greatly enhance the cost-savings of a cloud provider (by up to 40% in our implementation results) while providing the necessary performance SLA guarantees.
590
$a
School code: 0771.
650
4
$a
Computer science.
$3
523869
650
4
$a
Information technology.
$3
532993
650
4
$a
Data base management.
$3
528121
653
$a
Autoscaling
653
$a
Cloud computing
653
$a
Modeling
653
$a
Performance
653
$a
Resource management
690
$a
0984
690
$a
0489
690
$a
0454
710
2
$a
State University of New York at Stony Brook.
$b
Computer Science.
$3
1674709
773
0
$t
Dissertations Abstracts International
$g
82-04B.
790
$a
0771
791
$a
Ph.D.
792
$a
2020
793
$a
English
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=28091349
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9467025
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入