語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Acceleration of Machine Learning Alg...
~
Neshatpour, Katayoun.
FindBook
Google Book
Amazon
博客來
Acceleration of Machine Learning Algorithms for Big Data Applications.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Acceleration of Machine Learning Algorithms for Big Data Applications./
作者:
Neshatpour, Katayoun.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2018,
面頁冊數:
187 p.
附註:
Source: Dissertations Abstracts International, Volume: 80-09, Section: B.
Contained By:
Dissertations Abstracts International80-09B.
標題:
Computer Engineering. -
電子資源:
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=13420481
ISBN:
9780438887251
Acceleration of Machine Learning Algorithms for Big Data Applications.
Neshatpour, Katayoun.
Acceleration of Machine Learning Algorithms for Big Data Applications.
- Ann Arbor : ProQuest Dissertations & Theses, 2018 - 187 p.
Source: Dissertations Abstracts International, Volume: 80-09, Section: B.
Thesis (Ph.D.)--George Mason University, 2018.
This item must not be sold to any third party vendors.
Advances in various branches of technology - data sensing, data communication, data computation, and data storage have significantly changed machine learning in recent years. With new computing technologies allowing vast amount of data to be stored and creating opportunities to learn from the aggregated data, implementation of classical machine-learning algorithms including K-means, KNN, has been improved to meet the big data requirements of such applications. Moreover, the enhanced processing power of today's computing nodes enables training of more sophisticated learning models, paving the way for deep learning algorithms including convolutional neural networks (CNN). In big data domain, various frameworks have been developed that allow the processing of large data sets with parallel and distributed algorithms, with MapReduce being an example of such frameworks. The first section of this research focuses on classical Machine-Learning applications and their hardware acceleration in MapReduce platform. Profiling of applications in this research shows that the map and/or reduce phase of these applications take up most of the execution time. Subsequently, the map functions were accelerated to evaluate how a cluster of CPUs equipped with FPGAs uses the accelerated mapper to enhance the overall performance of MapReduce. Moreover, this research studies how the type of FPGA (low-end vs. high-end), and its integration with the CPU (on-chip vs. off-chip) along with the choice of CPU (high performance big vs. low power little servers) affects the speedup yield and power reduction. While FPGA acceleration of MapReduce offers superior energy efficiency, but MapReduce platform is not suitable for implementation of deep neural networks including CNNs. In the second part of this research an iterative approach is proposed to break down large CNNs into a sequence of smaller networks (uCNN), each processing a sub-sample of the input image, providing the ability to terminate the classification early or carry the classification to the next iteration in case of non-satisfactory confidence levels. Moreover, the contextual information resulted from early iterations of ICNN can be used to reduce the complexity of the subsequent iterations. To explore the complexity-accuracy tradeoff of ICNN, a dynamic deadline-driven exit policy for real-time applications, a confidence thresholding policy for dynamic complexity reduction, a context-aware pruning policy for parameter reduction and two hybrid pruning and thresholding policies for simultaneous parameter and complexity reduction were introduced. Simulation results on a case study with iterative AlexNet shows that with intelligent selection of the pruning and/or thresholding policies, ICNN reduces the average FLOP and parameter counts, and execution time across 50K validation images in ImageNet database by more than 25%, 80% and 38%, respectively, with negligible accuracy loss. Moreover, the real-time systems could exploit the dynamic structure of the ICNN by reducing the execution time by upto 12x by trading off accuracy with execution time.
ISBN: 9780438887251Subjects--Topical Terms:
1567821
Computer Engineering.
Acceleration of Machine Learning Algorithms for Big Data Applications.
LDR
:04202nmm a2200325 4500
001
2209177
005
20191025102849.5
008
201008s2018 ||||||||||||||||| ||eng d
020
$a
9780438887251
035
$a
(MiAaPQ)AAI13420481
035
$a
(MiAaPQ)gmu:11901
035
$a
AAI13420481
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Neshatpour, Katayoun.
$3
3436256
245
1 0
$a
Acceleration of Machine Learning Algorithms for Big Data Applications.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2018
300
$a
187 p.
500
$a
Source: Dissertations Abstracts International, Volume: 80-09, Section: B.
500
$a
Publisher info.: Dissertation/Thesis.
500
$a
Advisor: Homayoun, Houman;Sasan, Avesta.
502
$a
Thesis (Ph.D.)--George Mason University, 2018.
506
$a
This item must not be sold to any third party vendors.
520
$a
Advances in various branches of technology - data sensing, data communication, data computation, and data storage have significantly changed machine learning in recent years. With new computing technologies allowing vast amount of data to be stored and creating opportunities to learn from the aggregated data, implementation of classical machine-learning algorithms including K-means, KNN, has been improved to meet the big data requirements of such applications. Moreover, the enhanced processing power of today's computing nodes enables training of more sophisticated learning models, paving the way for deep learning algorithms including convolutional neural networks (CNN). In big data domain, various frameworks have been developed that allow the processing of large data sets with parallel and distributed algorithms, with MapReduce being an example of such frameworks. The first section of this research focuses on classical Machine-Learning applications and their hardware acceleration in MapReduce platform. Profiling of applications in this research shows that the map and/or reduce phase of these applications take up most of the execution time. Subsequently, the map functions were accelerated to evaluate how a cluster of CPUs equipped with FPGAs uses the accelerated mapper to enhance the overall performance of MapReduce. Moreover, this research studies how the type of FPGA (low-end vs. high-end), and its integration with the CPU (on-chip vs. off-chip) along with the choice of CPU (high performance big vs. low power little servers) affects the speedup yield and power reduction. While FPGA acceleration of MapReduce offers superior energy efficiency, but MapReduce platform is not suitable for implementation of deep neural networks including CNNs. In the second part of this research an iterative approach is proposed to break down large CNNs into a sequence of smaller networks (uCNN), each processing a sub-sample of the input image, providing the ability to terminate the classification early or carry the classification to the next iteration in case of non-satisfactory confidence levels. Moreover, the contextual information resulted from early iterations of ICNN can be used to reduce the complexity of the subsequent iterations. To explore the complexity-accuracy tradeoff of ICNN, a dynamic deadline-driven exit policy for real-time applications, a confidence thresholding policy for dynamic complexity reduction, a context-aware pruning policy for parameter reduction and two hybrid pruning and thresholding policies for simultaneous parameter and complexity reduction were introduced. Simulation results on a case study with iterative AlexNet shows that with intelligent selection of the pruning and/or thresholding policies, ICNN reduces the average FLOP and parameter counts, and execution time across 50K validation images in ImageNet database by more than 25%, 80% and 38%, respectively, with negligible accuracy loss. Moreover, the real-time systems could exploit the dynamic structure of the ICNN by reducing the execution time by upto 12x by trading off accuracy with execution time.
590
$a
School code: 0883.
650
4
$a
Computer Engineering.
$3
1567821
650
4
$a
Artificial intelligence.
$3
516317
690
$a
0464
690
$a
0800
710
2
$a
George Mason University.
$b
Electrical and Computer Engineering.
$3
3193956
773
0
$t
Dissertations Abstracts International
$g
80-09B.
790
$a
0883
791
$a
Ph.D.
792
$a
2018
793
$a
English
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=13420481
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9385726
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入