Language:
English
繁體中文
Help
回圖書館首頁
手機版館藏查詢
Login
Back
Switch To:
Labeled
|
MARC Mode
|
ISBD
Acceleration of Machine Learning Alg...
~
Neshatpour, Katayoun.
Linked to FindBook
Google Book
Amazon
博客來
Acceleration of Machine Learning Algorithms for Big Data Applications.
Record Type:
Electronic resources : Monograph/item
Title/Author:
Acceleration of Machine Learning Algorithms for Big Data Applications./
Author:
Neshatpour, Katayoun.
Published:
Ann Arbor : ProQuest Dissertations & Theses, : 2018,
Description:
187 p.
Notes:
Source: Dissertations Abstracts International, Volume: 80-09, Section: B.
Contained By:
Dissertations Abstracts International80-09B.
Subject:
Computer Engineering. -
Online resource:
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=13420481
ISBN:
9780438887251
Acceleration of Machine Learning Algorithms for Big Data Applications.
Neshatpour, Katayoun.
Acceleration of Machine Learning Algorithms for Big Data Applications.
- Ann Arbor : ProQuest Dissertations & Theses, 2018 - 187 p.
Source: Dissertations Abstracts International, Volume: 80-09, Section: B.
Thesis (Ph.D.)--George Mason University, 2018.
This item must not be sold to any third party vendors.
Advances in various branches of technology - data sensing, data communication, data computation, and data storage have significantly changed machine learning in recent years. With new computing technologies allowing vast amount of data to be stored and creating opportunities to learn from the aggregated data, implementation of classical machine-learning algorithms including K-means, KNN, has been improved to meet the big data requirements of such applications. Moreover, the enhanced processing power of today's computing nodes enables training of more sophisticated learning models, paving the way for deep learning algorithms including convolutional neural networks (CNN). In big data domain, various frameworks have been developed that allow the processing of large data sets with parallel and distributed algorithms, with MapReduce being an example of such frameworks. The first section of this research focuses on classical Machine-Learning applications and their hardware acceleration in MapReduce platform. Profiling of applications in this research shows that the map and/or reduce phase of these applications take up most of the execution time. Subsequently, the map functions were accelerated to evaluate how a cluster of CPUs equipped with FPGAs uses the accelerated mapper to enhance the overall performance of MapReduce. Moreover, this research studies how the type of FPGA (low-end vs. high-end), and its integration with the CPU (on-chip vs. off-chip) along with the choice of CPU (high performance big vs. low power little servers) affects the speedup yield and power reduction. While FPGA acceleration of MapReduce offers superior energy efficiency, but MapReduce platform is not suitable for implementation of deep neural networks including CNNs. In the second part of this research an iterative approach is proposed to break down large CNNs into a sequence of smaller networks (uCNN), each processing a sub-sample of the input image, providing the ability to terminate the classification early or carry the classification to the next iteration in case of non-satisfactory confidence levels. Moreover, the contextual information resulted from early iterations of ICNN can be used to reduce the complexity of the subsequent iterations. To explore the complexity-accuracy tradeoff of ICNN, a dynamic deadline-driven exit policy for real-time applications, a confidence thresholding policy for dynamic complexity reduction, a context-aware pruning policy for parameter reduction and two hybrid pruning and thresholding policies for simultaneous parameter and complexity reduction were introduced. Simulation results on a case study with iterative AlexNet shows that with intelligent selection of the pruning and/or thresholding policies, ICNN reduces the average FLOP and parameter counts, and execution time across 50K validation images in ImageNet database by more than 25%, 80% and 38%, respectively, with negligible accuracy loss. Moreover, the real-time systems could exploit the dynamic structure of the ICNN by reducing the execution time by upto 12x by trading off accuracy with execution time.
ISBN: 9780438887251Subjects--Topical Terms:
1567821
Computer Engineering.
Acceleration of Machine Learning Algorithms for Big Data Applications.
LDR
:04202nmm a2200325 4500
001
2209177
005
20191025102849.5
008
201008s2018 ||||||||||||||||| ||eng d
020
$a
9780438887251
035
$a
(MiAaPQ)AAI13420481
035
$a
(MiAaPQ)gmu:11901
035
$a
AAI13420481
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Neshatpour, Katayoun.
$3
3436256
245
1 0
$a
Acceleration of Machine Learning Algorithms for Big Data Applications.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2018
300
$a
187 p.
500
$a
Source: Dissertations Abstracts International, Volume: 80-09, Section: B.
500
$a
Publisher info.: Dissertation/Thesis.
500
$a
Advisor: Homayoun, Houman;Sasan, Avesta.
502
$a
Thesis (Ph.D.)--George Mason University, 2018.
506
$a
This item must not be sold to any third party vendors.
520
$a
Advances in various branches of technology - data sensing, data communication, data computation, and data storage have significantly changed machine learning in recent years. With new computing technologies allowing vast amount of data to be stored and creating opportunities to learn from the aggregated data, implementation of classical machine-learning algorithms including K-means, KNN, has been improved to meet the big data requirements of such applications. Moreover, the enhanced processing power of today's computing nodes enables training of more sophisticated learning models, paving the way for deep learning algorithms including convolutional neural networks (CNN). In big data domain, various frameworks have been developed that allow the processing of large data sets with parallel and distributed algorithms, with MapReduce being an example of such frameworks. The first section of this research focuses on classical Machine-Learning applications and their hardware acceleration in MapReduce platform. Profiling of applications in this research shows that the map and/or reduce phase of these applications take up most of the execution time. Subsequently, the map functions were accelerated to evaluate how a cluster of CPUs equipped with FPGAs uses the accelerated mapper to enhance the overall performance of MapReduce. Moreover, this research studies how the type of FPGA (low-end vs. high-end), and its integration with the CPU (on-chip vs. off-chip) along with the choice of CPU (high performance big vs. low power little servers) affects the speedup yield and power reduction. While FPGA acceleration of MapReduce offers superior energy efficiency, but MapReduce platform is not suitable for implementation of deep neural networks including CNNs. In the second part of this research an iterative approach is proposed to break down large CNNs into a sequence of smaller networks (uCNN), each processing a sub-sample of the input image, providing the ability to terminate the classification early or carry the classification to the next iteration in case of non-satisfactory confidence levels. Moreover, the contextual information resulted from early iterations of ICNN can be used to reduce the complexity of the subsequent iterations. To explore the complexity-accuracy tradeoff of ICNN, a dynamic deadline-driven exit policy for real-time applications, a confidence thresholding policy for dynamic complexity reduction, a context-aware pruning policy for parameter reduction and two hybrid pruning and thresholding policies for simultaneous parameter and complexity reduction were introduced. Simulation results on a case study with iterative AlexNet shows that with intelligent selection of the pruning and/or thresholding policies, ICNN reduces the average FLOP and parameter counts, and execution time across 50K validation images in ImageNet database by more than 25%, 80% and 38%, respectively, with negligible accuracy loss. Moreover, the real-time systems could exploit the dynamic structure of the ICNN by reducing the execution time by upto 12x by trading off accuracy with execution time.
590
$a
School code: 0883.
650
4
$a
Computer Engineering.
$3
1567821
650
4
$a
Artificial intelligence.
$3
516317
690
$a
0464
690
$a
0800
710
2
$a
George Mason University.
$b
Electrical and Computer Engineering.
$3
3193956
773
0
$t
Dissertations Abstracts International
$g
80-09B.
790
$a
0883
791
$a
Ph.D.
792
$a
2018
793
$a
English
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=13420481
based on 0 review(s)
Location:
ALL
電子資源
Year:
Volume Number:
Items
1 records • Pages 1 •
1
Inventory Number
Location Name
Item Class
Material type
Call number
Usage Class
Loan Status
No. of reservations
Opac note
Attachments
W9385726
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
On shelf
0
1 records • Pages 1 •
1
Multimedia
Reviews
Add a review
and share your thoughts with other readers
Export
pickup library
Processing
...
Change password
Login