語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Deep Generative Models for Image Rep...
~
Pu, Yunchen.
FindBook
Google Book
Amazon
博客來
Deep Generative Models for Image Representation Learning.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Deep Generative Models for Image Representation Learning./
作者:
Pu, Yunchen.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2018,
面頁冊數:
115 p.
附註:
Source: Dissertation Abstracts International, Volume: 79-09(E), Section: B.
Contained By:
Dissertation Abstracts International79-09B(E).
標題:
Artificial intelligence. -
電子資源:
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10745361
ISBN:
9780355872774
Deep Generative Models for Image Representation Learning.
Pu, Yunchen.
Deep Generative Models for Image Representation Learning.
- Ann Arbor : ProQuest Dissertations & Theses, 2018 - 115 p.
Source: Dissertation Abstracts International, Volume: 79-09(E), Section: B.
Thesis (Ph.D.)--Duke University, 2018.
Recently there has been increasing interest in developing generative models of data, offering the promise of learning based on the often vast quantity of unlabeled data. With such learning, one typically seeks to build rich, hierarchical probabilistic models that are able to fit to the distribution of complex real data, and are also capable of realistic data synthesis. In this dissertation, novel models and learning algorithms are proposed for deep generative models. This disseration consists of three main parts.
ISBN: 9780355872774Subjects--Topical Terms:
516317
Artificial intelligence.
Deep Generative Models for Image Representation Learning.
LDR
:04301nmm a2200337 4500
001
2162344
005
20180928111502.5
008
190424s2018 ||||||||||||||||| ||eng d
020
$a
9780355872774
035
$a
(MiAaPQ)AAI10745361
035
$a
(MiAaPQ)duke:14409
035
$a
AAI10745361
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Pu, Yunchen.
$3
3350329
245
1 0
$a
Deep Generative Models for Image Representation Learning.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2018
300
$a
115 p.
500
$a
Source: Dissertation Abstracts International, Volume: 79-09(E), Section: B.
500
$a
Adviser: Lawrence Carin.
502
$a
Thesis (Ph.D.)--Duke University, 2018.
520
$a
Recently there has been increasing interest in developing generative models of data, offering the promise of learning based on the often vast quantity of unlabeled data. With such learning, one typically seeks to build rich, hierarchical probabilistic models that are able to fit to the distribution of complex real data, and are also capable of realistic data synthesis. In this dissertation, novel models and learning algorithms are proposed for deep generative models. This disseration consists of three main parts.
520
$a
The first part developed a deep generative model joint analysis of images and associated labels or captions. The model is efficiently learned using variational autoencoder. A multilayered (deep) convolutional dictionary representation is employed as a decoder of the latent image features. Stochastic unpooling is employed to link consecutive layers in the image model, yielding top-down image generation. A deep Convolutional Neural Network (CNN) is used as an image encoder; the CNN is used to approximate a distribution for the latent DGDN features/code. The latent code is also linked to generative models for labels (Bayesian support vector machine) or captions (recurrent neural network). When predicting a label/caption for a new image at test, averaging is performed across the distribution of latent codes; this is computationally efficient as a consequence of the learned CNN-based encoder. Since the framework is capable of modeling the image in the presence/absence of associated labels/captions, a new semi-supervised setting is manifested for CNN learning with images; the framework even allows unsupervised CNN learning, based on images alone. Excellent results are obtained on several benchmark datasets, including ImageNet, demonstrating that the proposed model achieves results that are highly competitive with similarly sized convolutional neural networks.
520
$a
The second part developed a new method for learning variational autoencoders (VAEs), based on Stein variational gradient descent. A key advantage of this approach is that one need not make parametric assumptions about the form of the encoder distribution. Performance is further enhanced by integrating the proposed encoder with importance sampling. Excellent performance is demonstrated across multiple unsupervised and semi-supervised problems, including semi-supervised analysis of the ImageNet data, demonstrating the scalability of the model to large datasets.
520
$a
The third part developed a new form of variational autoencoder, in which the joint distribution of data and codes is considered in two (symmetric) forms: (i) from observed data fed through the encoder to yield codes, and (ii) from latent codes drawn from a simple prior and propagated through the decoder to manifest data. Lower bounds are learned for marginal log-likelihood fits observed data and latent codes. When learning with the variational bound, one seeks to minimize the symmetric Kullback-Leibler divergence of joint density functions from (i) and (ii), while simultaneously seeking to maximize the two marginal log-likelihoods. To facilitate learning, a new form of adversarial training is developed. An extensive set of experiments is performed, in which we demonstrate state-of-the-art data reconstruction and generation on several image benchmark datasets.
590
$a
School code: 0066.
650
4
$a
Artificial intelligence.
$3
516317
650
4
$a
Canadian history.
$3
3174996
690
$a
0800
690
$a
0334
710
2
$a
Duke University.
$b
Electrical and Computer Engineering.
$3
1032075
773
0
$t
Dissertation Abstracts International
$g
79-09B(E).
790
$a
0066
791
$a
Ph.D.
792
$a
2018
793
$a
English
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10745361
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9361891
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入