語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Learning Representations in Reinforc...
~
Rafati Heravi, Jacob.
FindBook
Google Book
Amazon
博客來
Learning Representations in Reinforcement Learning.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Learning Representations in Reinforcement Learning./
作者:
Rafati Heravi, Jacob.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2019,
面頁冊數:
155 p.
附註:
Source: Dissertations Abstracts International, Volume: 81-03, Section: B.
Contained By:
Dissertations Abstracts International81-03B.
標題:
Artificial intelligence. -
電子資源:
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=13880098
ISBN:
9781085630283
Learning Representations in Reinforcement Learning.
Rafati Heravi, Jacob.
Learning Representations in Reinforcement Learning.
- Ann Arbor : ProQuest Dissertations & Theses, 2019 - 155 p.
Source: Dissertations Abstracts International, Volume: 81-03, Section: B.
Thesis (Ph.D.)--University of California, Merced, 2019.
This item must not be sold to any third party vendors.
Reinforcement Learning (RL) algorithms allow artificial agents to improve their action selection policy to increase rewarding experiences in their environments. Temporal Difference (TD) learning algorithm, a model-free RL method, attempts to find an optimal policy through learning the values of agent's actions at any state by computing the expected future rewards without having access to a model of the environment. TD algorithms have been very successful on a broad range of control tasks, but learning can become intractably slow as the state space grows. This has motivated methods for using parameterized function approximation for the value function and developing methods for learning internal representations of the agent's state, to effectively reduce the size of state space and restructure state representations in order to support generalization. This dissertation investigates biologically inspired techniques for learning useful state representations in RL, as well as optimization methods for improving learning. There are three parts to this investigation. First, failures of deep RL algorithms to solve some relatively simple control problems are explored. Taking inspiration from the sparse codes produced by lateral inhibition in the brain, this dissertation offers a method for learning sparse state representations. Second, the challenges of RL in efficient exploration of environments with sparse delayed reward feedback, as well as the scalability issues in large-scale applications are addressed. The hierarchical structure of motor control in the brain prompts the consideration of approaches to learning action selection policies at multiple levels of temporal abstraction. That is learning to select subgoals separately from action selection policies that achieve those subgoals. This dissertation offers a novel model-free Hierarchical Reinforcement Learning framework, including approaches to automatic subgoal discovery based on unsupervised learning over memories of past experiences. Third, more complex optimization methods than those typically used in deep learning, and deep RL are explored, focusing on improving learning while avoiding the need to fine tune many hyperparameters. This dissertation offers limited-memory quasi-Newton optimization methods to efficiently solve highly nonlinear and nonconvex optimization problems for deep learning and deep RL applications. Together, these three contributions provide a foundation for scaling RL to more complex control problems through the learning of improved internal representations.
ISBN: 9781085630283Subjects--Topical Terms:
516317
Artificial intelligence.
Learning Representations in Reinforcement Learning.
LDR
:03557nmm a2200313 4500
001
2265335
005
20200514111955.5
008
220629s2019 ||||||||||||||||| ||eng d
020
$a
9781085630283
035
$a
(MiAaPQ)AAI13880098
035
$a
AAI13880098
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Rafati Heravi, Jacob.
$3
3542492
245
1 0
$a
Learning Representations in Reinforcement Learning.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2019
300
$a
155 p.
500
$a
Source: Dissertations Abstracts International, Volume: 81-03, Section: B.
500
$a
Advisor: Noelle, David C.
502
$a
Thesis (Ph.D.)--University of California, Merced, 2019.
506
$a
This item must not be sold to any third party vendors.
520
$a
Reinforcement Learning (RL) algorithms allow artificial agents to improve their action selection policy to increase rewarding experiences in their environments. Temporal Difference (TD) learning algorithm, a model-free RL method, attempts to find an optimal policy through learning the values of agent's actions at any state by computing the expected future rewards without having access to a model of the environment. TD algorithms have been very successful on a broad range of control tasks, but learning can become intractably slow as the state space grows. This has motivated methods for using parameterized function approximation for the value function and developing methods for learning internal representations of the agent's state, to effectively reduce the size of state space and restructure state representations in order to support generalization. This dissertation investigates biologically inspired techniques for learning useful state representations in RL, as well as optimization methods for improving learning. There are three parts to this investigation. First, failures of deep RL algorithms to solve some relatively simple control problems are explored. Taking inspiration from the sparse codes produced by lateral inhibition in the brain, this dissertation offers a method for learning sparse state representations. Second, the challenges of RL in efficient exploration of environments with sparse delayed reward feedback, as well as the scalability issues in large-scale applications are addressed. The hierarchical structure of motor control in the brain prompts the consideration of approaches to learning action selection policies at multiple levels of temporal abstraction. That is learning to select subgoals separately from action selection policies that achieve those subgoals. This dissertation offers a novel model-free Hierarchical Reinforcement Learning framework, including approaches to automatic subgoal discovery based on unsupervised learning over memories of past experiences. Third, more complex optimization methods than those typically used in deep learning, and deep RL are explored, focusing on improving learning while avoiding the need to fine tune many hyperparameters. This dissertation offers limited-memory quasi-Newton optimization methods to efficiently solve highly nonlinear and nonconvex optimization problems for deep learning and deep RL applications. Together, these three contributions provide a foundation for scaling RL to more complex control problems through the learning of improved internal representations.
590
$a
School code: 1660.
650
4
$a
Artificial intelligence.
$3
516317
650
4
$a
Computer science.
$3
523869
650
4
$a
Applied mathematics.
$3
2122814
690
$a
0800
690
$a
0984
690
$a
0364
710
2
$a
University of California, Merced.
$b
Electrical Engineering and Computer Science.
$3
2095014
773
0
$t
Dissertations Abstracts International
$g
81-03B.
790
$a
1660
791
$a
Ph.D.
792
$a
2019
793
$a
English
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=13880098
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9417569
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入