Frontiers of intelligent control and...
Liu, Derong, (1963-)

FindBook      Google Book      Amazon      博客來     
  • Frontiers of intelligent control and information processing
  • 紀錄類型: 書目-電子資源 : Monograph/item
    正題名/作者: Frontiers of intelligent control and information processing/ edited by Derong Liu
    其他作者: Liu, Derong,
    出版者: [Hackensack?] New Jersey :World Scientific, : 2014.,
    面頁冊數: 1 online resource.
    內容註: Preface; Contents; 1. Dynamic Graphical Games: Online Adaptive Learning Solutions Using Approximate Dynamic Programming; 1.1 Introduction; 1.2 Graphs and Synchronization of Multi-Agent Dynamical Systems; 1.2.1 Graphs; 1.2.2 Synchronization and tracking error dynamics; 1.3 Multiple Player CooperativeGames on Graphs; 1.3.1 Graphical games; 1.3.2 Comparison of graphical games with standard dynamic games; 1.3.3 Nash equilibrium for graphical games; 1.3.4 Hamiltonian equation for dynamic graphical games; 1.3.5 Bellman equation for dynamic graphical games.
    內容註: 1.3.6 Discrete Hamilton-Jacobi theory: Equivalence of Bellman and discrete-time Hamilton Jacobi equations1.3.7 Stability and Nash solution of the graphical games; 1.4 Approximate Dynamic Programming for Graphical Games; 1.4.1 Heuristic dynamic programming for graphical games; 1.4.2 Dual heuristic programming for graphical games; 1.5 Coupled Riccati Recursions; 1.6 Graphical Game Solutions by Actor-Critic Learning; 1.6.1 Actor-critic networks and tuning; 1.6.2 Actor-critic offline tuning with exploration; 1.6.3 Actor-critic online tuning in real-time.
    內容註: 1.7 Graphical Game Example and Simulation Results1.7.1 Riccati recursion offline solution; 1.7.2 Simulation results using offline actor-critic tuning; 1.7.3 Simulation results using online actor-critic tuning; 1.8 Conclusions; Acknowledgement; References; 2. Reinforcement-Learning-Based Online Learning Control for Discrete-Time Unknown Nonaffine Nonlinear Systems; 2.1 Introduction; 2.2 Problem Statement and Preliminaries; 2.2.1 Dynamics of nonaffine nonlinear discrete-time systems; 2.2.2 A single-hidden layer neural network; 2.3 Controller Design via Reinforcement Learning.
    內容註: 2.3.1 A basic controller design approach2.3.2 Critic neural network and weight update law; 2.3.3 Action neural network and weight update law; 2.4 Stability Analysis and Performance of the Closed-Loop System; 2.5 Numerical Examples; 2.5.1 Example 1; 2.5.2 Example 2; 2.6 Conclusions; Acknowledgement; References; 3. Experimental Studies on Data-Driven Heuristic Dynamic Programming for POMDP; 3.1 Introduction; 3.2 Markov Decision Process and Partially Observable Markov Decision Process; 3.2.1 Markov decision process; 3.2.2 Partially observable Markov decision process.
    內容註: 3.3 Problem Formulation with the State Estimator3.4 Data-Driven HDP Algorithm for POMDP; 3.4.1 Learning in the state estimator network; 3.4.2 Learning in the critic and the action network; 3.5 Simulation Study; 3.5.1 Case study one; 3.5.2 Case study two; 3.5.3 Case study three; 3.6 Conclusions and Discussion; Acknowledgement; References; 4. Online Reinforcement Learning for Continuous-State Systems; 4.1 Introduction; 4.2 Background of Reinforcement Learning; 4.3 RLSPI Algorithm; 4.3.1 Policy iteration; 4.3.2 RLSPI; 4.4 Examples of RLSPI; 4.4.1 Linear discrete-time system.
    標題: Automatic control. -
    電子資源: http://www.worldscientific.com/worldscibooks/10.1142/9243#t=toc
    ISBN: 9789814616881
館藏地:  出版年:  卷號: 
館藏
  • 1 筆 • 頁數 1 •
 
W9269111 電子資源 01.外借(書)_YB 電子書 EB TJ216 .F76 2014eb 一般使用(Normal) 在架 0
  • 1 筆 • 頁數 1 •
多媒體
評論
Export
取書館
 
 
變更密碼
登入