扫一扫
关注中图网
官方微博
本类五星书更多>
-
>
湖南省志(1978-2002)?铁路志
-
>
公路车宝典(ZINN的公路车维修与保养秘籍)
-
>
晶体管电路设计(下)
-
>
基于个性化设计策略的智能交通系统关键技术
-
>
德国克虏伯与晚清火:贸易与仿制模式下的技术转移
-
>
花样百出:贵州少数民族图案填色
-
>
识木:全球220种木材图鉴
Self-Learning Optimal Control of Nonlinear Systems 版权信息
- ISBN:9787030520609
- 条形码:9787030520609 ; 978-7-03-052060-9
- 装帧:一般胶版纸
- 册数:暂无
- 重量:暂无
- 所属分类:>
Self-Learning Optimal Control of Nonlinear Systems 目录
1 Principle of Adaptive Dynamic Programming 1.1 Dynamic Programming 1.1.1 Discretε-Time Systems 1.1.2 Continuous-Time Systems 1.2 Original Forms of Adaptive Dynamic Programming 1.2.1 Principle of Adaptive Dynamic Programming 1.3 Iterative Forms of Adaptive Dynamic Programming 1.3.1 Value Iteration 1.3.2 Policy Iteration 1.4 About This Book References2 An Iterative ε-Optimal Control Scheme for a Class of Discretε-Time Nonlinear Systems with Unfixed Initial State.
2.1 Introduction 2.2 Problem Statement 2.3 Properties of the Iterative Adaptive Dynamic Programming Algorithm 2.3.1 Derivation of the Iterative ADP Algorithm 2.3.2 Properties of the Iterative ADP Algorithm 2.4 The ε-Optimal Control Algorithm 2.4.1 The Derivation of the ε-Optimal Control Algorithm 2.4.2 Properties of the ε-Optimal Control Algorithm 2.4.3 The ε-Optimal Control Algorithm for Unfixed Initial State 2.4.4 The Expressions of the ε-Optimal Control Algorithm 2.5 Neural Network Implementation for the ε-Optimal Control Scheme 2.5.1 The Critic Network 2.5.2 The Action Network 2.6 Simulation Study 2.7 Conclusions References3 Discretε-Time Optimal Control of Nonlinear Systems via Value Iteration-Based Q-Learning 3.1 Introduction 3.2 Preliminaries and Assumptions 3.2.1 Problem Formulations 3.2.2 Derivation of the Discretε-Time Q-Learning Algorithm 3.3 Properties of the Discretε-Time Q-Learning Algorithm 3.3.1 Non-Discount Case 3.3.2 Discount Case 3.4 Neural Network Implementation for the Discretε-Time Q-Learning Algorithm 3.4.1 The Action Network 3.4.2 The Critic Network 3.4.3 Training Phase 3.5 Simulation Study 3.5.1 Example 1 3.5.2 Example 2 3.6 Conclusion References4 A Novel Policy Iteration-Based Deterministic Q-Learning for Discretε-Time Nonlinear Systems 4.1 Introduction 4.2 Problem Formulation 4.3 Policy Iteration-Based Deterministic Q-Learning Algorithm for Discretε-Time Nonlinear Systems 4.3.1 Derivation of the Policy Iteration-Based Deterministic Q-Learning Algorithm 4.3.2 Properties of the Policy Iteration-Based Deterministic Q-Learning Algorithm 4.4 Neural Network Implementation for the Policy Iteration-Based Deterministic Q-Learning Algorithm 4.4.1 The Critic Network 4.4.2 The Action Network 4.4.3 Summary of the Policy Iteration-Based Deterministic Q-Learning Algorithm 4.5 Simulation Study 4.5.1 Example 1 4.5.2 Example 2 4.6 Conclusion References5 Nonlinear Neuro-Optimal Tracking Control via Stable Iterative Q-Learning Algorithm 5.1 Introduction 5.2 Problem Statement 5.3 Policy Iteration Q-Learning Algorithm for Optimal Tracking Control 5.4 Properties of the Policy Iteration Q-Learning Algorithm 5.5 Neural Network Implementation for the Policy Iteration Q-Learning Algorithm 5.5.1 The Critic Network 5.5.2 The Action Network 5.6 Simulation Study 5.6.1 Example 1 5.6.2 Example 2 5.7 Conclusions References6 Model-Free Multiobjective Adaptive Dynamic Programming for Discretε-Time Nonlinear Systems with General Performance Index Functions 6.1 Introduction 6.2 Preliminaries 6.3 Multiobjective Adaptive Dynamic Programming Method 6.4 Model-Free Incremental Q-Learning Method 6.5 Neural Network Implementation for the Incremental Q-Learning Method 6.5.1 The Critic Network 6.5.2 The Action Network 6.5.3 The Procedure of the Model-Free Incremental Q-learning Method 6.6 Convergence Proof 6.7 Simulation Study 6.7.1 Example 1 6.7.2 Example 2 6.8 Conclusion References7 Multiobjective Optimal Control for a Class of Unknown Nonlinear Systems Based on Finitε-Approximation-Error ADP Algorithm 7.1 Introduction 7.2 General Formulation 7.3 Optimal Solution Based on Finitε-Approximation-Error ADP 7.3.1 Data-Based Identifier of Unknown System Dynamics 7.3.2 Derivation of the ADP Algorithm with Finite Approximation Errors 7.3.3 Convergence Analysis of the Iterative ADP Algorithm 7.4 Implementation of the Iterative ADP Algorithm 7.4.1 Critic Network 7.4.2 The Action Network 7.4.3 The Procedure of the ADP Algorithm 7.5 Simulation Study 7.5.1 Example 1 7.5.2 Example 2 7.6 Conclusions References8 A New Approach for a Class of Continuous-Time Chaotic Systems Optimal Control by Online ADP Algorithm 8.1 Introduction 8.2 Problem Statement 8.3 Optimal Control Based on Online ADP Algorithm 8.3.1 Design Method of the Critic Network and the Action Network 8.3.2 Stability Analysis 8.3.3 Online ADP Algorithm Implementation 8.4 Simulation Examples 8.4.1 Example 1 8.4.2 Example 2 8.5 Conclusions References9 Off-Pohcy IRL Optimal Tracking Control for Continuous-Time Chaotic Systems 9.1 Introduction 9.2 System Description and Problem Statement 9.3 Off-Policy IRL ADP Algorithm 9.3.1 Convergence Analysis of IRL ADP Algorithm 9.3.2 Off-Policy IRL Method 9.3.3 Methods for Updating Weights 9.4 Simulation Study 9.4.1 Example 1 9.4.2 Example 2 9.5 Conclusion References10 ADP-Based Optimal Sensor Scheduling for Target Tracking in Energy Harvesting Wireless Sensor Networks 10.1 Introduction 10.2 Problem Formulation 10.2.1 NN Model Description of Solar Energy Harvesting.
10.2.2 Sensor Energy Consumption 10.2.3 KF Technology 10.3 ADP-Based Sensor Scheduling for Maximum WSNs Residual Energy and Minimum Measuring Accuracy 10.3.1 Optimization Problem of the Sensor Scheduling 10.3.2 ADP-Based Sensor Scheduling with Convergence Analysis 10.3.3 Critic Network 10.3.4 Implementation Process 10.4 Simulation Study 10.5 Conclusion ReferencesIndex
展开全部
书友推荐
本类畅销
-
黑科技驱动世界的100项技术
¥21.9¥69.8 -
黑科技:驱动世界的100项技术(八品)
¥21.9¥69.8 -
指火控弹道与射击诸元理论
¥38.7¥49 -
智能照明工程手册
¥29.9¥49.8 -
光电直读光谱分析技术与应用
¥38.7¥49 -
服务型制造
¥76.3¥109