- Deep Reinforcement Learning Hands-On
- Maxim Lapan
- 179字
- 2021-06-25 20:46:58
Chapter 6. Deep Q-Networks
In the previous chapter, we became familiar with the Bellman equation and the practical method of its application called Value iteration. This approach allowed us to significantly improve our speed and convergence in the FrozenLake environment, which is promising, but can we go further?
In this chapter, we'll try to apply the same theory to problems of much greater complexity: arcade games from the Atari 2600 platform, which are the de-facto benchmark of the RL research community. To deal with this new and more challenging goal, we'll talk about problems with the Value iteration method and introduce its variation, called Q-learning. In particular, we'll look at the application of Q-learning to so-called "grid world" environments, which is called tabular Q-learning, and then we'll discuss Q-learning in conjunction with neural networks. This combination has the name DQN. At the end of the chapter, we'll reimplement a DQN algorithm from the famous paper, Playing Atari with Deep Reinforcement Learning by V. Mnih and others, published in 2013, which started a new era in RL development.
- 工業機器人虛擬仿真實例教程:KUKA.Sim Pro(全彩版)
- Hadoop 2.x Administration Cookbook
- Hands-On Cloud Solutions with Azure
- Windows XP中文版應用基礎
- 自動檢測與傳感技術
- ServiceNow Cookbook
- VB語言程序設計
- 單片機C語言應用100例
- 筆記本電腦電路分析與故障診斷
- 人工智能:智能人機交互
- Machine Learning with Spark(Second Edition)
- PostgreSQL 10 High Performance
- 基于Quartus Ⅱ的數字系統Verilog HDL設計實例詳解
- 軟件質量管理實踐
- 系統安裝、維護與數據備份技巧