官术网_书友最值得收藏!

Algorithms covered in this book

In Chapter 2, Temporal Difference, SARSA, and Q-Learning, we will look into our first two RL algorithms: Q-learning and SARSA. Both of these algorithms are tabular-based and do not require the use of neural networks. Thus, we will code them in Python and NumPy. In Chapter 3, Deep Q-Network, we will cover DQN and use TensorFlow to code the agent for the rest of the book. We will then train it to play Atari Breakout. In Chapter 4, Double DQN, Dueling Architectures, and Rainbow, we will cover double DQN, dueling network architectures, and rainbow DQN. In Chapter 5, Deep Deterministic Policy Gradient, we will look at our first Actor-Critic RL algorithm called DDPG, learn about policy gradients, and apply them to a continuous action problem. In Chapter 6, Asynchronous Methods – A3C and A2C, we will investigate A3C, which is another RL algorithm that uses a master and several worker processes. In Chapter 7, Trust Region Policy Optimization and Proximal Policy Optimization, we will investigate two more RL algorithms: TRPO and PPO. Finally, we will apply DDPG and PPO to train an agent to drive a car autonomously in Chapter 8, Deep RL Applied to Autonomous Driving. From Chapter 3, Deep Q-Network, to Chapter 8, Deep RL Applied to Autonomous Driving, we'll use TensorFlow agents. Have fun learning RL.

主站蜘蛛池模板: 敦煌市| 宜良县| 昌邑市| 定兴县| 双辽市| 曲阳县| 葵青区| 子长县| 太康县| 兰坪| 赤壁市| 濉溪县| 平谷区| 来宾市| 嘉义县| 永修县| 乌苏市| 甘南县| 大石桥市| 靖远县| 洱源县| 宝坻区| 磴口县| 乐清市| 温州市| 伊宁市| 邯郸县| 江源县| 馆陶县| 方正县| 廉江市| 盐城市| 华阴市| 天全县| 竹溪县| 丹东市| 嘉黎县| 晋州市| 襄城县| 抚顺市| 霍山县|