Amazon cover image
Image from Amazon.com

TensorFlow Reinforcement Learning Quick Start Guide : Get up and Running with Training and Deploying Intelligent, Self-Learning Agents Using Python.

By: Material type: TextTextPublication details: Birmingham : Packt Publishing Ltd, 2019.Description: 1 online resource (175 pages)Content type:
  • text
Media type:
  • computer
Carrier type:
  • online resource
ISBN:
  • 1789533449
  • 9781789533446
Subject(s): Genre/Form: Additional physical formats: Print version:: TensorFlow Reinforcement Learning Quick Start Guide : Get up and Running with Training and Deploying Intelligent, Self-Learning Agents Using Python.DDC classification:
  • 005.133 23
LOC classification:
  • QA76.73.P98
Online resources:
Contents:
Cover; Title Page; Copyright and Credits; Dedication; About Packt; Contributors; Table of Contents; Preface; Chapter 1: Up and Running with Reinforcement Learning; Why RL?; Formulating the RL problem; The relationship between an agent and its environment; Defining the states of the agent; Defining the actions of the agent; Understanding policy, value, and advantage functions; Identifying episodes; Identifying reward functions and the concept of discounted rewards; Rewards; Learning the Markov decision process ; Defining the Bellman equation; On-policy versus off-policy learning
On-policy methodOff-policy method; Model-free and model-based training; Algorithms covered in this book; Summary; Questions; Further reading; Chapter 2: Temporal Difference, SARSA, and Q-Learning; Technical requirements; Understanding TD learning; Relation between the value functions and state; Understanding SARSA and Q-Learning ; Learning SARSA ; Understanding Q-learning; Cliff walking and grid world problems; Cliff walking with SARSA; Cliff walking with Q-learning; Grid world with SARSA; Summary; Further reading; Chapter 3: Deep Q-Network; Technical requirements
Learning the theory behind a DQNUnderstanding target networks; Learning about replay buffer; Getting introduced to the Atari environment; Summary of Atari games; Pong; Breakout; Space Invaders; LunarLander; The Arcade Learning Environment ; Coding a DQN in TensorFlow; Using the model.py file; Using the funcs.py file; Using the dqn.py file; Evaluating the performance of the DQN on Atari Breakout; Summary; Questions; Further reading; Chapter 4: Double DQN, Dueling Architectures, and Rainbow; Technical requirements; Understanding Double DQN ; Coding DDQN and training to play Atari Breakout
Evaluating the performance of DDQN on Atari BreakoutUnderstanding dueling network architectures; Coding dueling network architecture and training it to play Atari Breakout; Combining V and A to obtain Q; Evaluating the performance of dueling architectures on Atari Breakout ; Understanding Rainbow networks; DQN improvements; Prioritized experience replay ; Multi-step learning; Distributional RL; Noisy nets; Running a Rainbow network on Dopamine; Rainbow using Dopamine; Summary; Questions; Further reading; Chapter 5: Deep Deterministic Policy Gradient; Technical requirements
Actor-Critic algorithms and policy gradientsPolicy gradient; Deep Deterministic Policy Gradient; Coding ddpg.py; Coding AandC.py; Coding TrainOrTest.py; Coding replay_buffer.py; Training and testing the DDPG on Pendulum-v0; Summary; Questions; Further reading; Chapter 6: Asynchronous Methods -- A3C and A2C; Technical requirements; The A3C algorithm; Loss functions; CartPole and LunarLander; CartPole; LunarLander; The A3C algorithm applied to CartPole; Coding cartpole.py; Coding a3c.py; The AC class; The Worker() class; Coding utils.py; Training on CartPole
Summary: This book is an essential guide for anyone interested in Reinforcement Learning. The book provides an actionable reference for Reinforcement Learning algorithms and their applications using TensorFlow and Python. It will help readers leverage the power of algorithms such as Deep Q-Network (DQN), Deep Deterministic Policy Gradients (DDPG), and ...
Item type:
Tags from this library: No tags from this library for this title. Log in to add tags.
Star ratings
    Average rating: 0.0 (0 votes)
Holdings
Item type Home library Collection Call number Materials specified Status Date due Barcode
Electronic-Books Electronic-Books OPJGU Sonepat- Campus E-Books EBSCO Available

Cover; Title Page; Copyright and Credits; Dedication; About Packt; Contributors; Table of Contents; Preface; Chapter 1: Up and Running with Reinforcement Learning; Why RL?; Formulating the RL problem; The relationship between an agent and its environment; Defining the states of the agent; Defining the actions of the agent; Understanding policy, value, and advantage functions; Identifying episodes; Identifying reward functions and the concept of discounted rewards; Rewards; Learning the Markov decision process ; Defining the Bellman equation; On-policy versus off-policy learning

On-policy methodOff-policy method; Model-free and model-based training; Algorithms covered in this book; Summary; Questions; Further reading; Chapter 2: Temporal Difference, SARSA, and Q-Learning; Technical requirements; Understanding TD learning; Relation between the value functions and state; Understanding SARSA and Q-Learning ; Learning SARSA ; Understanding Q-learning; Cliff walking and grid world problems; Cliff walking with SARSA; Cliff walking with Q-learning; Grid world with SARSA; Summary; Further reading; Chapter 3: Deep Q-Network; Technical requirements

Learning the theory behind a DQNUnderstanding target networks; Learning about replay buffer; Getting introduced to the Atari environment; Summary of Atari games; Pong; Breakout; Space Invaders; LunarLander; The Arcade Learning Environment ; Coding a DQN in TensorFlow; Using the model.py file; Using the funcs.py file; Using the dqn.py file; Evaluating the performance of the DQN on Atari Breakout; Summary; Questions; Further reading; Chapter 4: Double DQN, Dueling Architectures, and Rainbow; Technical requirements; Understanding Double DQN ; Coding DDQN and training to play Atari Breakout

Evaluating the performance of DDQN on Atari BreakoutUnderstanding dueling network architectures; Coding dueling network architecture and training it to play Atari Breakout; Combining V and A to obtain Q; Evaluating the performance of dueling architectures on Atari Breakout ; Understanding Rainbow networks; DQN improvements; Prioritized experience replay ; Multi-step learning; Distributional RL; Noisy nets; Running a Rainbow network on Dopamine; Rainbow using Dopamine; Summary; Questions; Further reading; Chapter 5: Deep Deterministic Policy Gradient; Technical requirements

Actor-Critic algorithms and policy gradientsPolicy gradient; Deep Deterministic Policy Gradient; Coding ddpg.py; Coding AandC.py; Coding TrainOrTest.py; Coding replay_buffer.py; Training and testing the DDPG on Pendulum-v0; Summary; Questions; Further reading; Chapter 6: Asynchronous Methods -- A3C and A2C; Technical requirements; The A3C algorithm; Loss functions; CartPole and LunarLander; CartPole; LunarLander; The A3C algorithm applied to CartPole; Coding cartpole.py; Coding a3c.py; The AC class; The Worker() class; Coding utils.py; Training on CartPole

The A3C algorithm applied to LunarLander

This book is an essential guide for anyone interested in Reinforcement Learning. The book provides an actionable reference for Reinforcement Learning algorithms and their applications using TensorFlow and Python. It will help readers leverage the power of algorithms such as Deep Q-Network (DQN), Deep Deterministic Policy Gradients (DDPG), and ...

Print version record.

Includes bibliographical references.

eBooks on EBSCOhost EBSCO eBook Subscription Academic Collection - Worldwide

There are no comments on this title.

to post a comment.

O.P. Jindal Global University, Sonepat-Narela Road, Sonepat, Haryana (India) - 131001

Send your feedback to glus@jgu.edu.in

Hosted, Implemented & Customized by: BestBookBuddies   |   Maintained by: Global Library