Reinforcement Learning (English): Master the Art of RL

Destiny For Everything


Reinforcement Studying

What you’ll study

Outline what’s Reinforcement Studying?

Apply all what’s discovered utilizing state-of-the artwork libraries like OpenAI Gymnasium, StabeBaselines, Keras-RL and TensorFlow Brokers

Outline what are the functions domains and success tales of RL?

Outline what are the distinction between Reinforcement and Supervised Studying?

Outline the primary parts of an RL drawback setup?

Outline what are the primary components of an RL agent and their taxonomy?

Outline what’s Markov Reward Course of (MRP) and Markov Choice Course of (MDP)?

Outline the answer area of RL utilizing MDP framework

Remedy the RL issues utilizing planning with Dynamic Programming algorithms, like Coverage Analysis, Coverage Iteration and Worth Iteration

Remedy RL issues utilizing mannequin free algorithms like Monte-Carlo, TD studying, Q-learning and SARSA

Differentiate On-policy and Off-policy algorithms

Grasp Deep Reinforcement Studying algorithms like Deep Q-Networks (DQN), and apply them to Massive Scale RL

Grasp Coverage Gradients algorithms and Actor-Critic (AC, A2C, A3C)

Grasp superior DRL algorithms like DDPG, TRPO and PPO

Outline what’s model-based RL, and differentiate it from planning, and what are their predominant algorithms and functions?

Description

Howdy and welcome to our course; Reinforcement Studying.

Reinforcement Studying is a really thrilling and essential subject of Machine Studying and AI. Some name it the crown jewel of AI.

On this course, we are going to cowl all of the features associated to Reinforcement Studying or RL. We’ll begin by defining the RL drawback, and evaluate it to the Supervised Studying drawback, and uncover the areas of functions the place RL can excel. This contains the issue formulation, ranging from the very fundamentals to the superior utilization of Deep Studying, resulting in the period of Deep Reinforcement Studying.

In our journey, we are going to cowl, as common, each the theoretical and sensible features, the place we are going to discover ways to implement the RL algorithms and apply them to the well-known issues utilizing libraries like OpenAI Gymnasium, Keras-RL, TensorFlow Brokers or TF-Brokers and Secure Baselines.

The course is split into 6 predominant sections:

1- We begin with an introduction to the RL drawback definition, primarily evaluating it to the Supervised studying drawback, and discovering the appliance domains and the primary constituents of an RL drawback. We describe right here the well-known OpenAI Gymnasium environments, which might be our playground in relation to sensible implementation of the algorithms that we study.

2- Within the second half we talk about the primary formulation of an RL drawback as a Markov Choice Course of or MDP, with easy answer to essentially the most fundamental issues utilizing Dynamic Programming.

3- After being armed with an understanding of MDP, we transfer on to discover the answer area of the MDP drawback, and what the completely different options past DP, which incorporates model-based and model-free options. We’ll focus on this half on model-free options, and defer model-based options to the final half. On this half, we describe the Monte-Carlo and Temporal-Distinction sampling based mostly strategies, together with the well-known and essential Q-learning algorithm, and SARSA. We’ll describe the sensible utilization and implementation of Q-learning and SARSA on management tabular maze issues from OpenAI Gymnasium environments.

4- To maneuver past easy tabular issues, we might want to study perform approximation in RL, which ends up in the mainstream RL strategies at present utilizing Deep Studying, or Deep Reinforcement Studying (DRL). We’ll describe right here the breakthrough algorithm of DeepMind that solved the Atari video games and AlphaGO, which is Deep Q-Networks or DQN. We additionally talk about how we are able to resolve Atari video games issues utilizing DQN in observe utilizing Keras-RL and TF-Brokers.

5- Within the fifth half, we transfer to Superior DRL algorithms, primarily beneath a household referred to as Coverage based mostly strategies. We talk about right here Coverage Gradients, DDPG, Actor-Critic, A2C, A3C, TRPO and PPO strategies. We additionally talk about the essential Secure Baseline library to implement all these algorithms on completely different environments in OpenAI Gymnasium, like Atari and others.

6- Lastly, we discover the model-based household of RL strategies, and importantly, differentiating model-based RL from planning, and exploring the entire spectrum of RL strategies.

Hopefully, you get pleasure from this course, and discover it helpful.

English
language

Content material

Introduction

Course introduction
Course overview

Introduction to Reinforcement Studying

Module intro and roadmap
What’s RL?
What RL can do?
The RL drawback setup (AREA)
Reward
RL vs. Supervised Studying
State
AREA examples and quizes
Gymnasium Environments
Inside RL agent – RL agent components
Coverage
Worth
Mannequin
RL brokers taxonomy
Prediction vs Management

Markov Choice Course of (MDP)

Module intro and roadmap
Markov Chain and Markov Course of (MP)
Markov Reward Course of (MRP)
Markov Choice Course of (MDP)
Prediction
Bellman Equations with action-value perform Q
Management

MDP options areas

Module intro and roadmap
Planning with Dynamic Programming (DP)
Prediction with DP – Coverage Analysis
Management with DP – Coverage Iteration and Worth Iteration
Worth Iteration instance
Prediction with Monte-Carlo – MC Coverage Analysis
Prediction with Temporal-Distinction (TD)
TD Lambda
Management with Monte-Carlo – MC Coverage Iteration
Management with TD – SARSA
On-policy vs. Off-policy
Q-learning
MDP options abstract

Deep Reinforcement Studying (DRL)

Module intro and roadmap
Massive Scale Reinforcement Studying
DNN as perform approximator
Worth Perform Approximation
DNN insurance policies
Worth perform approximation with DL encoder-decoder sample
Deep Q-Networks (DQN)
DQN Atari Instance with Keras-RL and TF-Brokers

Superior DRL

Module intro and roadmap
Worth-based vs Coverage based mostly vs Actor-Critic
Coverage Gradients (PG)
REINFORCE – Monte-Carlo PG
AC – Actor-Critic
A2C – Benefit Actor-Critic
A3C – Asynchronous Benefit Actor-Critic
TRPO – Trusted Area Coverage Optimization
PPO – Proximal Coverage Optimization
DDPG – Deep Determinstic Coverage Gradients
StableBaselines library overview
Atari instance with stable-baselines
Mario instance with stable-baselines
StreetFighter instance with stable-baselines

Mannequin-based Reinforcement Studying

Module intro and roadmap
Mannequin studying strategies
Mannequin studying with Supervised Studying and Perform Approximation
Pattern based mostly planning
Dyna – Intergation planning and Studying

Conclusion

Conclusion

Materials

Slides

The post Reinforcement Studying (English): Grasp the Artwork of RL appeared first on destinforeverything.com.

Please Wait 10 Sec After Clicking the "Enroll For Free" button.