Advanced Reinforcement Learning in Python: from DQN to SAC
Advanced Reinforcement Learning in Python: from DQN to SAC, available at $69.99, has an average rating of 4.5, with 121 lectures, based on 120 reviews, and has 1994 subscribers.
You will learn about Master some of the most advanced Reinforcement Learning algorithms. Learn how to create AIs that can act in a complex environment to achieve their goals. Create from scratch advanced Reinforcement Learning agents using Python's most popular tools (PyTorch Lightning, OpenAI gym, Brax, Optuna) Learn how to perform hyperparameter tuning (Choosing the best experimental conditions for our AI to learn) Fundamentally understand the learning process for each algorithm. Debug and extend the algorithms presented. Understand and implement new algorithms from research papers. This course is ideal for individuals who are Developers who want to get a job in Machine Learning. or Data scientists/analysts and ML practitioners seeking to expand their breadth of knowledge. or Robotics students and researchers. or Engineering students and researchers. It is particularly useful for Developers who want to get a job in Machine Learning. or Data scientists/analysts and ML practitioners seeking to expand their breadth of knowledge. or Robotics students and researchers. or Engineering students and researchers.
Enroll now: Advanced Reinforcement Learning in Python: from DQN to SAC
Summary
Title: Advanced Reinforcement Learning in Python: from DQN to SAC
Price: $69.99
Average Rating: 4.5
Number of Lectures: 121
Number of Published Lectures: 121
Number of Curriculum Items: 121
Number of Published Curriculum Objects: 121
Original Price: $199.99
Quality Status: approved
Status: Live
What You Will Learn
- Master some of the most advanced Reinforcement Learning algorithms.
- Learn how to create AIs that can act in a complex environment to achieve their goals.
- Create from scratch advanced Reinforcement Learning agents using Python's most popular tools (PyTorch Lightning, OpenAI gym, Brax, Optuna)
- Learn how to perform hyperparameter tuning (Choosing the best experimental conditions for our AI to learn)
- Fundamentally understand the learning process for each algorithm.
- Debug and extend the algorithms presented.
- Understand and implement new algorithms from research papers.
Who Should Attend
- Developers who want to get a job in Machine Learning.
- Data scientists/analysts and ML practitioners seeking to expand their breadth of knowledge.
- Robotics students and researchers.
- Engineering students and researchers.
Target Audiences
- Developers who want to get a job in Machine Learning.
- Data scientists/analysts and ML practitioners seeking to expand their breadth of knowledge.
- Robotics students and researchers.
- Engineering students and researchers.
This is the most complete Advanced Reinforcement Learning course on Udemy. In it, you will learn to implement some of the most powerful Deep Reinforcement Learning algorithms in Python using PyTorch and PyTorch lightning. You will implement from scratch adaptive algorithms that solve control tasks based on experience. You will learn to combine these techniques with Neural Networks and Deep Learning methods to create adaptive Artificial Intelligence agents capable of solving decision-making tasks.
This course will introduce you to the state of the art in Reinforcement Learning techniques. It will also prepare you for the next courses in this series, where we will explore other advanced methods that excel in other types of task.
The course is focused on developing practical skills. Therefore, after learning the most important concepts of each family of methods, we will implement one or more of their algorithms in jupyter notebooks, from scratch.
Leveling modules:
– Refresher: The Markov decision process (MDP).
– Refresher: Q-Learning.
– Refresher: Brief introduction to Neural Networks.
– Refresher: Deep Q-Learning.
– Refresher: Policy gradient methods
Advanced Reinforcement Learning:
– PyTorch Lightning.
– Hyperparameter tuning with Optuna.
– Deep Q-Learning for continuous action spaces (Normalized advantage function – NAF).
– Deep Deterministic Policy Gradient (DDPG).
– Twin Delayed DDPG (TD3).
– Soft Actor-Critic (SAC).
– Hindsight Experience Replay (HER).
Course Curriculum
Chapter 1: Introduction
Lecture 1: Introduction
Lecture 2: Reinforcement Learning series
Lecture 3: Google Colab
Lecture 4: Where to begin
Lecture 5: Complete code
Lecture 6: Connect with me on social media
Chapter 2: Refresher: The Markov Decision Process (MDP)
Lecture 1: Module Overview
Lecture 2: Elements common to all control tasks
Lecture 3: The Markov decision process (MDP)
Lecture 4: Types of Markov decision process
Lecture 5: Trajectory vs episode
Lecture 6: Reward vs Return
Lecture 7: Discount factor
Lecture 8: Policy
Lecture 9: State values v(s) and action values q(s,a)
Lecture 10: Bellman equations
Lecture 11: Solving a Markov decision process
Chapter 3: Refresher: Q-Learning
Lecture 1: Module overview
Lecture 2: Temporal difference methods
Lecture 3: Solving control tasks with temporal difference methods
Lecture 4: Q-Learning
Lecture 5: Advantages of temporal difference methods
Chapter 4: Refresher: Brief introduction to Neural Networks
Lecture 1: Module overview
Lecture 2: Function approximators
Lecture 3: Artificial Neural Networks
Lecture 4: Artificial Neurons
Lecture 5: How to represent a Neural Network
Lecture 6: Stochastic Gradient Descent
Lecture 7: Neural Network optimization
Chapter 5: Refresher: Deep Q-Learning
Lecture 1: Module overview
Lecture 2: Deep Q-Learning
Lecture 3: Experience Replay
Lecture 4: Target Network
Chapter 6: PyTorch Lightning
Lecture 1: PyTorch Lightning
Lecture 2: Link to the code notebook
Lecture 3: Introduction to PyTorch Lightning
Lecture 4: Create the Deep Q-Network
Lecture 5: Create the policy
Lecture 6: Create the replay buffer
Lecture 7: Create the environment
Lecture 8: Define the class for the Deep Q-Learning algorithm
Lecture 9: Define the play_episode() function
Lecture 10: Prepare the data loader and the optimizer
Lecture 11: Define the train_step() method
Lecture 12: Define the train_epoch_end() method
Lecture 13: [Important] Lecture correction.
Lecture 14: Train the Deep Q-Learning algorithm
Lecture 15: Explore the resulting agent
Chapter 7: Hyperparameter tuning with Optuna
Lecture 1: Hyperparameter tuning with Optuna
Lecture 2: Link to the code notebook
Lecture 3: Log average return
Lecture 4: Define the objective function
Lecture 5: Create and launch the hyperparameter tuning job
Lecture 6: Explore the best trial
Chapter 8: Deep Q-Learning for continuous action spaces (Normalized Advantage Function)
Lecture 1: Continuous action spaces
Lecture 2: The advantage function
Lecture 3: Normalized Advantage Function (NAF)
Lecture 4: Normalized Advantage Function pseudocode
Lecture 5: Link to the code notebook
Lecture 6: Hyperbolic tangent
Lecture 7: Creating the (NAF) Deep Q-Network 1
Lecture 8: Creating the (NAF) Deep Q-Network 2
Lecture 9: Creating the (NAF) Deep Q-Network 3
Lecture 10: Creating the (NAF) Deep Q-Network 4
Lecture 11: Creating the policy
Lecture 12: Create the environment
Lecture 13: Polyak averaging
Lecture 14: Implementing Polyak averaging
Lecture 15: Create the (NAF) Deep Q-Learning algorithm
Lecture 16: Implement the training step
Lecture 17: Implement the end-of-epoch logic
Lecture 18: Debugging and launching the algorithm
Lecture 19: Checking the resulting agent
Chapter 9: Refresher: Policy gradient methods
Lecture 1: Policy gradient methods
Lecture 2: Policy performance
Lecture 3: Representing policies using neural networks
Lecture 4: The policy gradient theorem
Lecture 5: Entropy Regularization
Chapter 10: Deep Deterministic Policy Gradient (DDPG)
Lecture 1: The Brax Physics engine
Lecture 2: Deep Deterministic Policy Gradient (DDPG)
Lecture 3: DDPG pseudocode
Lecture 4: Link to the code notebook
Lecture 5: Important – updated code
Lecture 6: Deep Deterministic Policy Gradient (DDPG)
Lecture 7: Create the gradient policy
Lecture 8: Create the gradient policy – Correction
Lecture 9: Create the Deep Q-Network
Lecture 10: Create the DDPG class
Lecture 11: Define the play method
Lecture 12: Define the play method – Correction
Instructors
-
Escape Velocity Labs
Hands-on, comprehensive AI courses
Rating Distribution
- 1 stars: 3 votes
- 2 stars: 1 votes
- 3 stars: 11 votes
- 4 stars: 30 votes
- 5 stars: 75 votes
Frequently Asked Questions
How long do I have access to the course materials?
You can view and review the lecture materials indefinitely, like an on-demand channel.
Can I take my courses with me wherever I go?
Definitely! If you have an internet connection, courses on Udemy are available on any device at any time. If you don’t have an internet connection, some instructors also let their students download course lectures. That’s up to the instructor though, so make sure you get on their good side!
You may also like
- Top 10 Video Editing Courses to Learn in November 2024
- Top 10 Music Production Courses to Learn in November 2024
- Top 10 Animation Courses to Learn in November 2024
- Top 10 Digital Illustration Courses to Learn in November 2024
- Top 10 Renewable Energy Courses to Learn in November 2024
- Top 10 Sustainable Living Courses to Learn in November 2024
- Top 10 Ethical AI Courses to Learn in November 2024
- Top 10 Cybersecurity Fundamentals Courses to Learn in November 2024
- Top 10 Smart Home Technology Courses to Learn in November 2024
- Top 10 Holistic Health Courses to Learn in November 2024
- Top 10 Nutrition And Diet Planning Courses to Learn in November 2024
- Top 10 Yoga Instruction Courses to Learn in November 2024
- Top 10 Stress Management Courses to Learn in November 2024
- Top 10 Mindfulness Meditation Courses to Learn in November 2024
- Top 10 Life Coaching Courses to Learn in November 2024
- Top 10 Career Development Courses to Learn in November 2024
- Top 10 Relationship Building Courses to Learn in November 2024
- Top 10 Parenting Skills Courses to Learn in November 2024
- Top 10 Home Improvement Courses to Learn in November 2024
- Top 10 Gardening Courses to Learn in November 2024