Skip to content

Latest commit

 

History

History

Cartpole

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 

CartPole RL Practices 🕹️

Python Status

Welcome to the CartPole RL Practices repository! This repository is divided into four directories, each focusing on a different Reinforcement Learning (RL) technique applied to the classic Cart Pole problem, a popular test environment in RL.

Table of Contents

1 - DQN (Deep Q-Networks)

  • 1 - DQN
  • Goals: Implement and evaluate the DQN algorithm. Focus on demonstrating how deep learning can be used to solve reinforcement learning problems efficiently.

DQN

Epoch 10
Epoch 10 Performance
Epoch 500
Epoch 500 Performance
Epoch 1000
Epoch 1000 Performance

2 - Hyperparameter Exploration

  • 2 - Hyperparameters
  • Goals: Analyze and understand the impact of different hyperparameters on the performance of RL algorithms. Using the DQN setup, explore variations in learning rates, discount factors, and update frequencies to optimize performance.

Learning Rate Variations

Learning Rate Loss Plot Reward Plot
1e-2
1e-4
1e-6

Discount Factor Variations

Discount Factor Loss Plot Reward Plot
0.997
0.97
0.9

Update Frequency Variations

Update Frequency Loss Plot Reward Plot
5
50
100

3 - Boltzmann Exploration

  • 3 - Boltzmann
  • Goals: Implement Boltzmann exploration strategy within a DQN setup to compare its effectiveness against the epsilon-greedy approach. Focus on the probabilistic approach to action selection based on Q-values.

boltzmann_formula

Set Temperature Plot Loss Plot Reward Plot
Set 1
Set 2
Set 3
Set 4

4 - SARSA (State-Action-Reward-State-Action)

  • 4 - SARSA
  • Goals: Apply the SARSA algorithm to the Cart Pole problem to evaluate its performance in a straightforward RL scenario. Focus on how SARSA's on-policy learning compares to other techniques like DQN.

SARSA

Each directory contains detailed implementations, experiments, and results that explore various aspects of RL algorithms. Feel free to explore each practice directory for in-depth code, insights, and performance analysis! 😃