Skip to content

Theoretical review and a implementation of the Soft Actor-Critic algorithm (Haarnoja et al., 2018). It includes a derivation of soft policy iteration, a tabular implementation on Taxi-v3, and experimental comparisons between SAC, PPO, and DDPG using Stable-Baselines3.

Notifications You must be signed in to change notification settings

alexandre-zenou/Reinforcement-Learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Soft Actor-Critic (SAC)

This repository contains the final project completed in the context of the Reinforcement Learning course at École Polytechnique

The project is based on the paper:

Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, Sergey Levine (2018)

It includes a theoretical study of maximum-entropy reinforcement learning, an implementation of soft policy iteration in a discrete setting, and empirical experiments with Soft Actor-Critic (SAC) in continuous control environments.

The work was concluded by an oral presentation delivered in front of a jury as part of the course evaluation.


Repository Structure

  • Code/ — Personal implementations and experiments
  • Report/ — Slides
  • Research Paper/ — Original research article (Haarnoja et al., 2018)

Authors

Lucien Le Gall
Alexandre Zenou
Remi Moshfeghi

École polytechnique

About

Theoretical review and a implementation of the Soft Actor-Critic algorithm (Haarnoja et al., 2018). It includes a derivation of soft policy iteration, a tabular implementation on Taxi-v3, and experimental comparisons between SAC, PPO, and DDPG using Stable-Baselines3.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published