SkillAgentSearch skills...

FloRL

Implicit Normalizing Flows + Reinforcement Learning

Install / Use

/learn @joeybose/FloRL
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

Improving Exploration in SAC with Normalizing Flows Policies

This codebase was used to generate the results documented in the paper "Improving Exploration in Soft-Actor-Critic with Normalizing Flows Policies". Patrick Nadeem Ward<sup>*12</sup>, Ariella Smofsky<sup>*12</sup>, Avishek Joey Bose<sup>12</sup>. INNF Workshop ICML 2019.

  • <sup>*</sup> Equal contribution, <sup>1</sup> McGill University, <sup>2</sup> Mila
  • Correspondence to:

Requirements

Run Experiments

Gaussian policy on Dense Gridworld environment with REINFORCE:

TODO

Gaussian policy on Sparse Gridworld environment with REINFORCE:

TODO

Gaussian policy on Dense Gridworld environment with reparametrization:

python main.py --namestr=G-S-DG-CG --make_cont_grid --batch_size=128 --replay_size=100000 --hidden_size=64 --num_steps=100000 --policy=Gaussian --smol --comet --dense_goals --silent

Gaussian policy on Sparse Gridworld environment with reparametrization:

python main.py --namestr=G-S-CG --make_cont_grid --batch_size=128 --replay_size=100000 --hidden_size=64 --num_steps=100000 --policy=Gaussian --smol --comet --silent

Normalizing Flow policy on Dense Gridworld environment:

TODO

Normalizing Flow policy on Sparse Gridworld environment:

TODO

To run an experiment with a different policy distribution, modify the --policy flag.

References

Related Skills

View on GitHub
GitHub Stars62
CategoryEducation
Updated2mo ago
Forks7

Languages

Python

Security Score

95/100

Audited on Jan 15, 2026

No findings