-
Notifications
You must be signed in to change notification settings - Fork 864
Clarify CleanRL is a non-modular library #200
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 1 commit
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -13,19 +13,21 @@ | |
|
||
CleanRL is a Deep Reinforcement Learning library that provides high-quality single-file implementation with research-friendly features. The implementation is clean and simple, yet we can scale it to run thousands of experiments using AWS Batch. The highlight features of CleanRL are: | ||
|
||
|
||
* Single-file Implementation | ||
* **Every detail about an algorithm is put into the algorithm's own file.** Therefore, it's easier for you to fully understand an algorithm and do research with it. | ||
* Benchmarked Implementation on 7+ algorithms and 34+ games | ||
* Tensorboard Logging | ||
* Local Reproducibility via Seeding | ||
* Videos of Gameplay Capturing | ||
* Experiment Management with [Weights and Biases](https://wandb.ai/site) | ||
* Cloud Integration with Docker and AWS | ||
* 📜 Single-file implementation | ||
* *Every detail about an algorithm variant is put into a single standalone file.* | ||
* For example, our `ppo_atari.py` only has 340 lines of code but contains all implementation details on how PPO works with Atari games, so it is a great reference implementation to read for folks who do not wish to read an entire modular library. | ||
* 📊 Benchmarked Implementation (7+ algorithms and 34+ games at https://benchmark.cleanrl.dev) | ||
* 📈 Tensorboard Logging | ||
* 🪛 Local Reproducibility via Seeding | ||
* 🎮 Videos of Gameplay Capturing | ||
* 🧫 Experiment Management with [Weights and Biases](https://wandb.ai/site) | ||
* 💸 Cloud Integration with docker and AWS | ||
|
||
You can read more about CleanRL in our [technical paper](https://arxiv.org/abs/2111.08819) and [documentation](https://docs.cleanrl.dev/). | ||
|
||
Good luck have fun 🚀 | ||
Good luck have fun :rocket: | ||
|
||
⚠️ **NOTE**: CleanRL is *not* a modular library and therefore it is not meant to be imported. At the cost of duplicate code, we make all implementation details of a DRL algorithm variant easy to understand, so CleanRL comes with its own pros and cons. You should consider using CleanRL if you want to 1) understand all implementation details of an algorithm's varaint or 2) do quick prototypes. | ||
|
||
|
||
## Citing CleanRL | ||
|
||
|
Uh oh!
There was an error while loading. Please reload this page.