A curated list of resources about multi-armed bandit (MAB).
-
Updated
Apr 7, 2020 - Jupyter Notebook
A curated list of resources about multi-armed bandit (MAB).
Combining Experience Replay with Exploration by Random Network Distillation
Add a description, image, and links to the exploitation-vs-exploration topic page so that developers can more easily learn about it.
To associate your repository with the exploitation-vs-exploration topic, visit your repo's landing page and select "manage topics."