Multi-Objective Multi-Armed Bandit
-
Updated
Jul 17, 2023 - Python
Multi-Objective Multi-Armed Bandit
This repository contains several implementations of multi-armed bandit (MAB) agents applied to a simulated cricket match where an agent selects among different strategies with the goal of maximizing runs while minimizing the risk of getting out.
An implementation of Bandit Algorithms, focusing on the case of Bernoulli Rewards.
Add a description, image, and links to the ucb-algorithm topic page so that developers can more easily learn about it.
To associate your repository with the ucb-algorithm topic, visit your repo's landing page and select "manage topics."