Skip to content

rpeddakama/Bandit-Learning-Research

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Bandit Learning Research

Multi-armed bandit algorithms; Research done at Dr Ji's SNAIL lab @ Virginia Tech

Implements different bandit algorithms that solve the exploration vs exploitation problem.

Algorithms

  • Epsilon-greedy: Random exploration with probability ε
  • UCB: Upper confidence bound approach
  • Thompson Sampling: Bayesian method
  • Contextual bandits: Uses additional info to make decisions

Requirements

numpy, matplotlib, scipy

About

🤖 Research on novel multi-armed bandit strategies

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published