Online Learning Algorithms in Varied Bandit Scenarios