Research Article

Combining Multiple Strategies for Multiarmed Bandit Problems and Asymptotic Optimality

Figure 2

Percentage of plays of the optimal arm of (tuned) -comb(3 -greedy strategies) and of (tuned) Exp4/NEXP/EEEs for the distribution ().