Research Article

Combining Multiple Strategies for Multiarmed Bandit Problems and Asymptotic Optimality

Figure 1

Percentage of selections of the best strategy of (tuned) -comb(3 -greedy strategies) and of (tuned) Exp4/NEXP/EEEs for the distribution ().