Bayesian experimental design (BED) is a methodology to identify designs that are expected to yield informative data. Recent work in cognitive science considered BED for cognitive models with tractable and known likelihood functions. However, as cognitive models have become more complex and richer, their likelihood functions are often intractable. In this work, we leverage recent advances in BED for intractable models and demonstrate their application on a set of multi-armed bandit tasks. We further propose a generalized latent state model that unifies two previously proposed models. Our experiments show that data gathered using optimal designs results in improved model discrimination and parameter estimation, as compared to naive designs. Furthermore, we find that increasing the number of bandit arms increases the expected information gain in experiments.