Multi-armed bandit experiments in the online service economy
From MaRDI portal
Publication:6574679
DOI10.1002/ASMB.2104MaRDI QIDQ6574679
Publication date: 18 July 2024
Published in: Applied Stochastic Models in Business and Industry (Search for Journal in Brave)
Cites Work
- Title not available (Why is that?)
- Title not available (Why is that?)
- Title not available (Why is that?)
- Title not available (Why is that?)
- Statistical methods for dynamic treatment regimes. Reinforcement learning, causal inference, and personalized medicine
- Asymptotically efficient adaptive allocation rules
- Bandit problems with infinitely many arms
- On the likelihood that one unkrown probability exeeds another in view of the evidence of two samples.
- Thompson Sampling: An Asymptotically Optimal Finite-Time Analysis
- Learning to Optimize via Posterior Sampling
- Some aspects of the sequential design of experiments
- Finite-time analysis of the multiarmed bandit problem
Related Items (5)
Visualizations for interrogations of multi-armed bandits ⋮ A Bayesian two-armed bandit model ⋮ Discussion of: ``Machine learning applications in nonlife insurance ⋮ Novelty and Primacy: A Long-Term Estimator for Online Experiments ⋮ Factorial Designs for Online Experiments
This page was built for publication: Multi-armed bandit experiments in the online service economy
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6574679)