Thompson sampling for zero-inflated count outcomes with an application to the Drink Less mobile health study.
CoRR(2023)
摘要
Mobile health (mHealth) technologies aim to improve distal outcomes, such as
clinical conditions, by optimizing proximal outcomes through just-in-time
adaptive interventions. Contextual bandits provide a suitable framework for
customizing such interventions according to individual time-varying contexts,
intending to maximize cumulative proximal outcomes. However, unique challenges
such as modeling count outcomes within bandit frameworks have hindered the
widespread application of contextual bandits to mHealth studies. The current
work addresses this challenge by leveraging count data models into online
decision-making approaches. Specifically, we combine four common offline count
data models (Poisson, negative binomial, zero-inflated Poisson, and
zero-inflated negative binomial regressions) with Thompson sampling, a popular
contextual bandit algorithm. The proposed algorithms are motivated by and
evaluated on a real dataset from the Drink Less trial, where they are shown to
improve user engagement with the mHealth system. The proposed methods are
further evaluated on simulated data, achieving improvement in maximizing
cumulative proximal outcomes over existing algorithms. Theoretical results on
regret bounds are also derived. A user-friendly R package countts that
implements the proposed methods for assessing contextual bandit algorithms is
made publicly available at https://cran.r-project.org/web/packages/countts.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要