Gianluca Baio
Department of Statistical Science | University College London
https://gianluca.statistica.it
https://egon.stats.ucl.ac.uk/research/statistics-health-economics
https://github.com/giabaio https://github.com/StatisticsHealthEconomics
@gianlubaio@mas.to @gianlubaio
Issue Panel
ISPOR Europe 2025, Glasgow
11 November 2025
Check out our departmental podcast “Random Talks” on Soundcloud!
Follow our departmental social media accounts + magazine “Sample Space”
Expanding the focus on multiple imputation and linking it up with Bayesian modelling
Is it more efficient, particularly in HTA?\(^{*}\)
Methodological approaches to handle situations where substantial amount of missingness is present, with particular relevance to PRO data accounting for the longitudinal nature of the data
An example…
\(^{*}\)Yes it is. 😉
We plan to observe \(n_{\rm{planned}}\) data points, but end up with a (much) lower number of observations \(n_{\rm{observed}}\)
We typically don’t know why the unobserved points are missing and what their value might have been
– Missingness can be differential in treatment/exposure groups
Multiple (stochastic) imputation (MI)
“Full Bayesian”
\[\class{myblue}{y_i = \beta_0 + \beta_1 x_i + \varepsilon_i \qquad \varepsilon_i \sim \dnorm(0,\sigma^2)}\]
\[\class{myblue}{y_i^{\rm{mis}}\sim \dnorm(\hat{\beta}_0 + \hat{\beta}_1 x_i,\hat\sigma^2)}\]
\[\class{myblue}{\hat{\mu}_{\txt{MI}} = \frac{1}{R}\sum_{r=1}^R \bar{y}_r}\]
with variance
\[\class{myblue}{\hat{\sigma}^2_{\text{MI}} = \underbrace{\left(1 + \frac{1}{R}\right)}_{\substack{\txt{finite sampling}\\ \txt{correction}}}\underbrace{\left[\frac{1}{R-1}\sum_{r=1}^R \left(\bar{y}_r - \hat{\mu}_{\txt{MI}}\right)^2\right]}_{\txt{between imputation}} + \underbrace{\left[ \frac{1}{R}\sum_{r=1}^R s^2_r \right]}_{\txt{within imputation}}}\]
(Generally) valid under MCAR and MAR assumptions
Makes use of the whole dataset
Can be extended to MNAR, although models become more complex and untestable assumptions are necessary
“Vanilla” implementations based on underlying Normality assumption
MI can be seen as an approximation to the full Bayesian modelling in which the missingness mechanism is formally modelled (if necessary, eg MNAR)
In any case, as the underlying model complexity increases, the Bayesian machinery becomes marginally less complicated…
The MCMC structure does not change dramatically, once it is in place and can handle the idiosyncrasy of the data
JAGS or Stan…)\[\class{myblue}{p(y^{\text{mis}}\mid y^{\text{obs}},\bm{x}) = \int p(y^{\text{mis}}\mid \bm\theta,\bm{x}) p(\bm\theta\mid y^{\text{obs}},\bm{x}) d\bm\theta}\]
Either way, the modelling happens “in one go” and uncertainty is fully propagated!
| Time | Type of outcome | observed (%) | observed (%) |
|---|---|---|---|
| Baseline | utilities | 72 (96%) | 72 (86%) |
| 3 months | utilities and costs | 34 (45%) | 23 (27%) |
| 6 months | utilities and costs | 35 (47%) | 23 (27%) |
| 12 months | utilities and costs | 43 (57%) | 36 (43%) |
| Complete cases | utilities and costs | 27 (44%) | 19 (23%) |
Gabrio et al (2018). https://doi.org/10.1002/sim.8045
© Gianluca Baio (UCL) | | Mind the Gaps | ISPOR 2025 | 11 Nov 2025 | Slides available at https://gianluca.statistica.it/slides/ispor-2025