Excellent question. You've correctly identified a key challenge in frequentist inference after multiple imputation (MI) and are asking about the standard, and most accepted, solution.
Yes, a generalization for multi-parameter tests (often called "chunk tests" or omnibus tests) exists, and it's a crucial part of the MI inference framework developed by Rubin and his colleagues. The resulting test is an F-test, which is a natural generalization of the t-test.
Let's quickly recap the single-parameter case to set the stage. For a single scalar parameter
The key insight is that the uncertainty from the missing data (captured by
The generalization for testing a set of
Let
The Wald test statistic for the null hypothesis
If there were no missing data,
Test Statistic and Reference Distribution:
The test statistic is defined as:
This statistic is compared to an F-distribution
Here,
The p-value for the test is then
Note: There are alternative, more complex formulas for the denominator degrees of freedom
You asked about adjusting the covariance matrix so that standard distributions could be used. The F-test approach is the direct analogue of this. Instead of adjusting the matrix
A
where
This procedure is standard in major statistical software:
mice package is the cornerstone.
pool.compare() or the D1() function (which implements this Wald test).x2 and x3 are jointly zero:
library(mice)
imp <- mice(nhanes, m = 20, seed = 123)
fit_full <- with(imp, lm(bmi ~ age + hyp + chl))
fit_reduced <- with(imp, lm(bmi ~ age))
# D1() performs the multivariate Wald test on the difference in coefficients
# It correctly calculates the F-statistic and degrees of freedom
D1(fit_full, fit_reduced)
mi estimate prefix automatically handles this. After running an MI estimation, the test command will correctly use the F-test with the adjusted degrees of freedom.
mi estimate: reg y x1 x2 x3
test x1 x2 // This automatically uses the correct F-test
PROC MIANALYZE with the MULT statement performs this multivariate test.Your follow-up is spot on. The complexities of Rubin's rules (which are based on a frequentist approximation to a Bayesian posterior) are elegantly resolved by a fully Bayesian approach.
When you use MCMC to fit a model, you can treat the missing data as just another set of unknown parameters to be estimated. At each step of the MCMC sampler, you draw a value for the model parameters and a value for each missing data point.
By marginalizing (i.e., integrating out) the imputed data values across all posterior draws, you get the posterior distribution for your parameters of interest. This posterior automatically and correctly incorporates the uncertainty from the missing data. The tails of the posterior distributions will naturally be heavier, reflecting this extra uncertainty, without any need for special t or F distribution adjustments.
"Posterior stacking" is a pragmatic way to combine results if you run MCMC on each of the
In summary, the path you're on is correct:
D3 statistic, based on likelihood ratios).Was this answer helpful?
version: June 2025 release
Status: UQ Validated
Validated: 8 months ago
Status: Human Verified
Verified: 7 months ago
Loading model reviews...
Loading reviews...