generated from IQSS/dss-template-quarto
-
Notifications
You must be signed in to change notification settings - Fork 1
/
respecification.qmd
19 lines (10 loc) · 4.65 KB
/
respecification.qmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
## Respecification {#sec-respecification}
If a given specification is not adequate in that balance is too poor, the effective sample size is too small, or the sample is no longer representative of the target population, one must respecify. Respecification can involve changing some aspect of the conditioning strategy, such as changing a parameter involved in the matching or weighting or changing the model used to estimate propensity scores, if any. Because there are so many parameters that can be changed and they can be changed in so many ways, it is impossible to give a complete account of the best way to respecify. One should try many specifications, examining patterns in how making those changes improves the quality of the resulting sample. As long as the outcome is not involved in this process, doing so will not invalidate inferences made at the end.
There are some common tricks that can be used to nudge the respecification process in the right direction. Below are some common issues and some potential solutions.
- Poor balance as measured by SMDs: consider using an optimization-based method, like entropy balancing [@hainmuellerEntropyBalancingCausal2012] or cardinality matching [@zubizarretaMatchingBalancePairing2014] or using a method that changes the estimand, like caliper matching or overlap weighting
- Poor balance beyond SMDs (e.g., on polynomial terms, variance ratios, or KS statistics): consider adding polynomial or interaction terms to the propensity score model; using a machine-learning method that flexibly models the propensity score [@leeImprovingPropensityScore2010]; using an optimization-based method that balances the full distribution, like energy balancing [@hulingEnergyBalancingCovariate2024]; using coarsened exact matching to balance the full distribution approximately [@iacusCausalInferenceBalance2012]; or adding an exact matching constraint to a matching specification
- Low ESS: consider using a method to regularize the propensity score model (e.g., ridge or lasso regression); increasing the matching ratio; using an optimization-based method that maximizes the ESS (e.g., profile matching or stable balancing weights [@zubizarretaStableWeightsThat2015]); relaxing the caliper (if used); trimming extreme weights; or using overlap weighting
- Poor representativeness: consider using a method that strongly respects the estimand (e.g., entropy balancing; not cardinality matching, caliper matching, or overlap weighting) or removing a caliper or exact matching restriction
Having broad experience with the variety of matching and weighting methods available makes this process quick. Fortunately, the software we recommend and use in the examples, the R packages `MatchIt` and `WeightIt`, make switching between various specifications easy.
To avoid endless respecification, it is a good idea to use methods designed to optimize the evaluation criteria in a simple way. Often, the oldest and most commonly used methods are the worst in that they perform poorly and require manual respecification to get right. For example, 1:1 propensity score matching with a caliper is the most commonly used propensity score method in medical research, but it is widely known to have many problems: it hampers representativeness because the caliper discards units from both treatment groups [@rosenbaumBiasDueIncomplete1985], it reduces the effective sample size by dropping many units from the sample, it can make balance worse when used thoughtlessly [@kingWhyPropensityScores2019], and it has many specification parameters that need to be adjusted arbitrarily (e.g., the propensity score model, caliper width, matching order, etc.). Another popular but old method, propensity score weighting, also has many problems, including inability to achieve balance, low ESS due to extreme weights, and reduced representativeness when measures are taken to rectify the other problems.
Methods that consistently perform well include entropy balancing [@hainmuellerEntropyBalancingCausal2012] and energy balancing [@hulingEnergyBalancingCovariate2022], as these ensure balance and representativeness without requiring major respecification. Entropy balancing guarantees exact balance as measured by the SMD, but it may be necessary to include other terms to fully balance the covariate distributions. Energy balancing balances the full covariate distribution, but can decrease ESS (though the trade-off between them can be managed with a single parameter). Though these methods are newer, they are beginning to see use in applied research [e.g., @bramante2022; @sharma2023] and should be the first line of defense when adjusting for confounders rather than poorly performing but older and more familiar methods.