Inspired by a post on Andrew Gelman’s blog, The harm done by tests of signficance:
The first four steps to understanding cause and effect:
- Formulate your signal hypotheses,
where i = 1, …, n.
- Fit your signal models to your data,
. Obtain model parameter values,
, under each model.
- Reality-check your fit results. Does at least one of the fit models do a decent job of fitting the data? (If
is crazy low for all of your signal hypotheses then either you’ve got a highly anomalous observation on your hands or your signal hypotheses do not include the one which gave rise to the data.)
- Compute posterior probabilities:
Get through those steps and you may have a story to tell. That said, your estimates of effects and of which signal hypothesis best explains the observation are only as good as your data models, the associated probability distribution functions for measurement noise/covariance, and your estimates of . Worth noting too that if your set of signal hypotheses is incomplete that could completely blow the
calculations.
I once heard someone remark, “Statistics is an arbitrary way of being reasonable.” With that in mind and the results of my fits in hand I might report the effect (and the associated uncertainty) corresponding to the largest ,
, or maybe I’d report the weighted value of the effect where
served as the weighting term, i.e.,
The goal should be to provide insight. There are a number of ways one could do that. If viewing one’s results from modestly different but seemingly reasonable perspectives yields significantly different conclusions then you should investigate and figure out why. Robust conclusions should be minimally sensitive to modest deviations from model assumptions and the fine points of how you interpret fit results.
Bottom line: The objective of any analysis should be to gain insight. Looking to p-values for Truth elevates dogma over insight.
PS To make Steps 2 and 4 above more Bayesian let
where is a prior for the distribution of
.