(advanced) Alternate ways to decorrelate repeated measures from transformations

The methods CM (Cousineau, 2005; Morey, 2008) and LM (Loftus & Masson, 1994) can be unified when the transformations they required are considered. In the Cousineau-Morey method, the raw data must be subject-centered and bias-corrected.

Subject centering is obtained from

\[ Y_{ij} = X_{ij} - \bar{X}_{i\cdot} + \bar{\bar{X}} \]

in which \(i = 1..n\) and \(j=1..C\) where \(n\) is the number of participants and \(C\) is the number of repeated measures (sometimes noted with \(J\)).

Bias-correction is obtained from

\[ Z_{ij} = \sqrt{\frac{C}{C-1}} \left( Y_{ij} - \bar{Y}_{\cdot{}j} \right) + \bar{Y}_{\cdot{}j} \]

These two operations can be performed with two matrix transformations. In comparison, the LM method requires one additional step, that is, pooling standard deviation, also achievable with the following transformation.

\[ W_{ij} = \sqrt{\frac{S_p^2}{S_i^2}} \left( Z_{ij} - \bar{Z}_{\cdot{}j} \right) + \bar{Z}_{\cdot{}j} \]

in which \(S_j^2\) is the variance in measurement \(j\) and \(S_p^2\) is the pooled variance across all \(j\) mesurements.

With this approach, we can categorize all the proposals to repeated measure precision as requiring or not certain transformations. Table 1 shows these.

Table 1. Transformations required to implement one of the repeated-measures method. Preprocessing must precede post-processing

Method preprocessing postprocessing
Stand-alone - -
Cousineau, 2005 Subject-centering - -
CM Subject-centering Bias-correction
NKM Subject-centering - pool standard deviations
LM Subject-centering Bias-correction pool standard deviations

From that point of view, we see that the Nathoo, Kilshaw and Masson NKM (Nathoo, Kilshaw, & Masson, 2018; but see Heck, 2019 ) method is missing a bias-correction transformation, which explains why these error bars are shorter. The original proposal found in Cousineau, 2005, is also missing the bias correction step, which led Morey (2008) to supplement this approach. With these four approaches, we have exhausted all the possible combinations regarding decorrelation methods based on subject-centering.

We added two arguments in superbPlot to handle this transformation approach, the first is preprocessfct and the second is postprocessfct.

Assuming a dataset dta with replicated measures stored in say columns called Score.1, Score.2 and Score.3, the command

pCM <- superb(
    crange(Score.1,Score.3) ~ .,
    dta, WSFactors = "moment(3)",     
    adjustments=list(decorrelation="none"),
    preprocessfct = "subjectCenteringTransform",
    postprocessfct = "biasCorrectionTransform",
    plotStyle = "pointjitter",
    errorbarParams = list(color="red", width= 0.1, position = position_nudge(-0.05) )
)

will reproduce the CM error bars because it decorrelates the data as per this method. With one additional transformation,

pLM <- superb(
    crange(Score.1,Score.3) ~ .,    
    dta, WSFactors = "moment(3)", 
    adjustments=list(decorrelation="none"),
    preprocessfct = "subjectCenteringTransform",
    postprocessfct = c("biasCorrectionTransform","poolSDTransform"),
    plotStyle = "line",
    errorbarParams = list(color="orange", width= 0.1, position = position_nudge(-0.0) )
)

the LM method is reproduced. Finally, if the biasCorrectionTransform is omitted, we get the NKM error bars with:

pNKM <- superb(
    crange(Score.1,Score.3) ~ .,
    dta, WSFactors = "moment(3)", 
    adjustments=list(decorrelation="none"),
    preprocessfct = "subjectCenteringTransform",
    postprocessfct = c("poolSDTransform"),
    plotStyle = "line",
    errorbarParams = list(color="blue", width= 0.1, position = position_nudge(+0.05) )
)

In what follow, I justapose the three plots to see the differences:

tlbl <- paste( "(red)    Subject centering & Bias correction == CM\n",
               "(orange) Subject centering, Bias correction & Pooling SDs == LM\n",
               "(blue)   Subject centering & Pooling SDs == NKM", sep="")

ornate <- list(
    xlab("Group"),
    ylab("Score"),
    labs(   title=tlbl),
    coord_cartesian( ylim = c(12,18) ),
    theme_light(base_size=10)
)

# the plots on top are made transparent
pCM2 <- ggplotGrob(pCM + ornate)
pLM2 <- ggplotGrob(pLM + ornate + makeTransparent() )
pNKM2 <- ggplotGrob(pNKM + ornate + makeTransparent() )

# put the grobs onto an empty ggplot 
ggplot() + 
    annotation_custom(grob=pCM2) + 
    annotation_custom(grob=pLM2) + 
    annotation_custom(grob=pNKM2)
Figure 1. Plot of the tree decorrelation methods based on subject transformation.
Figure 1. Plot of the tree decorrelation methods based on subject transformation.

The method from Cousineau (2005) missing the bias-correction step is not shown as it should not be used.

In summary

All the decorrelation methods based on transformations have (probably) been explored. An alternative approach using correlation was proposed in Cousineau (2019). All these approaches requires sphericity of the data. Other approaches are required to overcome this sphericity limitations.

References

Cousineau, D. (2005). Confidence intervals in within-subject designs: A simpler solution to Loftus and Masson’s method. Tutorials in Quantitative Methods for Psychology, 1, 42–45. https://doi.org/10.20982/tqmp.01.1.p042
Cousineau, D. (2019). Correlation-adjusted standard errors and confidence intervals for within-subject designs: A simple multiplicative approach. The Quantitative Methods for Psychology, 15, 226–241. https://doi.org/10.20982/tqmp.15.3.p226
Heck, D. W. (2019). Accounting for estimation uncertainty and shrinkage in bayesian within-subject intervals: A comment on Nathoo, Kilshaw, and Masson (2018). Journal of Mathematical Psychology, 88, 27–31. https://doi.org/10.1016/j.jmp.2018.11.002
Loftus, G. R., & Masson, M. E. J. (1994). Using confidence intervals in within-subject designs. Psychonomic Bulletin & Review, 1, 476–490. https://doi.org/10.3758/BF03210951
Morey, R. D. (2008). Confidence intervals from normalized data: A correction to Cousineau (2005). Tutorials in Quantitative Methods for Psychology, 4, 61–64. https://doi.org/10.20982/tqmp.04.2.p061
Nathoo, F. S., Kilshaw, R. E., & Masson, M. E. J. (2018). A better (bayesian) interval estimate for within-subject designs. Journal of Mathematical Psychology, 86, 1–9. https://doi.org/10.1016/j.jmp.2018.07.005