The methods `CM`

(Cousineau, 2005; Morey, 2008) and `LM`

(Loftus & Masson, 1994) can be unified when the transformations they required are considered. In the Cousineau-Morey method, the raw data must be subject-centered and bias-corrected.

Subject centering is obtained from

\[ Y_{ij} = X_{ij} - \bar{X}_{i\cdot} + \bar{\bar{X}} \]

in which \(i = 1..n\) and \(j=1..C\) where \(n\) is the number of participants and \(C\) is the number of repeated measures (sometimes noted with \(J\)).

Bias-correction is obtained from

\[ Z_{ij} = \sqrt{\frac{C}{C-1}} \left( Y_{ij} - \bar{Y}_{\cdot{}j} \right) + \bar{Y}_{\cdot{}j} \]

These two operations can be performed with two matrix transformations. In comparison, the `LM`

method requires one additional step, that is, pooling standard deviation, also achievable with the following transformation.

\[ W_{ij} = \sqrt{\frac{S_p^2}{S_i^2}} \left( Z_{ij} - \bar{Z}_{\cdot{}j} \right) + \bar{Z}_{\cdot{}j} \]

in which \(S_j^2\) is the variance in measurement \(j\) and \(S_p^2\) is the pooled variance across all \(j\) mesurements.

With this approach, we can categorize all the proposals to repeated measure precision as requiring or not certain transformations. Table 1 shows these.

**Table 1**. Transformations required to implement one of the repeated-measures method. Preprocessing must precede post-processing

Method | preprocessing | postprocessing | |
---|---|---|---|

Stand-alone | - | - | |

Cousineau, 2005 | Subject-centering | - | - |

CM | Subject-centering | Bias-correction | |

NKM | Subject-centering | - | pool S |

LM | Subject-centering | Bias-correction | pool S |

From that point of view, we see that the Nathoo, Kilshaw and Masson NKM (Nathoo, Kilshaw, & Masson, 2018; but see Heck, 2019 ) method is missing a bias-correction transformation, which explains why these error bars are shorter. The original proposal found in Cousineau, 2005, is also missing the bias correction step, which led Morey (2008) to supplement this approach. With these four approaches, we have exhausted all the possible combinations regarding decorrelation methods based on subject-centering.

We added two arguments in superbPlot to handle this transformation approach, the first is `preprocessfct`

and the second is `postprocessfct`

.

Assuming a dataset dta with replicated measures stored in say columns called `Score.1`

, `Score.2`

and `Score.3`

, the command

```
pCM <- superbPlot(dta, WSFactors = "moment(3)",
variables = c("Score.1","Score.2","Score.3"),
adjustments=list(decorrelation="none"),
preprocessfct = "subjectCenteringTransform",
postprocessfct = "biasCorrectionTransform",
plotStyle = "pointjitter",
errorbarParams = list(color="red", width= 0.1, position = position_nudge(-0.05) )
)
```

will reproduce the `CM`

error bars because it decorrelates the data as per this method. With one additional transformation,

```
pLM <- superbPlot(dta, WSFactors = "moment(3)",
variables = c("Score.1","Score.2","Score.3"),
adjustments=list(decorrelation="none"),
preprocessfct = "subjectCenteringTransform",
postprocessfct = c("biasCorrectionTransform","poolSDTransform"),
plotStyle = "line",
errorbarParams = list(color="orange", width= 0.1, position = position_nudge(-0.0) )
)
```

the `LM`

method is reproduced. Finally, if the `biasCorrectionTransform`

is omitted, we get the NKM error bars with:

```
pNKM <- superbPlot(dta, WSFactors = "moment(3)",
variables = c("Score.1","Score.2","Score.3"),
adjustments=list(decorrelation="none"),
preprocessfct = "subjectCenteringTransform",
postprocessfct = c("poolSDTransform"),
plotStyle = "line",
errorbarParams = list(color="blue", width= 0.1, position = position_nudge(+0.05) )
)
```

In what follow, I justapose the three plots to see the differences:

```
tlbl <- paste( "(red) Subject centering & Bias correction == CM\n",
"(orange) Subject centering, Bias correction & Pooling SDs == LM\n",
"(blue) Subject centering & Pooling SDs == NKM", sep="")
ornate <- list(
xlab("Group"),
ylab("Score"),
labs( title=tlbl),
coord_cartesian( ylim = c(12,18) ),
theme_light(base_size=16)
)
# the plots on top are made transparent
pCM2 <- ggplotGrob(pCM + ornate)
pLM2 <- ggplotGrob(pLM + ornate + makeTransparent() )
pNKM2 <- ggplotGrob(pNKM + ornate + makeTransparent() )
# put the grobs onto an empty ggplot
ggplot() +
annotation_custom(grob=pCM2) +
annotation_custom(grob=pLM2) +
annotation_custom(grob=pNKM2)
```

The method from Cousineau (2005) missing the bias-correction step is not shown as it should not be used.

All the decorrelation methods based on transformations have (probably) been explored. An alternative approach using correlation was proposed in Cousineau (2019). All these approaches requires sphericity of the data. Other approaches are required to overcome this sphericity limitations.

Cousineau, D. (2005). Confidence intervals in within-subject designs: A simpler solution to Loftus and Masson’s method. *Tutorials in Quantitative Methods for Psychology*, *1*, 42–45. https://doi.org/10.20982/tqmp.01.1.p042

Cousineau, D. (2019). Correlation-adjusted standard errors and confidence intervals for within-subject designs: A simple multiplicative approach. *The Quantitative Methods for Psychology*, *15*, 226–241. https://doi.org/10.20982/tqmp.15.3.p226

Heck, D. W. (2019). Accounting for estimation uncertainty and shrinkage in bayesian within-subject intervals: A comment on Nathoo, Kilshaw, and Masson (2018). *Journal of Mathematical Psychology*, *88*, 27–31. https://doi.org/10.1016/j.jmp.2018.11.002

Loftus, G. R., & Masson, M. E. J. (1994). Using confidence intervals in within-subject designs. *Psychonomic Bulletin & Review*, *1*, 476–490. https://doi.org/10.3758/BF03210951

Morey, R. D. (2008). Confidence intervals from normalized data: A correction to Cousineau (2005). *Tutorials in Quantitative Methods for Psychology*, *4*, 61–64. https://doi.org/10.20982/tqmp.04.2.p061

Nathoo, F. S., Kilshaw, R. E., & Masson, M. E. J. (2018). A better (bayesian) interval estimate for within-subject designs. *Journal of Mathematical Psychology*, *86*, 1–9. https://doi.org/10.1016/j.jmp.2018.07.005