If I wanted the Contrast that denotes the difference between ControlB and TreatB, or the difference between ControlA and TreatA, would I need to make a contrast matrix, or is that accessible through one of the columns in the design matrix?
In that design, the last three coefficients are comparisons to the baseline, which is TreatmentsControlA (e.g., TreatmentsControlB is actually TreatmentsControlB - TreatmentsControlA), so you already get the comparison of TreatmentsTreatA and TreatmentsControlA, but any other comparisons (TreatmentsTreatB - TreatmentsControlB) have to be done using a contrasts matrix. Since you need a contrasts matrix anyway, you could eliminate the intercept and just make all the comparisons you need.
So yes, you can model the experiments together, and contingent upon the intra-group variability being relatively similar between the different groups, this will increase power. If one of the experiments is more variable than the other, then you may reduce power in the less variable experiment because you may be introducing extra variability. Or you may be correcting for artificially low variability, depending on your interpretation.
Well, if I fit a none intercept model (~0 + Pairings + Treatments), then as expected, TreatmentsControlA gets dropped as the baseline. Would it make more sense to use the model ~0 + Treatments + Pairings so that all terms are present for the contrast matrix? Then it'd be a case of:
In fact, the design matrix isn't of full rank; you need to drop TreatmentsControlB for the coefficients to be estimable. The TreatmentsTreatA term represents the treatment effect in experiment A, while the TreatmentsTreatB term represents the effect in experiment B. You can drop them individually to test for the treatment effect in each experiment, or you can do fancier stuff, e.g., compare them to each other, drop both of them, test the average of them against zero.
Edit: Note that I'm referring to the design matrix in your original post. The design matrix in your comment would require you to drop, e.g., Pairingspair4 to get to full column rank. It's a lot harder to interpret the coefficients in the second design compared to the first, because you don't have a blocking term for each pairing.
Cheers Aaron, really useful. I agree that the first model is far easier to interpret after dropping those terms to get full rank. I've seen the post with Gordon's answer before, but useful to go over again!
Well, if I fit a none intercept model (~0 + Pairings + Treatments), then as expected, TreatmentsControlA gets dropped as the baseline. Would it make more sense to use the model ~0 + Treatments + Pairings so that all terms are present for the contrast matrix? Then it'd be a case of:
That should still utilise the information between the pairs wouldn't it?
In fact, the design matrix isn't of full rank; you need to drop
TreatmentsControlB
for the coefficients to be estimable. TheTreatmentsTreatA
term represents the treatment effect in experiment A, while theTreatmentsTreatB
term represents the effect in experiment B. You can drop them individually to test for the treatment effect in each experiment, or you can do fancier stuff, e.g., compare them to each other, drop both of them, test the average of them against zero.Also check out Gordon's comment (A: Correct assumptions of using limma moderated t-test) regarding unequal variances.
Edit: Note that I'm referring to the design matrix in your original post. The design matrix in your comment would require you to drop, e.g.,
Pairingspair4
to get to full column rank. It's a lot harder to interpret the coefficients in the second design compared to the first, because you don't have a blocking term for each pairing.Cheers Aaron, really useful. I agree that the first model is far easier to interpret after dropping those terms to get full rank. I've seen the post with Gordon's answer before, but useful to go over again!