# Factorial designs

## Puzzle 1

When Zach entered the JIG:SAW complex he tried garlic, rTMS and Tasers to deter the zombie security. Afterwards, he decided to do a study to test which of these weapons was best. Based on his experiences, he predicted that Tasers and rTMS would be better deterrents than garlic, but that there would be no significant difference between Tasers and rTMS. He randomly assigned 5 people a Taser, 5 people an rTMS device, and 5 people a necklace of garlic. He then put them in a pit with 10 hungry zombies. He counted how many of the 10 zombies were incapacitated for each participant. The data are in Table 16.10 (in the book). Compute the F-ratio. Were there significant differences between the mean number of zombies incapacitated by the different weapons?

The summary table for the model is as follows:

Table 1: Summary table of the model predicting the number of zombies incapacitated from the weapon used
term df sumsq meansq statistic p.value
Weapon 2 42.13 21.07 14.7 0.001
Residuals 12 17.20 1.43

Let’s see how we calculate these values. First, we compute the total sum of squares:

$$SS_\text{T} = \sum_{i = 1}^{N}(x_i - \bar{x}_\text{grand})^2 = 59.35$$

We arrive at this value (approx) by subtracting the grand mean from each score, squaring these differences and adding them up, as shown in the following table

Calculating sums of squares
ID Zombies ($x$) Weapon Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
1 5 Taser 3.667 1.333 1.777
2 2 Taser 3.667 -1.667 2.779
3 5 Taser 3.667 1.333 1.777
4 6 Taser 3.667 2.333 5.443
5 3 Taser 3.667 -0.667 0.445
6 6 rTMS 3.667 2.333 5.443
7 4 rTMS 3.667 0.333 0.111
8 7 rTMS 3.667 3.333 11.109
9 5 rTMS 3.667 1.333 1.777
10 5 rTMS 3.667 1.333 1.777
11 2 Garlic 3.667 -1.667 2.779
12 1 Garlic 3.667 -2.667 7.113
13 1 Garlic 3.667 -2.667 7.113
14 2 Garlic 3.667 -1.667 2.779
15 1 Garlic 3.667 -2.667 7.113
Sum 59.335

Next, we calculate the model sum of squares

\begin{aligned} SS_\text{M} &= \sum_{i=1}^{k} n_k(\bar{x}_k - \bar{x})^2 \\ &= 5(4.2-3.667)^2 + 5(5.4-3.667)^2 + 5(1.4-3.667)^2 \\ &= 42.13. \end{aligned}

Now, we calculate the residual sum of squares

\begin{aligned} SS_\text{R} &= \sum(Y_{ik}-\bar{Y}_k)^2 \\ &= 10.8 + 5.2 + 1.2 \\ &= 17.2. \end{aligned}

For each score we compute the difference between the score and the group mean, then square these differences and sum them, as shown in the following table:

Calculating sums of squares for Taser
ID Zombies ($x$) Weapon Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
1 5 Taser 4.2 0.8 0.64
2 2 Taser 4.2 -2.2 4.84
3 5 Taser 4.2 0.8 0.64
4 6 Taser 4.2 1.8 3.24
5 3 Taser 4.2 -1.2 1.44
Sum 10.800

Calculating sums of squares for rTMS
ID Zombies ($x$) Weapon Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
6 6 rTMS 5.4 0.6 0.36
7 4 rTMS 5.4 -1.4 1.96
8 7 rTMS 5.4 1.6 2.56
9 5 rTMS 5.4 -0.4 0.16
10 5 rTMS 5.4 -0.4 0.16
Sum 5.200

Calculating sums of squares for Garlic
ID Zombies ($x$) Weapon Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
11 2 Garlic 1.4 0.6 0.36
12 1 Garlic 1.4 -0.4 0.16
13 1 Garlic 1.4 -0.4 0.16
14 2 Garlic 1.4 0.6 0.36
15 1 Garlic 1.4 -0.4 0.16
Sum 1.200

We convert the sums of squares to mean squares by dividing by the degrees of freedom for each:

\begin{aligned} MS_\text{M} &= \frac{SS_\text{M}}{df_\text{M}} = \frac{42.13}{2} = 21.07 \\ MS_\text{R} &= \frac{SS_\text{R}}{df_\text{R}} = \frac{17.2}{12} = 1.433. \end{aligned}

We calculate F as

\begin{aligned} F &= \frac{MS_\text{M}}{MS_\text{R}} \\ &= \frac{21.07}{1.433} \\ &= 14.70. \end{aligned}

We conclude that there was a significant difference in the mean number of zombies incapacitated by different weapons, F(2, 12) = 14.70, p < 0.001.

## Puzzle 2

Based on Zach’s predictions, construct some contrasts and weights for those contrasts to test his predictions.

The contrasts and weights reflecting Zach’s predictions are in the following table:

Table 2: Table of contrast codes
Contrast Taser rTMS Garlic
Garlic compared to other weapon 1 1 -2
Taser compared to rTMS 1 -1 0

## Puzzle 3

Calculate omega squared for the zombie data, and also calculate Cohen’s $$d$$ for the difference between means of the three conditions (use the pooled estimate of the standard deviation).

First we calculate $^s$ as

\begin{aligned} \omega^2 &= \frac{SS_\text{M}-df_\text{M} \times MS_\text{R}}{SS_\text{T} + MS_\text{R}} \\ &= \frac{42.13-2 \times 1.433}{59.35+ 1.433} \\ &= \frac{39.264}{60.783} \\ &= 0.646. \end{aligned}

Cohen’s d is defined as

$$\hat{d} = \frac{\bar{X_1}-\bar{X_2}}{s}.$$

The pooled estimate of the standard deviation is

$$s_p = \sqrt{\frac{(N_1-1) s_1^2+(N_2-1) s_2^2}{N_1+N_2-2}}.$$

The values and resulting calculations are tabulated below

Effect size calculations
Comparison $M_1$ $M_2$ $SD_1$ $SD_2$ $n_1$ $n_2$ $M_1-M_2$ $s_p$ $d$
Taser - Garlic 4.2 1.4 2.7 0.3 5 5 2.8 1.921 1.458
rTMS - Garlic 5.4 1.4 1.3 0.3 5 5 4.0 0.943 4.242
Taser - rTMS 4.2 5.4 2.7 1.3 5 5 -1.2 2.119 -0.566

## Puzzle 4

Rob Nutcot used an oxytocin spray to try to make people trust him more. Zach set up an experiment to see whether oxytocin affected trust. He took 7 people who were delivered a speech about different products to keep you looking younger by three different people. Before each speech they were sprayed with either (1) a placebo that smelled like oxytocin spray but had no active ingredient, (2) a spray with a low dose of oxytocin in it, or (3) a spray with a high dose of oxytocin in it. Each participant was, therefore, exposed to each dose of oxytocin. The outcome was how much they believed the claims about the product (out of 10). The data are in Table 16.11. Calculate the F-ratio. Did oxytocin affect trust in the product?

The summary table for the model is as follows:

Table 3: Summary table of the model predicting the belief in the product from the dose of oxytocin administered before the speech
stratum term df sumsq meansq statistic p.value
ID Residuals 6 10.48 1.75
Within Dose 2 38.38 19.19 8.54 0.005
Within Residuals 12 26.95 2.25

Let’s see how we calculate these values. First, we compute the total sum of squares

\begin{aligned} SS_\text{T} &= \sum_{i = 1}^{N}(x_i - \bar{x}_\text{grand})^2 \\ &= 75.82. \end{aligned}

We arrive at this value by subtracting the grand mean from each score, squaring these differences and adding them up, as shown in the following table.

Calculating sums of squares
ID Dose Belief ($x$) Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
1 Placebo 5 6.238 -1.238 1.533
1 Low 5 6.238 -1.238 1.533
1 High 5 6.238 -1.238 1.533
2 Placebo 7 6.238 0.762 0.581
2 Low 5 6.238 -1.238 1.533
2 High 8 6.238 1.762 3.105
3 Placebo 6 6.238 -0.238 0.057
3 Low 4 6.238 -2.238 5.009
3 High 10 6.238 3.762 14.153
4 Placebo 4 6.238 -2.238 5.009
4 Low 8 6.238 1.762 3.105
4 High 9 6.238 2.762 7.629
5 Placebo 4 6.238 -2.238 5.009
5 Low 7 6.238 0.762 0.581
5 High 8 6.238 1.762 3.105
6 Placebo 6 6.238 -0.238 0.057
6 Low 5 6.238 -1.238 1.533
6 High 9 6.238 2.762 7.629
7 Placebo 4 6.238 -2.238 5.009
7 Low 4 6.238 -2.238 5.009
7 High 8 6.238 1.762 3.105
Sum 75.817

Next, we calculate the model sum of squares

\begin{aligned} SS_\text{M} &= \sum_{i=1}^{k}n_k(\bar{x}_k-\bar{x})^2 \\ &= 7(5.143-6.238)^2 + 7(5.429-6.238)^2 +7(8.143-6.238)^2 \\ &= 38.38. \end{aligned}

Next, we calculate the within-participant sum of squares by calculating the sum of squared error within each participant and adding them up. The table shows how to arrive at the sums of squares for each entity

Calculating sums of squares within participants
ID Mean ($\bar{X}$) Variance ($s^2$) df SS
1 5.000 0.000 2 0.000
2 6.667 2.333 2 4.666
3 6.667 9.333 2 18.666
4 7.000 7.000 2 14.000
5 6.333 4.333 2 8.666
6 6.667 4.333 2 8.666
7 5.333 5.333 2 10.666
Sum 65.330

We use these values as follows

\begin{aligned} SS_\text{W} &= SS_\text{Entity 1} + SS_\text{Entity 2} + \ldots + SS_{\text{Entity }i} \\ &= 0 + 4.666 + 18.666 + 14.000 + 8.666 + 8.666 + 10.666 \\ &= 65.33. \end{aligned}

The associated degrees of freedom are

\begin{aligned} df_\text{W} &= df_\text{Entity 1} + df_\text{Entity 2} + \ldots + df_{\text{Entity }i} \\ &= 2 + 2 + 2 + 2 + 2 + 2 + 2 \\ &= 14. \end{aligned}

Next, we calculate the residual sum of squares

\begin{aligned} SS_\text{R} &= SS_\text{W}-SS_\text{M} \\ &= 65.33 - 38.38 \\ &= 26.95, \end{aligned}

and their degrees of freedom

\begin{aligned} df_\text{R} &= df_\text{W}-df_\text{M} \\ &= 14 - 2 \\ &= 12. \end{aligned}

We convert the sums of squares to mean squares by dividing by the degrees of freedom for each:

\begin{aligned} MS_\text{M} &= \frac{SS_\text{M}}{df_\text{M}} = \frac{38.38}{2} = 19.190 \\ MS_\text{R} &= \frac{SS_\text{R}}{df_\text{R}} = \frac{26.95}{12} = 2.246. \end{aligned}

We calculate F as follows

$$F = \frac{MS_\text{M}}{MS_\text{R}} = \frac{19.190}{2.246} = 8.544$$

There was a significant difference in the belief scores after different doses of oxytocin, F(2, 12) = 8.54, p = 0.005.

## Puzzle 5

Earlier in his journey, Zach looked at some data showing the strength, footspeed and vision of different groups of JIG:SAW employees (code 1318, scientists, security, professional services and technicians; see Figure 9.1 in the book). There were 240 employees in total. Milton forced him reluctantly to test whether the mean strength differed across the five groups. Then, because he is an evil cat, he erased some of the values from the summary (Table 16.12 in the book). Can you help Zach to fill in the blanks? (Hint: think about the degrees of freedom first.)

• There were five groups being compared (code 1318, scientists, security, professional services and technicians), so the degrees of freedom for the model will be 4 ($k - 1$)
• The total degrees of freedom will be: $N - 1 = 240 - 1 = 239$
• The residual degrees of freedom will be: $df_ - df_ = 239 - 4 = 235$
• We can work out the mean squared error for the model by dividing the sum of squares by the degrees of freedom

$$MS_\text{M} = \frac{SS_\text{M}}{df_\text{M}} = \frac{8490624}{4} = 2122656$$

• We can work out the residual sum of squared errors by multiplying the mean squared error by the degrees of freedom

$$SS_\text{R} = MS_\text{R} \times df_\text{R} = 89636 \times 235 = 21064460$$

• We can work out the total sum of squares by adding the model and residual sum of squares

$$SS_\text{T} = SS_\text{M} + SS_\text{R} = 8490624 + 21064460 = 29555084$$

• The F statistic is the mean squared error for the model divided by that for the residual

$$F = \frac{MS_\text{M}}{MS_\text{R}} = \frac{2122656}{89636} = 23.68$$

## Puzzle 6

That cat, Milton, is such a japester that he has done the same thing to the summary table (Table 16.13 in the book) for the comparison between means of footspeed across the five job types (see puzzle 5). Help Zach to fill in the blanks before he dies of tedium.

• There were five groups being compared (code 1318, scientists, security, professional services and technicians), so the degrees of freedom for the model will be 4 ($k - 1$)
• The total degrees of freedom will be: $N - 1 = 240 - 1 = 239$
• The residual degrees of freedom will be: $df_ - df_ = 239 - 4 = 235$
• We can work out the model sum of squares by subtracting the residual sum of squares from the total sum of squares

$$SS_\text{M} = SS_\text{T} - SS_\text{R} = 3272 - 3256 = 16$$

• We can work out the mean squared error for the model by dividing the sum of squares by the degrees of freedom

$$MS_\text{M} = \frac{SS_\text{M}}{df_\text{M}} = \frac{16}{4} = 4$$

• We can work out the mean squared error for the residual by dividing the sum of squares by the degrees of freedom:

$$MS_\text{R} = \frac{SS_\text{R}}{df_\text{R}} = \frac{3256}{235} = 13.86$$

• The F statistic is the mean squared error for the model divided by that for the residual

$$F = \frac{MS_\text{M}}{MS_\text{R}} = \frac{4}{13.86} = 0.29$$

## Puzzle 7

When Celia was trying to woo Zach, she looked at whether there was a significant association between how you rated the costs and rewards of your relationship and openness to alternative relationships (see Reality Check 13.2 in the book). She decided to see whether you could manipulate a person’s openness to alternative relationships. She took five people and asked them to list all the costs and rewards of their current relationship, then asked them to rate their openness to alternative relationships. At another point in time she asked the same people to study a list of potential costs to being in a relationship and again got them to rate their openness to alternative relationships. At a final point in time she asked the same people to study a list of potential benefits or rewards to being in a relationship and again got them to rate their openness to alternative relationships. Each person completed these tasks a week apart and in random order. The data are in Table 16.14 (in the boiok). Use these values to compute the F-ratio. Is openness to alternative relationships affected by focusing on lists of potential costs or rewards compared to generating your own costs and rewards?

The summary table for the model is in the table below

Table 4: Summary table of the model predicting openness to a new relationship after studying different prompt lists
stratum term df sumsq meansq statistic p.value
ID Residuals 4 8.93 2.23
Within Prompt 2 12.40 6.20 6 0.026
Within Residuals 8 8.27 1.03

Let’s see how we calculate these values. First, we compute the total sum of squares:

$$SS_\text{T} = \sum_{i = 1}^{N}(x_i - \bar{x}_\text{grand})^2 = 29.60$$

We arrive at this value by subtracting the grand mean from each score, squaring these differences and adding them up, as shown in the following table

Calculating sums of squares for Openness
ID Prompt Openness ($x$) Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
1 Free_List 3 3.4 -0.4 0.16
1 Costs 5 3.4 1.6 2.56
1 Rewards 3 3.4 -0.4 0.16
2 Free_List 6 3.4 2.6 6.76
2 Costs 4 3.4 0.6 0.36
2 Rewards 3 3.4 -0.4 0.16
3 Free_List 4 3.4 0.6 0.36
3 Costs 4 3.4 0.6 0.36
3 Rewards 3 3.4 -0.4 0.16
4 Free_List 3 3.4 -0.4 0.16
4 Costs 5 3.4 1.6 2.56
4 Rewards 2 3.4 -1.4 1.96
5 Free_List 2 3.4 -1.4 1.96
5 Costs 4 3.4 0.6 0.36
5 Rewards 0 3.4 -3.4 11.56
Sum 29.600

Next, we calculate the model sum of squares:

\begin{aligned} SS_\text{M} &= \sum_{i=1}^{k}n_k(\bar{x}_k-\bar{x})^2 \\ &= 5(3.60-3.40)^2 + 5(4.40-3.40)^2 +5(2.20-3.40)^2 \\ &= 12.40 \end{aligned}

Next, we calculate the within-participant sum of squares by calculating the sum of squared error within each participant and adding them up. The sum of squared error for each participant is shown in the following table

Calculating sums of squares within participants
ID Mean ($\bar{X}$) Variance ($s^2$) df SS
1 3.667 1.333 2 2.666
2 4.333 2.333 2 4.666
3 3.667 0.333 2 0.666
4 3.333 2.333 2 4.666
5 2.000 4.000 2 8.000
Sum 20.664

We use these values as follows

\begin{aligned} SS_\text{W} &= SS_\text{Entity 1} + SS_\text{Entity 2} + \ldots + SS_{\text{Entity }i} \\ &= 2.666 + 4.666 + 0.666 + 4.666 + 8.000 \\ &= 20.664 \end{aligned}

The associated degrees of freedom are

\begin{aligned} df_\text{W} &= df_\text{Entity 1} + df_\text{Entity 2} + \ldots + df_{\text{Entity }i} \\ &= 2 + 2 + 2 + 2 + 2 \\ &= 10. \end{aligned}

Next, we calculate the residual sum of squares:

\begin{aligned} SS_\text{R} &= SS_\text{W}-SS_\text{M} \\ &= 20.664 - 12.40 \\ &= 8.264, \end{aligned}

with associated degrees of freedom

\begin{aligned} df_\text{R} &= df_\text{W}-df_\text{M} \\ &= 10 - 2 \\ &= 8. \end{aligned}

We convert the sums of squares to mean squares by dividing by the degrees of freedom for each

\begin{aligned} MS_\text{M} &= \frac{SS_\text{M}}{df_\text{M}} = \frac{12.40}{2} = 6.200 \\ MS_\text{R} &= \frac{SS_\text{R}}{df_\text{R}} = \frac{8.264}{8} = 1.033. \end{aligned}

We calculate F as

$$F = \frac{MS_\text{M}}{MS_\text{R}} = \frac{6.200}{1.033} = 6.002.$$

There was a significant difference in the openness to a new relationship after different prompts to thing about pros and cons of relationships, F(2, 8) = 6.00, p = 0.026.

## Puzzle 8

Output 16.13 shows the results of an analysis on the data in Table 16.14 (see puzzle 7), to compute a Bayes factor based on a reasonable default prior. Interpret this output.

The Bayes factor is 5.08, which means that the data are 5.08 times more likely given the alternative hypothesis compared to the null hypothesis. In other words, we should shift our belief in the alternative hypothesis (that these prompts change openness to new relationships) relative to the null by a factor of 5.08. This value ‘has substance’ but is not at all strong evidence for the alternative hypothesis (relative to the null). In other words we should shift our belief that these prompts change openness to new relationships a little, but not very much.

## Puzzle 9

The second experiment in Alice’s report (Alice’s Lab Notes 16.1 in the book), when Zach reanalysed it, seemed to suggest that sending visual images to someone’s ID chip while they recalled a memory would interfere with it (i.e., result in lower recall of the event). Roediger decided to do a follow-up study. He took 18 participants: 6 of them received no memory interference while they recalled an event, 6 received visual interference unrelated to the event they were recalling, and 6 received visual interference related to the event they were recalling. The data are in Table 16.15. Compute the F-ratio. Were there significant differences in recall between the three groups?

The summary table for the model is as follows:

Table 5: Summary table of the model predicting memory recall from the type of interference
term df sumsq meansq statistic p.value
Interference 2 80.11 40.06 12.78 0.001
Residuals 15 47.00 3.13

Let’s see how we calculate these values. First, we compute the total sum of squares:

$$SS_\text{T} = \sum_{i = 1}^{N}(x_i - \bar{x}_\text{grand})^2 = 127.106$$

We arrive at this value by subtracting the grand mean from each score, squaring these differences and adding them up, as shown in the following tables

Calculating sums of squares
ID Recall ($x$) Interference Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
1 6 No Interference 5.222 0.778 0.605
2 8 No Interference 5.222 2.778 7.717
3 8 No Interference 5.222 2.778 7.717
4 10 No Interference 5.222 4.778 22.829
5 7 No Interference 5.222 1.778 3.161
6 8 No Interference 5.222 2.778 7.717
7 6 Unrelated Interference 5.222 0.778 0.605
8 4 Unrelated Interference 5.222 -1.222 1.493
9 8 Unrelated Interference 5.222 2.778 7.717
10 6 Unrelated Interference 5.222 0.778 0.605
11 4 Unrelated Interference 5.222 -1.222 1.493
12 3 Unrelated Interference 5.222 -2.222 4.937
13 1 Related Interference 5.222 -4.222 17.825
14 0 Related Interference 5.222 -5.222 27.269
15 5 Related Interference 5.222 -0.222 0.049
16 2 Related Interference 5.222 -3.222 10.381
17 5 Related Interference 5.222 -0.222 0.049
18 3 Related Interference 5.222 -2.222 4.937
Sum 127.106

Next, we calculate the model sum of squares

\begin{aligned} SS_\text{M} &= \sum_{i=1}^{k}n_k(\bar{x}_k-\bar{x})^2 \\ &= 6(7.833-5.222)^2 + 6(5.167-5.222)^2 +6(2.667-5.222)^2 \\ &= 80.09 \end{aligned}

The associated degrees of freedom are $k-1 = 2$.

Next, we calculate the residual sum of squares. For each score we compute the difference between the score and the group mean, then square these differences and sum them, as shown in the following tables

Residual sum of squares for the no interference group
ID Recall ($x$) Interference Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
1 6 No Interference 7.833 -1.833 3.360
2 8 No Interference 7.833 0.167 0.028
3 8 No Interference 7.833 0.167 0.028
4 10 No Interference 7.833 2.167 4.696
5 7 No Interference 7.833 -0.833 0.694
6 8 No Interference 7.833 0.167 0.028
Sum 8.834

Residual sum of squares for the unrelated interference group
ID Recall ($x$) Interference Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
7 6 Unrelated Interference 5.167 0.833 0.694
8 4 Unrelated Interference 5.167 -1.167 1.362
9 8 Unrelated Interference 5.167 2.833 8.026
10 6 Unrelated Interference 5.167 0.833 0.694
11 4 Unrelated Interference 5.167 -1.167 1.362
12 3 Unrelated Interference 5.167 -2.167 4.696
Sum 16.834

Residual sum of squares for the related interference group
ID Recall ($x$) Interference Mean ($\bar{X}$) Error ($x-\bar{X}$) $\text{Error}^2$
13 1 Related Interference 2.667 -1.667 2.779
14 0 Related Interference 2.667 -2.667 7.113
15 5 Related Interference 2.667 2.333 5.443
16 2 Related Interference 2.667 -0.667 0.445
17 5 Related Interference 2.667 2.333 5.443
18 3 Related Interference 2.667 0.333 0.111
Sum 21.334

We use these sums as follows

\begin{aligned} SS_\text{R} &= \sum(Y_{ik}-\bar{Y}_k)^2 \\ &= 8.834 + 16.834 + 21.334 \\ &= 47.00 \end{aligned}

The associated degrees of freedom are

\begin{aligned} df &= df_\text{No interference} + df_\text{Unrelated interference} + df_\text{related interference} \\ &= (n_\text{No interference}-1) + (n_\text{Unrelated interference}-1) + (n_\text{related interference}-1) \\ &= 5 + 5 + 5 \\ &= 15 \end{aligned}

We convert the sums of squares to mean squares by dividing by the degrees of freedom for each

\begin{aligned} MS_\text{M} &= \frac{SS_\text{M}}{df_\text{M}} = \frac{80.09}{2} = 40.05 \\ MS_\text{R} &= \frac{SS_\text{R}}{df_\text{R}} = \frac{47.00}{15} = 3.13. \\ \end{aligned}

We calculate F as

$$F = \frac{MS_\text{M}}{MS_\text{R}} = \frac{40.05}{3.13} = 12.80$$

There was a significant difference in the mean recall after different types of interference, F(2, 15) = 12.80, p < 0.001.

## Puzzle 10

Output 16.14 shows the results of an analysis on the data in Table 16.15 (see puzzle 9), to compute a Bayes factor based on a reasonable default prior. Interpret this output.

The Bayes factor is 51.57, which means that the data are 51.57 times more likely given the alternative hypothesis compared to the null hypothesis. In other words, we should shift our belief in the alternative hypothesis (that the type of interference affects memory) relative to the null by a factor of 51.57 This value is strong evidence for the alternative hypothesis (relative to the null). In other words we should shift our belief that the type of interference affects memory a fair amount.

Previous
Next