While a tree test is running, it’s good practice to log in every day or two and see how things are coming along.

We all hope that our studies attract lots of participants, of each type we ask for, and that the results show how much better our new tree is than the old one.

While that does sometimes happen, more often we find that things are not quite running to plan. Here are the problems we encounter most when we check the progress of a study.

 

Not enough responses

Low participant numbers are the most common worry for any online study.

If you’re halfway through your test period and you only have a quarter of the participants you planned for, you may need to work harder to find people.

Missing user groups

When we target several user groups in a single tree test, sometimes we get lots of people from group A and B, but hardly any from group C. If you included a survey question that identified the participant’s user group, you can check that now to see if any groups are lagging and need more recruiting effort.

Obviously, the best way to boost a certain group’s numbers is to invite more of that group.

Unbalanced numbers between tests

Earlier we talked about splitting users randomly among tests. Usually this is an even split (e.g. two tests would each have a 50% chance of being selected), but sometimes we find that, halfway through the test period, test A has two thirds of the responses for some reason.

Whether you used code or a set of arbitrarily split links (first name A-M, first name N-Z, etc.), you can change this partway through the test to even up the numbers. For example, you may change the split so that test A now gets 30% of the clicks, while test B (the one that’s lagging) gets 70%.

 

Low success rates at first

Besides the number of participants, the other big thing you’re sure to check is the scoring – how well your tree is performing overall, and how individual tasks are doing.

Very often, you’ll be surprised (and appalled) by how low your interim scores are. Some part of the low scores will be justified – especially in a first-round test of a new tree, parts of that tree will simply not work well for your participants. Testing simply lets you identify the parts that need rethinking.

However, we also find that interim scores are often lower than expected because:

Very high task scores

Ideally, a high task score means that you did your job well when you created the tree.

Unfortunately, it can also mean that you included a “giveaway” word (and didn’t spot it during piloting). If you did, then this isn’t a fair measure of the real-word effectiveness of your tree.

Again, you shouldn’t edit the task’s wording while the test is running (unless you spot it very early); fix it in the next round.

 

High drop-out rates

You may find that lots of participants start your study, but many drop out before they finish it.

You’ll always have some drop-off (it’s the nature of online studies), but if it exceeds about 25%, you should investigate.

 


Next: Keeping stakeholders informed