While a tree test is running, it’s good practice to log in every day or two and see how things are coming along.

We all hope that our studies attract lots of participants, of each type we ask for, and that the results show how much better our new tree is than the old one.

While that does sometimes happen, more often we find that things are not quite running to plan. Here are the problems we encounter most when we check the progress of a study.

 

Not enough responses

Low participant numbers are the most common worry for any online study.

If we’re halfway through the test period and we only have a quarter of the participants we hoped for, we may need to work harder to find people.

Missing user groups

When we target several user groups in a single tree test, sometimes we get lots of people from group A and B, but hardly any from group C. If we included a survey question that identified the participant’s user group, we can check that now to see if any groups are lagging and need more recruiting effort.

Obviously, the best way to boost a certain group’s numbers is to invite more of that group.

Unbalanced numbers between tests

Earlier we talked about splitting users randomly among tests. Usually this is an even split (e.g. two tests would each have a 50% chance of being selected), but sometimes we find that, halfway through the test period, test A has two thirds of the responses for some reason.

Whether we used code or a set of arbitrarily split links, we can change this partway through the test to even up the numbers.

 

Low success rates at first

Besides the number of participants, the other big thing we’re sure to check is the scoring – how well our tree is performing overall, and how individual tasks are doing.

Very often, we’ll be surprised (and appalled) by how low the interim scores are. Some part of the low scores will be justified – especially in a first-round test of a new tree, parts of that tree will simply not work well for participants. Testing simply lets us identify the parts that need rethinking.

However, we also find that interim scores are often lower than expected because:

Very high task scores

Ideally, a high task score means that we did our job well when we created the tree.

Unfortunately, it can also mean that we included a “giveaway” word (and didn’t spot it during piloting). If we did, then this isn’t a fair measure of the real-word effectiveness of the tree.

Again, we shouldn’t edit the task’s wording while the test is running (unless we spot it very early); fix it in the next round.

 

High drop-out rates

We may find that lots of participants start our study, but many drop out before they finish it.

We’ll always have some drop-off (it’s the nature of online studies), but if it exceeds about 25%, we should investigate.

 


Next: Keeping stakeholders informed