Better planning
The solution is not to abandon statistics. The statistical analysis, if done right, can go a long way in helping us to talk less nonsense.
- We need shorter, more focused questionnaires
- We need to put thought into what we want to know, from who, and how to extract it reliably
- We need to plan the full analysis in detail before collecting data
- In medical trials the full analysis is often planned out to the finest detail, including power analysis, a year before the first participant is recruited
The best surveys
The best surveys in the world have 3 questions:
- Informed consent
- Rate your perception of X on a scale of 0 to 10
- Explain your rating
- The absolute best surveys in the world have 0 questions, where the responses to the above are implied by directly observed behaviour following randomised assignment
- A/B testing is a form of survey and can produce very definitive results if the research question can be formulated in a suitable way
Let’s be practical
- Obviously we can’t do a whole new survey every time we want to ask a slightly different question
- It is far more convenient to us to go out once and get responses about everything we are interested in
But if we don’t have a sound statistical analysis plan in place in advance then we are wasting peoples’ time because we are not getting real information out of them
Data is not the same as information and information is what we need because that can be turned into knowledge
Sample size / power analysis
- There are good sample size calculators and bad sample size calculators
- If your sample size calculator asks for the population size then it is a bad one and you need to throw it away immediately
- The population size is not a term in any standard statistical test or approach and is almost always irrelevant
- A good sample size calculator will ask you exactly what numbers you expect to see in your responses and how much you expect them to vary
- It must also consider how many tests you will end up actually doing on the same data
Connecting the dots
People avoid doing prospective power analysis, even though it is a critical step in scientific study, because it seems daunting to ask what results you expect to see in terms of numbers while drawing up a questionnaire about human perceptions of human endeavours.
The solution is simple though: take time when setting up the questionnaire to formally connect every question to its research objective
- You will have to do this anyway at some point
- It’s better to do it before sending out the questionnaire because then you can remove questions that are not relevant and stop wasting participants’ time
- Remember that every unnecessary question you ask reduces the quality of your feedback by inducing fatigue or frustration in respondents
- e.g. if you make a question I don’t want to answer compulsory that does not induce me to ponder longer and overcome my reluctance, instead it induces me to close the survey window (biasing your results)
Details
Suppose you want to know whether there is a difference on average between men and women in terms of their propensity to resign in the face of being denied promotion repeatedly
- You first decide on how to measure it. Suppose we use a 7 step Likert scale question.
- Then you ask yourself what responses you expect to see on this scale and how much you expect those responses to vary from person to person
- Perhaps we expect mostly “Slightly agree” from women and “Neutral” from men, but with two thirds of women varying between “Slightly disagree” and “Strongly Agree” and two thirds of men varying between “Disagree” and “Agree”
- Then we can use a t-test power calculation to approximate how many independent people we need to approach in order to be say 80% sure of detecting a difference at a 5% significance level
Multiple items
- The power curve is for a single item
- With multiple items additional considerations come up:
- Do you want any of them to be significant?
- Do you want all of them to be significant?
- Are the items correlated?
- Are you adjusting for that in your testing?