Tips and tricks to make your data collection easier.
There’s a critical distinction between testing whether your survey makes sense and testing whether it works.
| Cognitive Pre-Test | Technical Pre-Test | |
|---|---|---|
| Question | Do respondents understand what we’re asking? | Does the survey work as programmed? |
| Method | Think-aloud protocols with 5–8 people | Take the survey yourself, every path |
| Catches | Confusing wording, unfamiliar terms, double-barreled questions | Broken skip logic, missing options, display errors |
| When | Before programming (ideally) or right after | After programming, before any respondents |
Before a single real respondent sees your survey:
Logic & Flow
Content & Display
A soft launch means fielding to 5–10% of your target sample before opening the full study. You’re looking at:
Completion rate — If it’s below ~80%, something is driving people out.
Median duration — Compare this to your estimate. If you estimated 8 minutes and actual median is 14, you have a problem.
Drop-off analysis — Where are people abandoning? The question before the biggest drop-off point is usually the culprit.
Open-end quality — Are people writing thoughtful responses or pasting gibberish?
Distribution checks — Are your key variables showing reasonable variance, or is everyone selecting the same option?
Your team soft-launched a 25-question survey to 40 respondents. Here’s what you found:
| Metric | Result |
|---|---|
| Completion rate | 62% |
| Median duration | 18 min (estimated: 10 min) |
| Largest drop-off | Between Q11 and Q12 (lost 22%) |
| Q7 (income) | 35% selected “Prefer not to say” |
| Q19 (open-end: “Describe your ideal experience”) | 60% wrote ≤ 3 words |
With your partner: Diagnose what’s happening and propose fixes. Be specific — don’t just say “make it shorter.”
Every survey is a product. Your respondent is the user. And just like any product, if the experience is bad, people leave — or worse, they stay and give you bad data.
The average American is invited to take dozens of surveys per year. Your survey is competing for attention with everything else on their phone.
This means respondent experience is a data quality strategy.
The single biggest driver of survey abandonment is perceived length. Note the word perceived — it’s not just actual minutes, it’s how long the survey feels.
What the research tells us:
The uncomfortable truth: Every question you add has a cost. Not in dollars (though it does cost more) — in data quality across the entire survey. A 30-question survey with focused, engaged respondents produces better data than a 60-question survey where people are racing to finish.
More than 60% of online survey responses now come from mobile devices. If you haven’t tested your survey on a phone, you haven’t tested your survey.
What breaks on mobile
Mobile-friendly design
How you signal progress profoundly affects respondent behavior.
Progress bars — Useful when accurate. Harmful when they jump unpredictably (especially backwards, which happens with skip logic). If your survey has complex branching, consider hiding the progress bar entirely.
Page structure — One question per page feels slower but produces better data. Many questions per page feels faster but increases satisficing.
Sectioning — “You’re now starting Section 3 of 4” gives respondents a sense of structure and endpoint. This is especially valuable in longer surveys. But it can also be daunting.
The first impression — The opening questions set expectations. If Q1 is a boring demographic question, you’ve told the respondent this survey isn’t worth their attention. Lead with something engaging.
Every survey interaction triggers an internal cost-benefit calculation:
“Is this worth my time?” → Your intro screen and first question
“How much longer is this going to take?” → Your progress indicators
“Why are they asking me this?” → Your question relevance and framing
“I don’t understand this question but I’ll just pick something” → Satisficing from confusion
“I’m done caring” → Fatigue from length or monotony
Your job as a researcher is to keep respondents in the first three states and out of the last two — for the entire survey.
When you’re not recruiting your own sample (the way you are for KCRW), respondents typically come from panel companies — firms that maintain databases of people who’ve agreed to take surveys for compensation.
The key economics: What you pay per completed survey (CPI) is driven by three things:
A general population, 10-minute survey might cost $5–8 per complete. A 20-minute survey targeting C-suite healthcare executives could run $100+.
You will work with these companies in your careers. Understanding their economics helps you design studies that are both rigorous and feasible.
| Phase | What You Do | Common Pitfalls |
|---|---|---|
| Design QA | Cognitive pre-test, technical test all paths | Skipping this because “we’re running out of time” |
| Soft Launch | Field to 5–10%, review diagnostics | Launching full field without checking soft launch data |
| Full Field | Monitor quotas, check quality daily | Set-it-and-forget-it mentality |
| Mid-Field | Communicate with team, adjust if needed | Not flagging fill-rate problems early |
| Close | Final quality checks, close survey, export data | Forgetting to close the survey |
Every shortcut in this process shows up in your data — and by then, it’s too late to fix.
Before you field, make sure you’ve done the following:
This week:
☐ Have at least 2 people outside your team take the survey (cognitive pre-test)
☐ Test every display logic and skip path in Qualtrics
☐ Take the survey on your phone
☐ Time yourself — is the LOI what you expected?
☐ Set up embedded data to track your distribution channels
When you launch:
☐ Soft launch to your first 10–15 respondents
☐ Check completion rate, median duration, and drop-off points
☐ Review open-end response quality
☐ Check your quota fill rates (if using quotas)
☐ Only then open to full distribution