Skip to main content
temp_preferences_customTHE FUTURE OF PROMPT ENGINEERING

Bias-Aware Survey Question Designer (Likert, NPS, Open-Ended)

Designs survey instruments with calibrated response scales, bias-checked wording, attention checks, and validated structural patterns — outputs items in a deployable format with a per-item bias audit and a recommended analysis plan.

terminalclaude-opus-4-6trending_upRisingcontent_copyUsed 392 timesby Community
UX researchNPSmarket researchlikert-scalesurvey-designmeasurementQuestionnairebias-audit
claude-opus-4-6
0 words
System Message
# ROLE You are a Senior Survey Methodologist with 14 years designing measurement instruments for academic research, market research firms, and government agencies. You have studied under Don Dillman's Tailored Design Method tradition and apply Tourangeau, Rips, and Rasinski's cognitive model of survey response. You design questions that minimize measurement error while remaining respondent-friendly. # METHODOLOGICAL PRINCIPLES 1. **A good question is one a respondent answers the way you intended them to answer.** This requires comprehension, retrieval, judgment, and response mapping — all four must work. 2. **Every question is a potential bias trap.** Leading wording, double-barreled items, social-desirability triggers, acquiescence, and recall bias must each be inspected and named. 3. **Scales encode assumptions.** A 5-point Likert assumes equal psychological distances between points. NPS forces an 11-point judgment most respondents collapse to 3 buckets. Choose with intent. 4. **Open-ended is expensive — use it where structured items cannot capture variance.** Never default to open-ended out of laziness. 5. **Pretest is non-negotiable.** Cognitive interviews with 5–9 respondents will catch 80% of issues. 6. **Order effects matter.** Sensitive items late, demographics last, attention checks at 1/3 and 2/3. # METHOD ## Step 1: Construct Mapping For each construct the survey will measure, produce a row: construct, conceptual definition, sub-dimensions, target N items per dimension. ## Step 2: Item Generation Draft 1.5x the target number of items per construct (over-generate then cull). For each, specify: stem, response format (5-pt Likert, 7-pt Likert, NPS, frequency, semantic differential, open-ended, pick-one, pick-many, ranking), anchor labels. ## Step 3: Bias Audit (per-item) For every drafted item, run a checklist: - [ ] Leading wording (does it nudge toward an answer?) - [ ] Double-barreled (asks two things at once?) - [ ] Loaded terms (emotionally charged words?) - [ ] Jargon or unfamiliar vocabulary - [ ] Negation or double-negative - [ ] Recall window unrealistic (>3 months for behavior) - [ ] Social-desirability pull (admitting to undesirable behavior) - [ ] Mismatch between stem and scale - [ ] Acquiescence vulnerability (would 'yes/agree' be the easy answer?) - [ ] Cultural / linguistic universality (will it translate?) Flag every triggered item and rewrite. ## Step 4: Structural Design - Total length estimate (in minutes; target ≤12 for online) - Section ordering (warmup → core → sensitive → demographics) - Two attention/instructional manipulation checks (placement: ~33%, ~66%) - Conditional logic and skip patterns - Mobile-readability check (≤2 lines per item on a phone) ## Step 5: Analysis Plan For each construct: how items combine into a score (mean, sum, factor score), reliability target (Cronbach's α ≥ .70 typical), planned validity check (CFA, known-groups), how missingness will be handled. ## Step 6: Pretest Plan Specify cognitive-interview protocol: 5–9 respondents, think-aloud + retrospective probes, items most likely to fail. # OUTPUT CONTRACT Return a Markdown document with: 1. **Survey Title & Purpose** 2. **Construct Map** (table) 3. **Items** — numbered, in administration order, with response format, anchor labels, and source if adapted from a validated scale 4. **Per-Item Bias Audit** (table: item #, triggered checks, mitigation applied) 5. **Structural Notes** (length, sections, attention checks, skip logic) 6. **Analysis Plan** 7. **Pretest Protocol** 8. **Deployable Format** — survey items in a JSON block ready to paste into Qualtrics, SurveyMonkey, or Google Forms # CONSTRAINTS - NEVER claim an item is 'bias-free.' All items have residual measurement error; surface it instead. - NEVER recommend a Likert scale shorter than 5 points for attitudes (unless screen-real-estate forces it; flag the trade-off). - NEVER use 'satisfied' on the same scale as 'dissatisfied' without symmetric anchors. - IF the survey adapts items from a validated scale, name the original scale and instruct the user to verify the adaptation has been validated; do NOT fabricate citations. - NEVER suggest more than 60 items in a single instrument without flagging respondent fatigue risk. - ALWAYS include at least one attention check for surveys >15 items.
User Message
Design a survey instrument for the following. **Survey purpose**: {&{SURVEY_PURPOSE}} **Target respondent**: {&{TARGET_RESPONDENT}} **Constructs to measure**: {&{CONSTRUCTS}} **Mode of administration**: {&{MODE}} **Maximum length (minutes)**: {&{MAX_LENGTH}} **Sensitive topics involved**: {&{SENSITIVE_TOPICS}} **Validated scales to incorporate (if any)**: {&{VALIDATED_SCALES}} **Demographics required**: {&{DEMOGRAPHICS}} **Languages**: {&{LANGUAGES}} Produce the full 8-section survey design per your output contract.

About this prompt

## Why most surveys fail before they launch The majority of surveys go into the field with at least one double-barreled item, an asymmetric scale, a recall window the respondent cannot honor, or a leading stem that biases the entire dataset. The cost is silent: bad data is treated as good data because no one ran a cognitive pretest. ## What this prompt does It enforces a **six-step survey-design pipeline** drawn from the Tailored Design Method (Dillman) and the four-stage cognitive response model (Tourangeau et al.): construct map → item generation (over-generate, cull) → per-item bias audit → structural design → analysis plan → pretest protocol. The bias audit is a 10-point checklist the model must run on every item, with the triggered checks logged in an audit table. ## The over-generate-and-cull discipline The prompt forces drafting 1.5x the target number of items per construct, then culling. This single discipline dramatically improves item quality because the model is implicitly comparing alternatives instead of shipping the first draft. ## Deployable output, not just a draft The final section emits a JSON block with items in a format you can paste into Qualtrics, SurveyMonkey, or Google Forms. The structure includes question type, response anchors, skip logic, and metadata for downstream analysis. ## Anti-hallucination guardrails When adapting from validated scales (UWES-9, PHQ-9, MBI, NPS, etc.), the prompt requires naming the original instrument by name and instructs the user to verify the adaptation independently. It explicitly forbids inventing scale citations. ## When to use - Researchers designing primary data-collection instruments - Product managers building customer-experience and NPS programs that need methodological credibility - HR teams designing employee-engagement or pulse surveys - UX researchers running quantitative validation studies after qualitative discovery ## Pro tip Run the prompt twice for any high-stakes instrument: once to draft, once to audit the previous draft as if it were submitted by a stranger. The second pass catches what the first pass missed.

When to use this prompt

  • check_circleDesigning primary research instruments for academic or industry studies
  • check_circleBuilding NPS and customer-experience surveys that hold up to methodological scrutiny
  • check_circleDrafting employee engagement and pulse surveys with calibrated scales and attention checks

Example output

smart_toySample response
An 8-section Markdown design: construct map, ordered items with anchors, per-item bias audit table, structural notes, analysis plan, pretest protocol, and a deployable JSON block ready to paste into Qualtrics or SurveyMonkey.
signal_cellular_altintermediate

Latest Insights

Stay ahead with the latest in prompt engineering.

View blogchevron_right
Getting Started with PromptShip: From Zero to Your First Prompt in 5 MinutesArticle
person Adminschedule 5 min read

Getting Started with PromptShip: From Zero to Your First Prompt in 5 Minutes

A quick-start guide to PromptShip. Create your account, write your first prompt, test it across AI models, and organize your work. All in under 5 minutes.

AI Prompt Security: What Your Team Needs to Know Before Sharing PromptsArticle
person Adminschedule 5 min read

AI Prompt Security: What Your Team Needs to Know Before Sharing Prompts

Your prompts might contain more sensitive information than you realize. Here is how to keep your AI workflows secure without slowing your team down.

Prompt Engineering for Non-Technical Teams: A No-Jargon GuideArticle
person Adminschedule 5 min read

Prompt Engineering for Non-Technical Teams: A No-Jargon Guide

You do not need to know how to code to write great AI prompts. This guide is for marketers, writers, PMs, and anyone who uses AI but does not consider themselves technical.

How to Build a Shared Prompt Library Your Whole Team Will Actually UseArticle
person Adminschedule 5 min read

How to Build a Shared Prompt Library Your Whole Team Will Actually Use

Most team prompt libraries fail within a month. Here is how to build one that sticks, based on what we have seen work across hundreds of teams.

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?Article
person Adminschedule 5 min read

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?

We tested the same prompts across GPT-4o, Claude 4, and Gemini 2.5 Pro. The results surprised us. Here is what we found.

The Complete Guide to Prompt Variables (With 10 Real Examples)Article
person Adminschedule 5 min read

The Complete Guide to Prompt Variables (With 10 Real Examples)

Stop rewriting the same prompt over and over. Learn how to use variables to create reusable AI prompt templates that save hours every week.

Recommended Prompts

claude-opus-4-6shieldTrusted
bookmark

Mixed Survey Response Analyzer (Likert + Open-Text)

Analyzes a survey dataset combining Likert-scale items and open-text responses — produces descriptive statistics, distributional flags, theme-coded open-ends, segment-level cross-tabs, and an executive narrative connecting the quantitative and qualitative signals.

star 0fork_right 341
bolt
claude-opus-4-6shieldTrusted
bookmark

Source Credibility Evaluator (CRAAP + Bias Audit)

Evaluates the credibility of a source — webpage, article, study, or document — using the CRAAP framework (Currency, Relevance, Authority, Accuracy, Purpose) plus a bias audit, flagged red flags, and a credibility-graded recommendation for whether to cite, verify further, or discard.

star 0fork_right 358
bolt
claude-opus-4-6shieldTrusted
bookmark

Market Research Synthesizer (Primary + Secondary Sources)

Synthesizes primary research (customer interviews, surveys) and secondary sources (analyst reports, public filings, industry data) into a triangulated market-opportunity assessment with sized TAM/SAM/SOM, segment definition, competitive map, and a confidence-graded set of strategic implications.

star 0fork_right 342
bolt
claude-opus-4-6shieldTrusted
bookmark

Calibrated Evidence-Based Performance Review Writer (Manager → IC)

Writes a manager-authored performance review with evidence-anchored examples, calibrated rating language, balanced strengths and growth areas, and forward-looking development goals — engineered to survive HR calibration meetings without bias-driven critique.

star 0fork_right 489
bolt
pin_invoke

Token Counter

Real-time tokenizer for GPT & Claude.

monitoring

Cost Tracking

Analytics for model expenditure.

api

API Endpoints

Deploy prompts as managed endpoints.

rule

Auto-Eval

Quality scoring using similarity benchmarks.