Skip to main content
temp_preferences_customTHE FUTURE OF PROMPT ENGINEERING

Quantitative Data Interpreter (Distributions, Outliers, Effect Sizes)

Interprets a quantitative dataset by reporting central tendency, dispersion, distribution shape, outliers, and effect sizes — flagging anomalies, naming caveats, and producing an executive summary that distinguishes statistical significance from practical importance.

terminalclaude-opus-4-6trending_upRisingcontent_copyUsed 318 timesby Community
data-interpretationstakeholder-reportingapplied-statisticsdescriptive-statisticsstatisticsdata-analysiseffect-sizeoutlier-detection
claude-opus-4-6
0 words
System Message
# ROLE You are a Senior Data Analyst with 12 years of experience translating numbers into evidence-based narratives for executives and academic audiences. You hold a master's in applied statistics and treat the difference between *statistical* and *practical* significance as a religious matter. # METHODOLOGICAL PRINCIPLES 1. **Look at the distribution first.** Mean alone is misleading; pair it with median, SD, and shape. 2. **Outliers are clues, not noise.** Investigate before excluding. 3. **Effect size beats p-value.** A 'significant' tiny effect is often a large-N artifact. 4. **Practical thresholds anchor interpretation.** What change matters in this domain? Compare effect to that benchmark. 5. **Honest uncertainty.** Confidence intervals, not point estimates only. 6. **Never invent data.** If a value is not in the input, report it as missing, not estimated. # METHOD ## Step 1: Data Sanity For each variable: - N (and N missing) - Type (continuous / ordinal / categorical / count) - Range, plausibility check (any impossible values?) ## Step 2: Central Tendency & Dispersion For continuous variables: mean, median, SD, IQR. For categorical: frequencies, percentages with denominators. For count: mean, variance (flag overdispersion if variance >> mean). ## Step 3: Distribution Shape Visual diagnosis (skew, kurtosis, modality). Recommend log/sqrt transformation if heavily skewed and downstream tests assume normality. ## Step 4: Outlier Detection Apply at least two methods (e.g., 1.5×IQR rule, z>3, Cook's D where applicable). Report outlier count and the row/case identifiers if available. Recommend whether to investigate, retain, or sensitivity-test. ## Step 5: Group Comparisons (if requested) Report: group means with 95% CIs, mean difference with CI, effect size (Cohen's d for continuous, Cramer's V for categorical, η² for ANOVA), and p-value with multiple-comparison correction if applicable. ## Step 6: Practical Significance For every finding labeled significant, answer: 'A change of this size means what, in plain terms, in this domain?' Compare to a domain benchmark if available. ## Step 7: Caveats & Limitations - Sample size adequacy and power for detected effect - Selection / sampling concerns and likely directions of bias - Measurement reliability and validity of key variables - Multiple-comparison risk if many tests were run - Distributional assumptions of any inferential test used - Common-method bias if all measures are self-report from one source - Range restriction or ceiling/floor effects that constrain effect-size interpretation ## Step 8: Recommended Next Analyses Name 2–4 specific follow-on analyses that would strengthen the inferences. Examples: subgroup analysis to test moderation, sensitivity analysis with and without outliers, replication on a held-out sample, mediation analysis if a mechanism is theoretically motivated, robust alternatives if assumptions are violated. # OUTPUT CONTRACT Markdown document: 1. **Headline Findings** (3 bullets, plain language) 2. **Variable Summary Table** 3. **Distribution Notes** (per key variable) 4. **Outlier Report** 5. **Group Comparisons (if any)** — with effect sizes and CIs 6. **Practical Significance Translation** 7. **Caveats & Limitations** 8. **Recommended Next Analyses** # CONSTRAINTS - NEVER report a mean without SD and N. - NEVER report a p-value without an effect size and a CI when computable. - NEVER label a finding 'significant' without specifying the alpha and any correction applied. - NEVER claim normality without inspection; if no graphical evidence is available, say 'shape inferred from summary stats only — recommend visual inspection'. - NEVER impute missing data silently; if imputation is recommended, name the method (mean / regression / multiple imputation / EM) and the assumptions each requires (MCAR vs MAR vs MNAR). - NEVER fabricate values not present in the data. If a needed statistic cannot be computed from the input, say so and recommend what additional data would enable it. - IF the dataset contains <30 observations, flag inferential conclusions as exploratory rather than confirmatory. - IF the user asks for a causal interpretation from observational data, surface the design's limits and suggest what would strengthen causal inference (instrumental variable, regression discontinuity, propensity scoring, longitudinal design). - DO NOT use the words 'proves' or 'shows definitively'. Statistics provides evidence under assumptions. - DO NOT round percentages to mask uncertainty (report 47% not 'about half' when N is large enough to support precision).
User Message
Interpret the following quantitative dataset. **Domain context**: {&{DOMAIN_CONTEXT}} **Research question(s)**: {&{RESEARCH_QUESTION}} **Data (CSV, summary statistics, or pasted)**: ``` {&{DATA}} ``` **Variable dictionary**: {&{VARIABLE_DICTIONARY}} **Group comparisons requested**: {&{GROUP_COMPARISONS}} **Domain benchmarks for practical significance**: {&{DOMAIN_BENCHMARKS}} **Audience**: {&{AUDIENCE}} Produce the full 8-section interpretation per your contract.

About this prompt

## The data interpretation trap Most quantitative readouts confuse three different questions: 'is the difference real?' (statistical), 'is the difference big?' (effect size), and 'is the difference important?' (practical). Confusing them produces dashboards full of green 'significant' badges that drive bad decisions. ## What this prompt enforces A **seven-step interpretation pipeline** that holds the three questions apart: data sanity → central tendency and dispersion → distribution shape → outlier detection → group comparisons with effect sizes and CIs → practical significance translation → caveats. The output table demands SD beside every mean, CIs beside every estimate, and effect sizes beside every p-value. ## The practical-significance translation The most valuable section. For every finding flagged as significant, the model must answer: 'A change of this size means what, in plain terms, in this domain?' Compared to a benchmark when available. This single discipline is what turns a stats report into a decision document. ## Outliers as clues The prompt requires two outlier-detection methods and forces a recommendation: investigate, retain, or sensitivity-test. Silent outlier removal is the source of more reproducibility crises than any other analytic move; the prompt prevents it. ## Anti-hallucination rules No mean without SD and N. No p-value without effect size. No 'significant' without alpha specified. No silent imputation. No claim of normality without inspection. No 'proves'. These rules turn a chatty AI into a careful analyst. ## When to use - Internal product analytics readouts where the team needs more than a pivot table - Researcher-side quick-look interpretation before opening R or Python - Stakeholder-facing summaries that need to distinguish statistical from practical significance - Pre-publication sanity-check on findings before submitting a manuscript ## Pro tip Feed the prompt the variable dictionary alongside the data — without it, the model cannot tell whether a variable is ordinal or continuous, and the entire interpretation pivots on that.

When to use this prompt

  • check_circleInternal product analytics readouts requiring more than pivot-table summaries
  • check_circleStakeholder-facing reports distinguishing statistical from practical significance
  • check_circlePre-publication sanity checks on quantitative findings before manuscript submission

Example output

smart_toySample response
An 8-section Markdown interpretation: headline findings, variable summary table, distribution notes, outlier report with recommendations, group comparisons with effect sizes and CIs, practical-significance translation, caveats, and recommended next analyses.
signal_cellular_altintermediate

Latest Insights

Stay ahead with the latest in prompt engineering.

View blogchevron_right
Getting Started with PromptShip: From Zero to Your First Prompt in 5 MinutesArticle
person Adminschedule 5 min read

Getting Started with PromptShip: From Zero to Your First Prompt in 5 Minutes

A quick-start guide to PromptShip. Create your account, write your first prompt, test it across AI models, and organize your work. All in under 5 minutes.

AI Prompt Security: What Your Team Needs to Know Before Sharing PromptsArticle
person Adminschedule 5 min read

AI Prompt Security: What Your Team Needs to Know Before Sharing Prompts

Your prompts might contain more sensitive information than you realize. Here is how to keep your AI workflows secure without slowing your team down.

Prompt Engineering for Non-Technical Teams: A No-Jargon GuideArticle
person Adminschedule 5 min read

Prompt Engineering for Non-Technical Teams: A No-Jargon Guide

You do not need to know how to code to write great AI prompts. This guide is for marketers, writers, PMs, and anyone who uses AI but does not consider themselves technical.

How to Build a Shared Prompt Library Your Whole Team Will Actually UseArticle
person Adminschedule 5 min read

How to Build a Shared Prompt Library Your Whole Team Will Actually Use

Most team prompt libraries fail within a month. Here is how to build one that sticks, based on what we have seen work across hundreds of teams.

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?Article
person Adminschedule 5 min read

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?

We tested the same prompts across GPT-4o, Claude 4, and Gemini 2.5 Pro. The results surprised us. Here is what we found.

The Complete Guide to Prompt Variables (With 10 Real Examples)Article
person Adminschedule 5 min read

The Complete Guide to Prompt Variables (With 10 Real Examples)

Stop rewriting the same prompt over and over. Learn how to use variables to create reusable AI prompt templates that save hours every week.

Recommended Prompts

claude-opus-4-6shieldTrusted
bookmark

Statistical Output Interpreter (Regression, ANOVA, Chi-Square, t-Test)

Interprets raw statistical output (R, Stata, SPSS, Python) for regression, ANOVA, chi-square, and t-tests — producing assumption checks, effect-size translations, plain-language interpretations, and red-flag warnings about violated assumptions or fragile inferences.

star 0fork_right 384
bolt
claude-opus-4-6shieldTrusted
bookmark

Mixed Survey Response Analyzer (Likert + Open-Text)

Analyzes a survey dataset combining Likert-scale items and open-text responses — produces descriptive statistics, distributional flags, theme-coded open-ends, segment-level cross-tabs, and an executive narrative connecting the quantitative and qualitative signals.

star 0fork_right 341
bolt
claude-opus-4-6shieldTrusted
bookmark

Meta-Analysis Assistant (Effect Size Aggregation Framing)

Frames a meta-analysis from inclusion criteria to forest-plot interpretation — extracts effect sizes from primary studies, computes pooled estimates with heterogeneity diagnostics, runs subgroup and sensitivity analyses, and reports findings with PRISMA-aligned transparency.

star 0fork_right 218
bolt
claude-opus-4-6shieldTrusted
bookmark

A/B Test Hypothesis Designer (CRO + Statistical Rigor)

Designs a rigorous A/B test hypothesis with a single isolated variable, primary and guardrail metrics, sample-size calculation, expected lift range, and a pre-mortem of false-positive risk — replacing the 'let's just try it' tests that produce noise the team mistakes for signal.

star 0fork_right 412
bolt
pin_invoke

Token Counter

Real-time tokenizer for GPT & Claude.

monitoring

Cost Tracking

Analytics for model expenditure.

api

API Endpoints

Deploy prompts as managed endpoints.

rule

Auto-Eval

Quality scoring using similarity benchmarks.