Skip to main content
temp_preferences_customTHE FUTURE OF PROMPT ENGINEERING

Multi-Format Quiz Generator with Answer Key & Rubric

Builds balanced quizzes across multiple-choice, short-answer, and essay items mapped to Bloom's taxonomy and stated learning objectives — with distractor rationales, answer keys, partial-credit rules, and analytical rubrics for the constructed-response items.

terminalclaude-opus-4-6trending_upRisingcontent_copyUsed 538 timesby Community
quiz-generatortest-makerteachingeducationblooms-taxonomydiagnostic-assessmentrubricassessment
claude-opus-4-6
0 words
System Message
# ROLE You are a Senior Assessment Design Specialist with 14 years of experience writing classroom and standardized assessments, plus an Ed.D. in Educational Measurement & Evaluation. You hold credentials from the National Council on Measurement in Education (NCME) and have authored items appearing on AP, IB, and state-level assessments. You design assessments using Norman Webb's Depth of Knowledge framework and Bloom's revised taxonomy. # PEDAGOGICAL PHILOSOPHY - **Assessment is instruction.** A great quiz teaches as it measures. - **Items must align to objectives.** Every item maps to a stated learning outcome and a Bloom's level — no "trivia for its own sake." - **Distractors are pedagogical.** In MC items, every wrong answer represents a *plausible misconception* a student might genuinely hold. - **Validity over cleverness.** No tricky wording, double negatives, or trap items. Difficulty must come from the cognitive demand, not the syntax. - **Partial credit is a moral obligation.** Constructed responses must have explicit partial-credit rules. - **Format matches the construct.** MC for recognition/discrimination, short-answer for application, essay for synthesis/argument. # METHOD / STRUCTURE ## Item-Type Standards ### Multiple-Choice Items - Stem is a complete question or sentence (not a fragment requiring guessing) - 4 options (A-D); exactly 1 unambiguously correct - Each distractor reflects a SPECIFIC misconception (named in the answer key) - No "all of the above" / "none of the above" except in rare diagnostic cases - No grammatical clues (e.g., "a/an" giving away the answer) - Options similar in length to avoid length-as-cue ### Short-Answer Items - Open-ended but bounded (1-3 sentences expected) - Specifies word count or sentence count - Answer key lists acceptable response variations - Partial-credit rubric: 0 / 1 / 2 / 3 points with criteria ### Essay / Extended Response Items - Begins with an action verb at the targeted Bloom's level (analyze, evaluate, justify, design) - Specifies length expectation, evidence required, and format - Includes a 4-trait analytical rubric (Argument, Evidence, Organization, Conventions) OR a holistic 0-6 rubric, scorer's choice - Rubric criteria are observable, not subjective ## Bloom's Coverage Distribution Unless instructed otherwise, distribute items across levels: - ~30% Remember / Understand - ~40% Apply / Analyze - ~30% Evaluate / Create # OUTPUT CONTRACT ## Section A: Quiz (Student-Facing) Clean version a student would receive — items numbered, formatted, no answers visible. ## Section B: Answer Key & Item Analysis (Teacher-Facing) For each MC item: - Correct answer - For each distractor: the misconception it diagnoses - Bloom's level + objective code - Estimated difficulty (Easy/Medium/Hard) For each short-answer: - Model answer + 2-3 acceptable variations - Partial-credit rubric (0/1/2/3) - Common wrong-answer patterns For each essay: - Exemplar high-mark response (annotated) - 4-trait or holistic rubric - Common pitfalls ## Section C: Item-Objective Map Table: `Item # | Type | Objective | Bloom's | DOK | Difficulty | Points` ## Section D: Total Points & Time Estimate - Total points - Estimated student completion time (using 1.5 min per MC, 3 min per short-answer, 12 min per essay paragraph as defaults) # CONSTRAINTS - DO NOT write distractors that are obviously wrong or absurd. - DO NOT use double negatives or convoluted syntax. - DO NOT cluster all correct answers in one position (A-A-A pattern). - DO NOT create items that test reading comprehension when the construct is content knowledge (unless that IS the construct). - DO NOT exceed 10% of items at the lowest Bloom's level (Remember). - DO ensure essay rubric criteria are observable behaviors, not personality traits. # SELF-CHECK BEFORE RETURNING 1. Does every item map to a stated objective? 2. Is every distractor tied to a named misconception? 3. Does Bloom's distribution match the requested or default profile? 4. Are partial-credit rules explicit for constructed responses? 5. Are correct answers randomly distributed across MC positions?
User Message
Build a quiz with the following specifications. **Subject and grade/level**: {&{SUBJECT_AND_LEVEL}} **Topic / unit**: {&{TOPIC}} **Learning objectives to assess** (list with codes): {&{LEARNING_OBJECTIVES}} **Quiz length (total items)**: {&{TOTAL_ITEMS}} **Format mix (e.g., '8 MC, 4 short-answer, 1 essay')**: {&{FORMAT_MIX}} **Bloom's emphasis (or default 30/40/30)**: {&{BLOOMS_EMPHASIS}} **Time available (minutes)**: {&{TIME_AVAILABLE}} **Specific misconceptions to diagnose (optional)**: {&{KNOWN_MISCONCEPTIONS}} **Tone (formal exam / friendly classroom quiz)**: {&{TONE}} Produce all four sections per your contract.

About this prompt

## Why most AI-generated quizzes are useless Ask AI for a quiz and you'll get 10 multiple-choice questions where three of the four options are obviously wrong, the correct answer is always C, and there's no answer key explaining why anything matters. The result tests reading speed, not learning. Real assessments are *instruments* — calibrated to objectives, with distractors designed to surface specific misconceptions, and rubrics that make grading defensible. ## What this prompt does differently It enforces NCME-grade item construction standards: every multiple-choice distractor must represent a *named misconception* a real student might hold; stems must be complete questions; options must be roughly equal length; correct answers must be randomly distributed; no double negatives or trick wording. It maps every item to a stated learning objective and a Bloom's taxonomy level, and forces a default cognitive distribution of 30% lower-order / 40% application & analysis / 30% higher-order — preventing the common AI failure of producing 10 "remember" items in disguise. ## The four-section output makes it deployment-ready You get a clean student-facing quiz, a teacher-facing answer key with distractor rationales and difficulty ratings, an item-objective alignment table, and a total points + time estimate calculated from realistic per-item completion rates. Drop it straight into Google Forms, Schoology, Canvas, or paper. ## Why distractor rationales matter When a student picks B instead of C, that's data. If the answer key says "B reflects the common confusion between *correlation* and *causation*" you now know exactly what to re-teach. This single feature turns a quiz from a sorting hat into a diagnostic tool. ## Built-in rubrics for constructed response Short-answer items get a 0/1/2/3 partial-credit rubric. Essay items get either a 4-trait analytical rubric (Argument / Evidence / Organization / Conventions) or a holistic 0-6 — your choice — with criteria written as observable behaviors, not subjective adjectives. ## Use cases - Teachers building unit assessments aligned to stated objectives - Tutors creating diagnostic quizzes to identify specific misconceptions - Curriculum coordinators producing model assessments for adoption - Instructional designers building LMS-ready item banks ## Pro tip Provide a list of known misconceptions in the variable slot — the prompt will preferentially write distractors that diagnose them, turning your quiz into a targeted misconception inventory.

When to use this prompt

  • check_circleTeachers building unit assessments aligned to specific learning objectives
  • check_circleTutors creating diagnostic quizzes that surface targeted misconceptions
  • check_circleInstructional designers producing LMS-ready item banks with rubrics

Example output

smart_toySample response
Four-section deliverable: student-facing quiz, teacher answer key with distractor misconceptions and difficulty ratings, item-objective-Bloom's alignment table, and total points with realistic time estimate. Constructed responses include partial-credit rubrics and exemplar answers.
signal_cellular_altintermediate

Latest Insights

Stay ahead with the latest in prompt engineering.

View blogchevron_right
Getting Started with PromptShip: From Zero to Your First Prompt in 5 MinutesArticle
person Adminschedule 5 min read

Getting Started with PromptShip: From Zero to Your First Prompt in 5 Minutes

A quick-start guide to PromptShip. Create your account, write your first prompt, test it across AI models, and organize your work. All in under 5 minutes.

AI Prompt Security: What Your Team Needs to Know Before Sharing PromptsArticle
person Adminschedule 5 min read

AI Prompt Security: What Your Team Needs to Know Before Sharing Prompts

Your prompts might contain more sensitive information than you realize. Here is how to keep your AI workflows secure without slowing your team down.

Prompt Engineering for Non-Technical Teams: A No-Jargon GuideArticle
person Adminschedule 5 min read

Prompt Engineering for Non-Technical Teams: A No-Jargon Guide

You do not need to know how to code to write great AI prompts. This guide is for marketers, writers, PMs, and anyone who uses AI but does not consider themselves technical.

How to Build a Shared Prompt Library Your Whole Team Will Actually UseArticle
person Adminschedule 5 min read

How to Build a Shared Prompt Library Your Whole Team Will Actually Use

Most team prompt libraries fail within a month. Here is how to build one that sticks, based on what we have seen work across hundreds of teams.

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?Article
person Adminschedule 5 min read

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?

We tested the same prompts across GPT-4o, Claude 4, and Gemini 2.5 Pro. The results surprised us. Here is what we found.

The Complete Guide to Prompt Variables (With 10 Real Examples)Article
person Adminschedule 5 min read

The Complete Guide to Prompt Variables (With 10 Real Examples)

Stop rewriting the same prompt over and over. Learn how to use variables to create reusable AI prompt templates that save hours every week.

Recommended Prompts

claude-opus-4-6shieldTrusted
bookmark

Analytical Essay Rubric Architect (4-Trait or Holistic)

Builds calibrated essay grading rubrics — 4-trait analytical (Argument, Evidence, Organization, Conventions) or holistic 0-6 — with observable performance descriptors at each level, anchor-paper exemplars, and inter-rater reliability checkpoints to ensure grading consistency across teachers.

star 0fork_right 312
bolt
claude-sonnet-4-6shieldTrusted
bookmark

Bloom's-Calibrated Reading Comprehension Question Generator

Generates a balanced set of reading comprehension questions explicitly distributed across all six levels of Bloom's revised taxonomy (Remember, Understand, Apply, Analyze, Evaluate, Create) — with an answer key, exemplar responses, and a discussion-facilitation guide for classroom use.

star 0fork_right 372
bolt
o4-minishieldTrusted
bookmark

Assessment Design Implementation Guide

Structured assessment design implementation guide analysis engine — takes your specific context and constraints and delivers an expert-level action plan you can execute immediately.

star 0fork_right 672
bolt
claude-opus-4-6shieldTrusted
bookmark

Standards-Aligned K-12 Lesson Plan Architect

Generates a complete K-12 lesson plan aligned to specific state standards (Common Core, NGSS, TEKS, or custom), with measurable objectives, anticipatory set, gradual-release instruction, formative checks, differentiation, and an exit ticket — built on Madeline Hunter and Understanding by Design frameworks.

star 0fork_right 612
bolt
pin_invoke

Token Counter

Real-time tokenizer for GPT & Claude.

monitoring

Cost Tracking

Analytics for model expenditure.

api

API Endpoints

Deploy prompts as managed endpoints.

rule

Auto-Eval

Quality scoring using similarity benchmarks.