Generate structured research questions, testable hypotheses, and empirical strategies from a topic or dataset
Generate structured research questions, testable hypotheses, and empirical strategies from a topic, phenomenon, or dataset.
Input: $ARGUMENTS — a topic (e.g., "minimum wage effects on employment"), a phenomenon (e.g., "why do firms cluster geographically?"), or a dataset description (e.g., "panel of US counties with pollution and health outcomes, 2000-2020").
Understand the input. Read $ARGUMENTS and any referenced files. Check master_supporting_docs/ for related papers. Check .claude/rules/ for domain conventions.
Generate 3-5 research questions ordered from descriptive to causal:
Tag each RQ with a likely paper type (drawn from methods-referee.md):
reduced-form (DiD, IV, RD, event study, synthetic control)structural (estimation of a fully-specified model)theory+empirics (formal model + empirical test of its predictions)descriptive (measurement, data construction, pattern documentation)formal-theory (pure theory, no empirical test in this paper)survey-experiment (vignette, conjoint, list-experiment)unsure (when multiple types are plausible — the user can pick later via /interview-me)Use .claude/references/discipline-cards.md to bias the distribution by field (econ vs poli-sci default frequencies differ — e.g., poli-sci skews more toward survey-experiment and formal-theory than econ does).
For each research question, develop:
Rank the questions by feasibility and contribution.
Save the output to quality_reports/research_ideation_[sanitized_topic].md
# Research Ideation: [Topic]
**Date:** [YYYY-MM-DD]
**Input:** [Original input]
## Overview
[1-2 paragraphs situating the topic and why it matters]
## Research Questions
### RQ1: [Question] (Feasibility: High/Medium/Low)
**Type:** Descriptive / Correlational / Causal / Mechanism / Policy
**Paper type:** reduced-form / structural / theory+empirics / descriptive / formal-theory / survey-experiment / unsure
**Hypothesis:** [Testable prediction]
**Identification Strategy:**
- **Method:** [e.g., Difference-in-Differences]
- **Treatment:** [What varies and when]
- **Control group:** [Comparison units]
- **Key assumption:** [e.g., Parallel trends]
**Data Requirements:**
- [Dataset 1 — what it provides]
- [Dataset 2 — what it provides]
**Potential Pitfalls:**
1. [Threat 1 and possible mitigation]
2. [Threat 2 and possible mitigation]
**Related Work:** [Author (Year)], [Author (Year)]
---
[Repeat for RQ2-RQ5]
## Ranking
| RQ | Feasibility | Contribution | Priority |
|----|-------------|-------------|----------|
| 1 | High | Medium | ... |
| 2 | Medium | High | ... |
## Suggested Next Steps
1. [Most promising direction and immediate action]
2. [Data to obtain]
3. [Literature to review deeper]
Before returning the ideation report, run the Post-Flight Verification protocol from .claude/rules/post-flight-verification.md. Research ideation is hallucination-prone in three specific ways:
educ_attain" can be confidently wrong about variable names, coverage years, or restricted-access status.educ_attain variable 1990–2024?"claim-verifier via Task with subagent_type=claim-verifier and context=fork. Hand it claims + questions + source pointers (WebSearch allowed, NBER/SSRN URLs preferred, dataset codebooks preferred). Do NOT include the draft.--no-verify flag