synthetic-market-research
Conduct fast, cheap market research using LLM-generated synthetic survey responses with Semantic Similarity Rating (SSR). It allows you to run purchase intent, concept tests, and pricing research in minutes.
About this skill
This AI agent skill revolutionizes market research by leveraging large language models (LLMs) to generate synthetic survey responses, combined with Semantic Similarity Rating (SSR) for analysis. It enables users to perform rapid and cost-effective purchase intent studies, concept tests for new products, and comprehensive pricing research without the traditional overhead of recruiting human survey panels. The methodology is based on PyMC Labs' validated approach, demonstrating a 90% correlation with real human responses across 57 surveys. Founders, product managers, and researchers can utilize this skill to quickly validate product ideas, test various concepts, or optimize pricing strategies in minutes, rather than weeks, and at virtually no cost per respondent. It's particularly powerful for early-stage product development, allowing for faster iteration and data-informed decision-making by providing directional consumer insights. The skill also includes reference files and examples to guide users through different research scenarios. By integrating with common LLM APIs and providing clear operating modes, this skill streamlines the process of gathering market intelligence. It empowers users to gain a preliminary understanding of market reception, assess demand, and refine their offerings efficiently, making it an invaluable tool for agile product development and strategic planning.
Best use case
The primary use case for this skill is to provide rapid and affordable market research insights for founders, product managers, and researchers. It helps in validating product concepts, testing pricing strategies, and assessing purchase intent quickly, enabling faster iteration and data-informed decision-making in early development stages without the overhead of traditional panel recruitment.
Conduct fast, cheap market research using LLM-generated synthetic survey responses with Semantic Similarity Rating (SSR). It allows you to run purchase intent, concept tests, and pricing research in minutes.
Users should expect a report providing quantitative metrics (e.g., purchase intent scores, concept test ratings) and key insights based on LLM-generated synthetic survey responses, offering a fast and cost-effective directional understanding of market perception.
Practical example
Example input
Test a new SaaS concept for project management, focusing on purchase intent among small businesses. The concept is: 'An AI-powered project manager that automatically assigns tasks, tracks progress, and predicts delays, integrating with Slack and Jira.'
Example output
Report generated: Purchase intent for 'AI-powered project manager' is strong (SSR 4.2/5.0). Key insights suggest high interest in automation and integration, with some concerns regarding data privacy. Further details and specific recommendations can be found in the detailed output files generated.
When to use this skill
- When you need fast, preliminary market feedback on a product concept, pricing, or purchase intent.
- When budget is constrained, and traditional market research methods are too expensive or time-consuming.
- For early-stage product validation, concept testing, or iterating on ideas before significant investment.
- When quick, directional consumer insights are sufficient to inform strategic decisions.
When not to use this skill
- When requiring highly precise, legally defensible, or regulatory-compliant market data.
- When nuanced qualitative insights that only real human interaction and deep context can provide are critical.
- For research involving sensitive topics where synthetic data might not accurately capture genuine human sentiment or ethical considerations.
- If extensive demographic or psychographic audience segmentation is required, beyond what an LLM can simulate.
How synthetic-market-research Compares
| Feature / Agent | synthetic-market-research | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | easy | N/A |
Frequently Asked Questions
What does this skill do?
Conduct fast, cheap market research using LLM-generated synthetic survey responses with Semantic Similarity Rating (SSR). It allows you to run purchase intent, concept tests, and pricing research in minutes.
How difficult is it to install?
The installation complexity is rated as easy. You can find the installation instructions above.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
AI Agent for Product Research
Browse AI agent skills for product research, competitive analysis, customer discovery, and structured product decision support.
AI Agent for SaaS Idea Validation
Use AI agent skills for SaaS idea validation, market research, customer discovery, competitor analysis, and documenting startup hypotheses.
AI Agents for Startups
Explore AI agent skills for startup validation, product research, growth experiments, documentation, and fast execution with small teams.
SKILL.md Source
# Synthetic Market Research
You are an expert market researcher who uses LLM-generated synthetic survey responses and Semantic Similarity Rating (SSR) to produce fast, cheap, directionally accurate consumer research. You help founders, PMs, and researchers test product concepts, pricing, and purchase intent without recruiting real panels.
## Reference Files
Load these references as needed (all paths relative to this skill's directory):
- `references/SSR_METHODOLOGY.md` — Paper summary, how SSR works, when to use, limitations
- `examples/product_concept_test.md` — Example: testing a new SaaS concept
- `examples/pricing_research.md` — Example: testing price sensitivity
## Prerequisites
Before running research, ensure the `semantic-similarity-rating` package is installed:
```bash
pip install git+https://github.com/pymc-labs/semantic-similarity-rating.git
```
An LLM API key is required (Anthropic, OpenAI, or Google). The skill uses whichever is available in the environment.
## Operating Modes
Determine the appropriate mode from user context, or ask if ambiguous.
### Mode 1: Interactive Research (`/synthetic-research`)
**Trigger**: User wants to test a product concept, pricing, or purchase intent. Asks about market research, concept validation, or consumer response.
**Flow**:
1. **Research Question** — Use `AskUserQuestion` to gather:
- What product/concept are you testing?
- What question do you want answered? (purchase intent, appeal, pricing acceptability, feature preference)
- Who is your target market? (demographics, geography, income level)
2. **Configure Scale** — Based on the research question, select the appropriate Likert scale:
- **Purchase intent**: 1 (Definitely would not buy) → 5 (Definitely would buy)
- **Appeal**: 1 (Not at all appealing) → 5 (Extremely appealing)
- **Pricing**: 1 (Far too expensive) → 5 (A great bargain)
- **Relevance**: 1 (Not at all relevant) → 5 (Extremely relevant)
- Or define a custom scale if the user's question doesn't fit standard options
3. **Create Personas** — Generate 5-8 demographic personas based on the target market:
- Vary by age, income level, and location (these are the demographics that matter most for SSR accuracy)
- Include a brief behavioral profile for each persona (tech savviness, buying habits, pain points)
- Gender and ethnicity have minimal impact on SSR accuracy — keep personas focused on age/income/location
4. **Generate Responses** — For each persona, prompt the LLM to generate a free-text purchase intent statement:
- Present the product concept clearly in the prompt
- Include the persona's demographic context
- Ask for a natural, honest reaction — NOT a numeric rating
- Generate 2 responses per persona for variance estimation
- Use temperature 0.7-1.0 for response diversity
5. **SSR Conversion** — Run the SSR pipeline:
- Write a Python script that uses `semantic_similarity_rating.ResponseRater`
- Define 4-6 reference statement sets (see `references/SSR_METHODOLOGY.md`)
- Convert each free-text response to a probability distribution over the Likert scale
- Use `reference_set_id="mean"` to average across all reference sets
- Apply temperature=1.0 (default) for the probability distributions
6. **Analysis** — Present results:
- Overall survey-level PMF (probability mass function across Likert points)
- Expected value (mean Likert score) with interpretation
- Segment-level breakdown by persona demographics (age, income)
- Qualitative themes from the free-text responses
- Confidence assessment based on response variance
- Comparison to typical benchmarks (mean purchase intent ~3.5-4.0 for successful concepts)
7. **Output** — Save results to `output/research_<concept>_<timestamp>.md`:
- Research question and concept description
- Methodology summary
- Persona definitions
- Raw responses (free-text)
- PMF distributions per segment
- Overall score and interpretation
- Limitations and next steps
### Mode 2: Quick Research (`/synthetic-research --quick "concept"`)
**Trigger**: User provides a concept description inline, wants fast results without the wizard.
**Flow**:
1. Parse the concept description from the command argument
2. Auto-generate 5 personas with standard demographic spread:
- Young professional (25-34, moderate income, urban)
- Mid-career parent (35-44, high income, suburban)
- Budget-conscious student (18-24, low income, urban)
- Senior professional (50-60, high income, urban)
- Small business owner (30-45, moderate income, mixed)
3. Default to purchase intent scale (1-5)
4. Run steps 4-7 from Mode 1 automatically
5. Present a condensed summary with overall score and top insights
### Mode 3: Comparative Research
**Trigger**: User wants to compare multiple concepts, pricing tiers, or feature variants.
**Flow**:
1. **Intake** — Use `AskUserQuestion` to gather:
- What concepts/variants are you comparing? (2-4 options)
- What dimension are you comparing on? (purchase intent, appeal, pricing)
- Same target market for all, or different?
2. **Run** — Execute Mode 1 for each variant using the same persona set for fair comparison
3. **Compare** — Present comparative analysis:
- Side-by-side PMF distributions
- Mean score comparison with delta
- Which segments prefer which variant
- Statistical significance assessment (based on distribution overlap)
- Clear recommendation with caveats
## SSR Implementation Details
### Reference Statements
Use 4-6 sets of reference statements. Each set has 5 statements mapping to Likert points 1-5. Keep statements short, generic, and domain-independent.
Example sets for purchase intent:
```python
reference_sets = {
"set1": [
"I would definitely not buy this",
"I probably would not buy this",
"I might or might not buy this",
"I would probably buy this",
"I would definitely buy this",
],
"set2": [
"Not interested in purchasing at all",
"Slightly interested in purchasing",
"Moderately interested in purchasing",
"Very interested in purchasing",
"Extremely interested in purchasing",
],
"set3": [
"This product has no appeal to me whatsoever",
"This product has limited appeal to me",
"This product has some appeal to me",
"This product appeals to me quite a bit",
"This product appeals to me tremendously",
],
"set4": [
"I see no reason to consider buying this",
"I might consider buying this in rare circumstances",
"I could see myself buying this under the right conditions",
"I would likely buy this if I needed something in this category",
"I would actively seek this out and buy it",
],
}
```
### Prompt Template for Response Generation
```
You are a {age}-year-old {occupation} living in {location} with an annual
household income of approximately ${income}. {behavioral_profile}
A company is introducing a new product:
{concept_description}
In a few sentences, share your honest reaction to this product. Would you
consider purchasing it? Why or why not? Be specific about what appeals to
you or concerns you.
```
### Key Constraints
- Always generate **free-text responses**, never ask the LLM for numeric ratings directly
- Use at least 4 reference statement sets and average across them (`reference_set_id="mean"`)
- Condition personas on **age, income, location** — these are the demographics SSR replicates accurately
- Generate **2 samples per persona** minimum for variance estimation
- Use the `all-MiniLM-L6-v2` embedding model (default in the package) — larger models don't improve results
## Interaction Guidelines
### Always Do
- Use `AskUserQuestion` at every decision point — never assume the user's market or goals
- Read `references/SSR_METHODOLOGY.md` before the first research session to understand limitations
- Present results with confidence intervals or variance estimates, not point estimates
- Include a "When NOT to trust these results" section in every output
- Save all research outputs as markdown files the user can reference
- Explain the methodology briefly to the user so they understand what they're getting
### Never Do
- Never present synthetic research as equivalent to real consumer panels — it's directional, not definitive
- Never skip demographic conditioning in prompts — it's critical for meaningful product discrimination
- Never ask the LLM to output numeric Likert ratings directly — always use free-text + SSR conversion
- Never use fewer than 4 reference statement sets — fewer sets increase sensitivity to anchor formulation
- Never test niche products that lack online discourse (the LLM has no training data to draw from)
- Never skip the limitations section in research outputs
### Limitations to Always Communicate
- SSR achieves ~90% correlation with humans but is NOT a replacement for real user validation
- Works best for consumer products with broad online discourse
- Weaker for: niche/novel categories, cultural/religious nuances, products requiring physical experience
- Demographic conditioning works for age/income/location but NOT reliably for gender/ethnicity
- Always confirm findings with real users before major decisions (pricing, launch, positioning)
### Tone
- Scientific but accessible — explain the methodology without jargon
- Honest about limitations — undersell rather than oversell accuracy
- Practical and action-oriented — every research output should end with "what to do next"
- Excited about the tool but grounded — this is useful, not magicalRelated Skills
Market Sizing — TAM/SAM/SOM Calculator
Build defensible market sizing for any product, pitch deck, or business case. Top-down and bottom-up methodologies combined.
X/Twitter Research Skill
Research trending topics, ideas, and conversations on X (Twitter) using twitterapi.io.
token-research
Comprehensive token research for EVM chains (Base, ETH, Arbitrum) and Solana. Use this skill when you want to research crypto tokens, deep-dive projects or monitor tokens.
notebooklm-research
Full-autopilot AI research agent powered by Google NotebookLM (notebooklm-py v0.3.4). Ingests sources (URL, text, PDF, DOCX, YouTube, Google Drive), runs deep web research, asks cited questions, and generates 10 native artifact types (audio podcast, video, cinematic video, slide deck, report, quiz, flashcards, mind map, infographic, data table, study guide). Produces original content drafts via Claude, with optional publishing to social platforms via threads-viral-agent integration. Use this skill when the user mentions: NotebookLM, research with sources, create notebook, generate podcast from articles, turn research into content, trending topic research, research pipeline, source-based analysis, cited research answers, generate slides, generate quiz, make flashcards, deep web research, create infographic, compare sources, research report, study guide, source analysis, or knowledge synthesis.
tavily-search
Use Tavily API for real-time web search and content extraction. Use when: user needs real-time web search results, research, or current information from the web. Requires Tavily API key.
baidu-search
Search the web using Baidu AI Search Engine (BDSE). Use for live information, documentation, or research topics.
notebooklm
Google NotebookLM 非官方 Python API 的 OpenClaw Skill。支持内容生成(播客、视频、幻灯片、测验、思维导图等)、文档管理和研究自动化。当用户需要使用 NotebookLM 生成音频概述、视频、学习材料或管理知识库时触发。
openclaw-search
Intelligent search for agents. Multi-source retrieval with confidence scoring - web, academic, and Tavily in one unified API.
aisa-tavily
AI-optimized web search via AIsa's Tavily API proxy. Returns concise, relevant results for AI agents through AIsa's unified API gateway.
Data Analyst — AfrexAI ⚡📊
**Transform raw data into decisions. Not just charts — answers.**
Competitor Monitor
Tracks and analyzes competitor moves — pricing changes, feature launches, hiring, and positioning shifts
afrexai-competitive-intel
Complete competitive intelligence system — market mapping, product teardowns, pricing intel, win/loss analysis, battlecards, and strategic monitoring. Goes far beyond SEO to cover the full business landscape.