Back to Blog

50 ChatGPT SEO Prompts That Actually Work (Because They Produce a Backlog)

A ChatGPT SEO prompt library that creates a backlog, not just text. Rerunnable prompts for tracking AI visibility, forcing citations, and shipping fixes.

January 4, 202626 min read
Medieval scriptorium workspace with glowing prompt scroll, signal fires representing AI platforms, and frosted glass overlay showing article title

"How do you know when ChatGPT is mentioning your brand? Specifically what queries."

That question from r/bigseo captures why most ChatGPT SEO prompts don't matter. If you can't answer "which queries?", you're prompting blind.

ChatGPT SEO prompts are templates used to query AI models for visibility data (mentions, citations, recommendations) and to generate content that gets cited. The prompts that work aren't one-off spells; they're rerunnable tests you run weekly to track changes and produce execution tickets.

Here's the measurement reality: Pew Research found that clicks drop from 15% to 8% when an AI summary appears. Seer Interactive tracked a 61% decline in organic CTR for queries with AI Overviews. AI summaries are eating clicks, and the only brands getting traffic are the ones getting cited.

A prompt library is cheap. The hard part is building a repeatable system that tracks your visibility, forces citations in your content, and turns gaps into shipping tickets you can run every week.

This guide gives you 50 prompts grouped by what they produce:

  • Tracking prompts that tell you where you're mentioned, cited, or recommended
  • Citation-forcing prompts that show you what sources AI models trust
  • Recommendation prompts for "best X" queries where competitors win
  • Leverage prompts that turn SME time into reusable assets
  • Backlog prompts that convert outputs into tickets you can ship

Let's get into it.

Check if your brand appears in ChatGPT, Perplexity, and Google AI Overviews →

What makes a ChatGPT SEO prompt "work"?

A prompt works if it produces a decision you can repeat and measure. That's it.

Most prompt libraries don't do this. As one r/PromptEngineering user put it: "Perfect prompting strategists and prompt aggregators vibe like witches writing spell books now."

They're right. A prompt that outputs "here's a paragraph about X" doesn't change anything. A prompt that outputs "you're not cited, your competitors are, here are the sources you need to match, here's your first ticket" does.

A prompt works when it creates three things:

  1. Outcomes you can measure: mentioned, cited, recommended, or missing
  2. Artifacts you can version: a scorecard, a gap list, a ticket backlog
  3. A cadence you can repeat: weekly reruns that track deltas over time

The prompts in this guide are designed to run repeatedly. You're not looking for "the perfect prompt." You're building an audit harness.

The scorecard rubric

Before running any prompt, define what you're measuring. Here's the rubric we use:

QuestionWhat to record
Are we mentioned?Yes/No + exact phrase
Are we cited (with a link)?Yes/No + URL if present
Are we recommended?Yes/No + position in list
What competitors appear?List of brands + their positions
What sources does the model cite?URLs + domains
What's missing from our content?Gap notes for backlog

Run the same prompts weekly. Track the changes. That's what makes prompts work.

Start here: 10 prompts to run every week (your AI visibility harness)

Before you write anything, you need to know where you stand. These prompts give you query-level visibility data you can track over time.

The measurement reality is messy. As one r/TechSEO user noted: "LLMs don't pass referrers, everything just gets dumped into direct." Chris Long, VP of Marketing at Go Fish Digital, confirmed this: "ChatGPT does tag itself and is making efforts to add more tracking. Occasionally, you'll see it in GA. But most of the time, LLM referrals show up as homepage traffic."

You can't rely on analytics for this. You measure presence directly.

Prompt 1: Query discovery

I run [COMPANY] in the [INDUSTRY] space. Our main competitors are [COMPETITOR 1], [COMPETITOR 2], and [COMPETITOR 3].
 
List 20 queries a potential customer might ask when researching [PRODUCT/SERVICE CATEGORY]. Focus on:
- Definitional queries ("What is...")
- Comparison queries ("X vs Y", "Best X for...")
- Problem queries ("How to solve...")
- Vendor queries ("Who offers...", "Top companies for...")
 
Format as a numbered list with query intent noted in parentheses.

Prompt 2: Brand mention check

For the query "[YOUR TARGET QUERY]", answer as you would for a user.
 
Then answer these questions:
1. Did you mention [YOUR BRAND]? If yes, in what context?
2. Did you cite any content from [YOUR DOMAIN]?
3. What competitors did you mention?
4. What sources would you cite if asked to provide references?

Prompt 3: Competitor presence audit

For the query "[YOUR TARGET QUERY]":
 
1. List all brands/companies you would mention or recommend
2. For each brand, note:
   - Position in your response (first mentioned, recommended, etc.)
   - Why you included them
   - What sources informed this inclusion
 
Format as a table with columns: Brand | Position | Reason | Source.

Prompt 4: Citation source analysis

For the query "[YOUR TARGET QUERY]":
 
1. What sources would you cite to support your answer?
2. For each source, explain why you trust it
3. What types of sources are missing that would strengthen the answer?
4. If my company [YOUR BRAND] published content on this topic, what would it need to include to be cited?

Prompt 5: Recommendation criteria extraction

When recommending [PRODUCT/SERVICE CATEGORY] providers, what criteria do you use to evaluate them?
 
List each criterion and explain:
- How you verify it (what sources you check)
- What threshold makes a brand "recommendable"
- What would disqualify a brand
 
Format as a table with columns: Criterion | Verification Method | Threshold | Disqualifiers.

Prompt 6: Gap identification

I want my brand [YOUR BRAND] to be mentioned/cited/recommended for the query "[YOUR TARGET QUERY]".
 
Currently, you [mention/cite/recommend] [COMPETITOR] instead.
 
What specific content, credentials, or third-party signals would [YOUR BRAND] need to appear alongside or instead of [COMPETITOR]?
 
Be specific about:
- Content requirements (what to publish)
- Third-party requirements (reviews, lists, mentions)
- Authority signals (credentials, case studies, data)

Prompt 7: Source trust hierarchy

For the query "[YOUR TARGET QUERY]", rank these source types by how much you trust them:
 
- Company websites
- Industry publications (name examples)
- Review sites (name examples)
- Academic/research sources
- Government/regulatory sources
- Community forums (Reddit, Quora)
- Comparison sites
 
For each, explain what makes a source in that category trustworthy or untrustworthy.

Prompt 8: Content gap analysis

Here's the current content on [YOUR URL] about [TOPIC]:
 
[PASTE CONTENT SUMMARY OR KEY POINTS]
 
For the query "[YOUR TARGET QUERY]":
 
1. What's missing that would make this content citable?
2. What claims need sources?
3. What statistics would strengthen it?
4. What expert quotes would add credibility?
5. What comparisons or alternatives should be mentioned?

Prompt 9: Weekly delta prompt

Last week, for the query "[YOUR TARGET QUERY]", you said:
 
[PASTE LAST WEEK'S RESPONSE SUMMARY]
 
Answer the same query again today. Then compare:
1. Did the brands mentioned change?
2. Did the sources cited change?
3. Did the recommendations change?
4. What might have caused any changes?

Prompt 10: Backlog generation

Based on the gaps identified for [YOUR BRAND] across these queries:
- [QUERY 1]
- [QUERY 2]
- [QUERY 3]
 
Generate a prioritized backlog of content and visibility tasks. For each task:
- Describe the work
- Estimate effort (small/medium/large)
- Note the expected impact on visibility
- Identify dependencies or blockers
 
Format as a table with columns: Task | Effort | Impact | Dependencies.

What to record every week

After running these prompts, fill out your scorecard:

QueryMentioned?Cited?Recommended?CompetitorsSources citedGap notes

Track changes week over week. That's your visibility harness.

Prompts that force citations (and show you the sources you need)

If you don't force the model to cite sources, you can't learn what it's weighting. These prompts expose the citation hierarchy.

The Princeton GEO study found that including citations, quotations, and statistics can boost source visibility by 30-40% in generative answers. The mechanism matters: AI models prefer content that looks citable because it has the proof density they need to synthesize answers.

Prompt 11: Cite-first definitional query

Define [TERM/CONCEPT] as you would for a professional audience.
 
Requirements:
- Cite at least 3 sources for your definition
- Include the URL for each source
- Explain why each source is authoritative
- Note if any sources conflict and how you resolved it
 
If you cannot find sources, say "No authoritative source found" rather than inventing one.

Prompt 12: Source verification request

For the claim "[SPECIFIC CLAIM ABOUT YOUR INDUSTRY]":
 
1. Is this claim accurate?
2. What sources support or refute it?
3. Provide URLs for each source
4. Rate your confidence in each source (high/medium/low) and explain why
5. What additional sources would you need to be certain?

Prompt 13: Reverse-engineer citation needs

I want to publish content about [TOPIC] that AI models will cite when answering [QUERY TYPE] questions.
 
What would that content need to include to be considered citable?
 
Be specific about:
- Types of data/statistics required
- Credentialing needed (author expertise, organizational authority)
- Format requirements (structure, headers, extractable elements)
- Third-party validation needed

Prompt 14: Comparison with sources

Compare [OPTION A] vs [OPTION B] for [USE CASE].
 
Requirements:
- Support each comparison point with a source
- Provide URLs for sources
- If no source exists for a comparison point, note it as "unsupported observation"
- Include a summary table with sources linked

Prompt 15: Expert quote extraction

For the topic [TOPIC], who are the recognized experts whose opinions you would cite?
 
For each expert:
- Name and credentials
- Why they're considered authoritative
- A notable position or quote (with source)
- How I could get quoted alongside them

The operational reality: Understanding what AI cites is table stakes. The execution—tracking visibility across engines, engineering proof density into content, building third-party corroboration—is where most teams get stuck. That's the Track → Engineer → Leverage → Own system we build for clients.

Prompt 16: Missing proof identification

Here's a claim I want to make in my content:
 
"[YOUR CLAIM]"
 
1. What evidence would you need to cite this claim?
2. What types of sources would make it credible?
3. If I can't find external evidence, how should I reframe it?
4. What internal data or case study would serve as acceptable proof?

Prompt 17: Fact-check format

Evaluate these claims from my content for citability:
 
[PASTE 3-5 CLAIMS FROM YOUR CONTENT]
 
For each claim:
- Is it verifiable? (Yes/No)
- What source would verify it?
- If unverifiable, suggest how to reframe it as a qualified opinion or remove it

Prompt 18: Citation density analysis

Here's a piece of content about [TOPIC]:
 
[PASTE CONTENT]
 
Analyze its citation density:
1. How many claims are made?
2. How many have supporting evidence?
3. What's the citation ratio?
4. Which unsupported claims most need sources?
5. What statistics would strengthen the weakest sections?
 
The Princeton GEO study suggests content with higher proof density is more likely to be cited by AI. Rate this content's "citability" on a 1-10 scale.

When ChatGPT recommends competitors: prompts for "best X" and vendor lists

"Best X" queries are where the money is. And where your off-site footprint matters most.

AI referrals are growing fast. TechCrunch reported that AI platforms generated 1.13 billion referrals to top sites in June 2025—up 357% year-over-year. Still small compared to Google's 191 billion, but the growth curve matters.

When someone asks "best [your category] for [use case]," the AI isn't just checking your website. It's synthesizing from review sites, comparison articles, community discussions, and directory listings.

Prompt 19: Recommendation criteria audit

When a user asks "What's the best [PRODUCT/SERVICE CATEGORY] for [USE CASE]?", what criteria do you use to make recommendations?
 
For each criterion:
- How do you verify it?
- What sources inform it?
- What weight does it carry in your final recommendation?
 
Format as a ranked list with verification sources.

Prompt 20: Competitor differentiator extraction

For the query "best [PRODUCT/SERVICE CATEGORY] for [USE CASE]":
 
List all competitors you would recommend. For each:
- Their primary differentiator
- The source of that differentiator (their site, reviews, third-party comparisons)
- What would need to change for them to lose that position

Prompt 21: Third-party corroboration check

For my brand [YOUR BRAND] to be recommended for "[BEST X QUERY]":
 
1. What third-party sources mention us currently?
2. What third-party sources mention our competitors that don't mention us?
3. Specifically, which of these would matter most:
   - Industry comparison articles
   - Review sites (which ones?)
   - Directory listings (which ones?)
   - Community discussions (which platforms?)
   - Expert recommendations (whose?)

Prompt 22: Review and rating synthesis

For [YOUR PRODUCT CATEGORY], what review platforms do you check when forming recommendations?
 
For each platform:
- How much weight do you give it?
- What rating threshold matters?
- Do you factor in review volume?
- Do you analyze sentiment beyond star ratings?

Prompt 23: List inclusion gap

I found these "best [CATEGORY]" articles that include competitors but not my brand:
 
- [URL 1]
- [URL 2]
- [URL 3]
 
For each article:
1. Why might I be missing from this list?
2. What criteria would I need to meet for inclusion?
3. Who should I contact to request consideration?
4. What format do they expect from submissions?

Prompt 24: Competitive positioning prompt

Position [YOUR BRAND] against [COMPETITOR 1], [COMPETITOR 2], and [COMPETITOR 3] for the query "[BEST X QUERY]".
 
Create a comparison table with:
- Key decision criteria (use the criteria AI models weight from Prompt 19)
- How each brand performs on each criterion
- Gaps where [YOUR BRAND] underperforms
- Specific actions to close each gap

Prompt 25: "Who should use this" framing

For [YOUR PRODUCT/SERVICE]:
 
1. What type of user/company should you recommend this to?
2. What type should you NOT recommend this to?
3. What are the honest limitations?
4. How would you phrase a recommendation that acknowledges trade-offs?
 
Write a sample recommendation paragraph that's honest about fit.

Prompt inputs you need before you prompt (or your output will be generic)

Prompts are not magic. Inputs win.

The quality of any prompt output depends entirely on what you feed it. Generic inputs produce generic outputs. If you want prompts that produce actionable work, you need to build these input assets first.

This is where you leverage SME time. Instead of asking your experts to write content (slow, expensive), you extract their knowledge into reusable formats that make every prompt output better.

Prompt 26: Claim ledger builder

I'm building a "claim ledger" for [YOUR BRAND]. This is a list of things we can credibly claim, with supporting evidence.
 
For each claim below, help me categorize and strengthen it:
 
Claim: "[YOUR CLAIM]"
 
1. Is this a fact, opinion, or outcome?
2. What evidence supports it?
3. How should it be phrased to be both accurate and compelling?
4. What caveats or qualifications are needed?
5. What source should be cited?

Prompt 27: Stat bank creation

I need to build a stat bank for content about [TOPIC].
 
For each statistic:
- State the stat with exact numbers
- Provide the original source URL
- Note the date of the data
- Explain the methodology briefly
- Rate its shelf life (how soon will it be outdated?)
 
Help me convert this raw data into usable content statistics:
 
[PASTE YOUR RAW DATA OR FINDINGS]

Prompt 28: Quote bank from SME interview

Here's a transcript from an interview with [SME NAME, TITLE]:
 
[PASTE TRANSCRIPT EXCERPT]
 
Extract quotable statements that:
1. Contain specific insights (not generic advice)
2. Could stand alone as cited quotes
3. Demonstrate expertise
4. Are controversial or contrarian enough to be interesting
 
For each quote, suggest:
- The pull quote (exact words)
- A lead-in that provides context
- Topics where this quote would be relevant

Prompt 29: Criteria library for "best X" content

When we evaluate [PRODUCT/SERVICE CATEGORY], what criteria do we use?
 
For each criterion:
1. What is it?
2. How do we measure it?
3. What's "good" vs "acceptable" vs "poor"?
4. What do competitors claim vs deliver?
5. Where do we genuinely outperform?
 
Format as a reference table we can use across comparison content.

Prompt 30: Red flags and disqualifiers

When evaluating [PRODUCT/SERVICE CATEGORY] for clients, what are our red flags?
 
List the signals that indicate a poor choice:
- Vendor red flags
- Product red flags
- Pricing red flags
- Support/reliability red flags
 
For each, explain why it matters and give an example.
 
This becomes our "honest evaluator" reference when we write comparisons.

Prompt 31: FAQ bank from customer conversations

Here are questions we get repeatedly from customers:
 
[PASTE QUESTIONS]
 
For each question:
1. Write a direct answer (2-3 sentences)
2. Identify follow-up questions it might trigger
3. Note if it reveals a misconception to address
4. Tag which buyer stage it represents (awareness/consideration/decision)
 
Format as an FAQ bank we can reuse across content.

Prompt → score → backlog → ship: the loop that makes prompts matter

If prompt outputs aren't tickets, nothing changes.

The entire point of this prompt library is to produce work—not paragraphs, not "insights," but specific tasks your team can ship. Seer Interactive found that being cited in an AI Overview means 35% more organic clicks and 91% more paid clicks. Citations matter. But you have to earn them by shipping.

Prompt 32: Gap-to-ticket converter

Based on this visibility gap analysis:
 
[PASTE GAP ANALYSIS FROM TRACKING PROMPTS]
 
Convert each gap into a specific ticket with:
- Title (action verb + target)
- Description (what exactly to do)
- Success criteria (how we'll know it's done)
- Estimated effort (hours)
- Dependencies (what needs to happen first)
- Priority (high/medium/low based on query volume and current gap severity)

Prompt 33: Content update prioritization

Here are content updates identified from our AI visibility audit:
 
[PASTE LIST OF POTENTIAL UPDATES]
 
Prioritize them using this matrix:
- Impact: How much will visibility improve? (high/medium/low)
- Effort: How long will it take? (small/medium/large)
- Proof availability: Do we have the data/sources to support the update? (yes/partial/no)
 
Create a prioritized backlog with the top 5 tickets to ship this week.

Prompt 34: Weekly sprint generator

It's Monday. Here's our current visibility state:
 
- Queries we're mentioned in: [LIST]
- Queries we're NOT mentioned in but should be: [LIST]
- Competitor gains this week: [ANY CHANGES]
 
Generate this week's sprint:
1. 2-3 content updates to improve existing pages
2. 1-2 new content pieces to fill gaps
3. 1-2 outreach tasks (get listed, get reviewed, get cited)
 
Include estimated hours for each task.

Prompt 35: Delta analysis prompt

Last week's visibility scorecard:
 
[PASTE LAST WEEK'S DATA]
 
This week's visibility scorecard:
 
[PASTE THIS WEEK'S DATA]
 
Analyze the changes:
1. What improved? Why?
2. What declined? Why?
3. What stayed the same despite our work?
4. What external factors might have influenced changes?
5. What does this tell us about next week's priorities?

Prompt 36: Proof requirement ticket

This content piece needs more proof density:
 
[PASTE CONTENT URL OR SUMMARY]
 
Create tickets for each proof type needed:
- Statistics to find or create
- Expert quotes to source
- Case studies to develop
- Third-party citations to add
- Data visualizations to create
 
For each ticket, include: source requirements, effort estimate, and assignee suggestions.

Prompt 37: Versioning documentation

We're updating our prompt library and visibility tracking. Create a version log entry:
 
Date: [TODAY]
Prompts updated: [LIST]
Scorecard changes: [LIST]
Backlog template changes: [LIST]
Reason for updates: [EXPLAIN]
Expected impact: [DESCRIBE]
 
This keeps our system documented and auditable.

Ready to see where you're invisible?

We'll run your key queries through ChatGPT, Perplexity, and Google AI Overviews and show you exactly where competitors get cited and you don't. Takes 30 minutes.

Get your AI visibility audit →


Common mistakes (why most "ChatGPT SEO prompts" fail)

Most people prompt for text. They should prompt for decisions and proof gaps.

Mistake 1: "Prompts are the work"

Prompts are a spec. The work is converting prompt output into edits, experiments, and distribution tickets.

If you run 50 prompts and produce zero shipped changes, you've accomplished nothing. The prompt library exists to create a backlog. The backlog creates velocity. Velocity creates visibility.

Mistake 2: "Schema is the lever"

Schema helps machines parse your content. It doesn't make you recommendable.

One r/TechSEO user asked: "Does extensive Schema markup actually help Large Language Models understand your entity better, or is it just for Google Rich Snippets?" The honest answer: schema helps parsing, but citations and recommendations are driven by proof density, extractable answers, and third-party corroboration.

Google's documentation is clear: "There are no additional requirements to appear in AI Overviews or AI Mode, nor other special optimizations necessary."

Schema is hygiene. It's not the lever.

Mistake 3: "Just publish a machine-readable version"

Another r/GenEngineOptimization user wondered: "Should I edit existing posts or publish a second 'machine-readable' version?"

Don't create duplicate pages. Apply extractability patterns to your canonical page:

  • Clear section headings
  • Answer-first paragraphs
  • Definition boxes for key terms
  • Comparison tables where relevant
  • Citations inline (not just a reference section)
  • Statistics with sources and dates

Mistake 4: "Referral traffic proves visibility"

Presence in answers is the first metric. Referral traffic is lagging and incomplete.

You might be mentioned in a hundred ChatGPT answers and see zero traceable referrals. Measure presence directly by running the tracking prompts weekly. Don't wait for analytics to tell you something they can't measure.

Mistake 5: "One prompt, run once"

One prompt run once tells you nothing. The same prompt run weekly tells you everything.

Visibility changes. Models update. Competitors ship. Your tracking prompts need to run on a cadence so you can measure deltas and correlate them with your work.

The remaining prompts: your complete library

Here are prompts 38-50 to round out your library:

Content engineering prompts

Prompt 38: Extractable answer block

Rewrite this paragraph as an extractable answer block:
 
[PASTE PARAGRAPH]
 
Requirements:
- Lead with the direct answer
- Add supporting evidence
- Include a citation
- Keep it under 100 words
- Make it standalone (could be quoted without context)

Prompt 39: Comparison table generator

Create a comparison table for [TOPIC] covering:
- [OPTION A]
- [OPTION B]
- [OPTION C]
 
Columns: [CRITERIA 1], [CRITERIA 2], [CRITERIA 3], Best For
 
Include source notes for any factual claims.

Prompt 40: FAQ from query intent

For the query "[TARGET QUERY]", generate 5 FAQ questions a searcher would want answered.
 
For each question:
- Write a direct 2-3 sentence answer
- Suggest a source to cite
- Identify follow-up content to link

Off-site visibility prompts

Prompt 41: Directory presence audit

What directories and listing sites matter for [INDUSTRY/CATEGORY]?
 
For each:
- Name and URL
- How to get listed
- What information they require
- How prominent listings affect AI recommendations

Prompt 42: Review site strategy

For [PRODUCT/SERVICE CATEGORY], which review platforms influence AI recommendations most?
 
For each platform:
- Name and URL
- How reviews factor into AI responses
- Minimum review volume/rating thresholds (if observable)
- How to encourage customer reviews ethically

Prompt 43: Community presence mapping

What online communities discuss [TOPIC]?
 
For each:
- Platform and specific community
- Types of discussions relevant to [YOUR BRAND]
- Rules around promotional content
- How to add value without violating norms

Content refresh prompts

Prompt 44: Stale content identifier

Here's content published [DATE]:
 
[PASTE CONTENT OR SUMMARY]
 
What needs updating?
- Outdated statistics (cite newer sources)
- Outdated advice (what's changed)
- Missing topics (what's emerged since publication)
- Competitive gaps (what competitors now cover that we don't)

Prompt 45: Proof upgrade list

This content makes these claims:
 
[PASTE CLAIMS]
 
For each claim:
- Current proof (what we cite now)
- Better proof available (suggest alternatives)
- Proof we should create (internal data, case studies)

Prompt 46: Internal linking opportunities

Given this content about [TOPIC]:
 
[PASTE CONTENT SUMMARY]
 
And these other pages on our site:
- [URL 1]: [TOPIC 1]
- [URL 2]: [TOPIC 2]
- [URL 3]: [TOPIC 3]
 
Where should we add internal links? For each opportunity:
- The anchor text to use
- Where in the content to place it
- Why this link adds value for readers

Distribution prompts

Prompt 47: Reddit answer draft

Here's a Reddit post asking about [TOPIC]:
 
[PASTE POST]
 
Draft a helpful response that:
- Directly answers the question
- Shares genuine insight from experience
- Does NOT pitch or link (unless explicitly relevant)
- Positions [YOUR BRAND] as knowledgeable without being promotional
 
Follow Reddit norms: be helpful first, be human, avoid marketing speak.

Prompt 48: Guest post angle generator

We want to contribute to publications covering [TOPIC].
 
Generate 5 guest post angles that:
- Offer unique insight we can credibly deliver
- Would interest their audience (not just ours)
- Include data or research we can provide
- Position [YOUR BRAND] as a thought leader without being promotional
 
For each angle: title, 2-sentence pitch, data/proof we'd include.

Measurement prompts

Prompt 49: Visibility trend report

Based on 4 weeks of visibility data:
 
[PASTE WEEKLY SCORECARDS]
 
Generate a trend report:
1. Overall trajectory (improving/declining/flat)
2. Queries with biggest gains
3. Queries with biggest losses
4. Correlation with our shipped work
5. Recommendations for next month

Prompt 50: Quarterly visibility review

Based on our visibility tracking over [TIMEFRAME]:
 
[PASTE SUMMARY DATA]
 
Generate a quarterly review:
1. Executive summary (3 bullets)
2. Key wins (with evidence)
3. Persistent gaps
4. Recommendations for next quarter
5. Resource requirements to execute recommendations

Frequently asked questions

How do you know when ChatGPT is mentioning your brand? Specifically what queries.

You can't know passively. LLMs don't pass referrers consistently, and analytics won't show you query-level data.

The solution is active monitoring: run the same tracking prompts (1-10 above) weekly across ChatGPT, Perplexity, and Claude for your target queries. Record mentions, citations, and recommendations in a scorecard. Track changes over time.

Best AEO/GEO tracker?

Most practitioners feel lost about AI visibility tools—"they all have pretty similar features." The honest answer: the tool matters less than the system.

Start with the prompt harness in this guide (free, rerunnable). If you want automation, look for tools that track mentions + citations + recommendations across multiple AI platforms, with query-level granularity. But the prompts come first.

Is anyone else confused by AI traffic? ChatGPT is clearly sending visits but analytics shows nothing.

Yes. This is a common frustration. AI traffic often shows up as direct or homepage traffic. Don't wait for analytics to solve this.

Measure presence directly. Referral traffic is a lagging, incomplete signal. The prompts in this guide let you measure visibility without depending on referrer data.

Does extensive schema markup actually help LLMs understand your entity better?

Schema helps parsing; it doesn't guarantee recommendations. This r/TechSEO thread captures the debate well.

Schema markup is hygiene—do it. But citations, recommendations, and mentions are driven by proof density, extractable answers, and third-party corroboration. Don't over-index on schema as the "one lever."

Should I edit existing posts or publish a second "machine-readable" version?

Edit the canonical page. Don't create duplicate content.

Apply extractability patterns: clear headings, answer-first paragraphs, definition boxes, comparison tables, inline citations with dates. These make your existing content more citable without creating a confusing duplicate.

How do LLMs read websites? Is a certain website structure working well for LLM visibility?

This r/bigseo thread discusses it. Structure helps, but it's not the whole story.

What matters: clear HTML hierarchy, descriptive headings, answer-first content, proof density (stats + citations + quotes), and third-party corroboration. The structure makes content parseable. The proof makes it citable.

What to do next

You now have 50 prompts. But prompts without a system are just a list.

Here's the system that makes them work:

  1. Run prompts 1-10 this week to establish your baseline visibility
  2. Fill out the scorecard for your top 10 target queries
  3. Identify your 3 biggest gaps (queries where competitors appear and you don't)
  4. Generate tickets using prompts 32-37 for each gap
  5. Ship the tickets over the next 2-3 weeks
  6. Rerun the tracking prompts and measure the delta
  7. Repeat

The prompts produce a backlog. The backlog produces velocity. Velocity produces visibility.

Key takeaways:

  • Prompts work when they're rerunnable and produce decisions
  • Tracking visibility (mentioned/cited/recommended) matters more than waiting for referral data
  • Proof density (stats + citations + quotes) drives citability
  • Off-site corroboration (reviews, lists, directories) influences recommendations
  • The backlog loop is the system: prompt → score → ticket → ship → measure → repeat

No one can guarantee AI citations. What you can do is the work that makes citations possible: being everywhere, with quality content, consistently. Prompts are the audit. The backlog is the work.



Typescape makes expert brands visible everywhere AI looks. Get your AI visibility audit →