Back to Blog
SEO

Why ChatGPT and Perplexity Don't Recommend Your Website (And How to Fix It)

16 min read
Published:
Share:
Why ChatGPT and Perplexity Don't Recommend Your Website (And How to Fix It)

Why ChatGPT and Perplexity Don't Recommend Your Website (And How to Fix It)

The short version: Your competitor shows up in AI recommendations. You don't. This isn't random. Here are 7 specific things that cause AI engines to skip websites — with exact fixes for each one. Get your GEO diagnosis →

Test Yourself First

Before diving into diagnosis, you need a baseline. Here's exactly how to test whether you're currently appearing in AI search results.

In Perplexity (perplexity.ai): Open a new conversation and run these searches. Replace the bracketed text with your own specifics:

  • "[your service] agency [your city]"
  • "best [your service] company for [your target industry]"
  • "[main service you offer] guide"
  • "how to choose a [your service] provider"

Note whether your brand name, website URL, or content appears in the answer or sources.

In ChatGPT (chatgpt.com — with web search enabled): Use the same queries. Make sure web search is turned on in ChatGPT's settings — the default mode draws from training data rather than real-time results, so you need to test both modes.

In Google (google.com): Search your core queries and look at the top of the results. If there's a gray box with a summarized answer, that's an AI Overview. Check whether any of your content appears as a source.

Do this test now before reading further. It takes 10 minutes and gives you a concrete picture of where you stand. You might discover you appear in some but not all AI engines, which narrows the diagnosis considerably.

If you appear in Google AI Overviews but not Perplexity, the issue is likely content structure or brand authority — not bot access. If you appear in neither, start with Mistake #1.

Now, the seven mistakes. Work through each one methodically.

Mistake #1: AI Bots Are Blocked in robots.txt

Symptom: You don't appear in any AI search engine, regardless of query.

This is the most common mistake and the easiest to fix. Many sites accidentally block the bots used by AI companies in their robots.txt file.

Check your robots.txt right now at yourdomain.com/robots.txt. Look specifically for these bot names:

  • GPTBot (OpenAI/ChatGPT)
  • ClaudeBot (Anthropic/Claude)
  • PerplexityBot (Perplexity)
  • Google-Extended (Google AI/Gemini)
  • CCBot (Common Crawl, used by many AI systems)

If you see any of these in a Disallow rule, you've found the problem.

Here are the most common ways sites accidentally block AI bots:

```

Blocking pattern 1: Wildcard disallow (blocks everything including AI bots)

User-agent: * Disallow: /

Blocking pattern 2: Specific bot listed in error

User-agent: GPTBot Disallow: /

Blocking pattern 3: Left in from a "block all scrapers" tutorial

User-agent: CCBot Disallow: / ```

The fix:

``` User-agent: * Disallow: /admin/ Disallow: /wp-admin/ Allow: /

User-agent: GPTBot Allow: /

User-agent: ClaudeBot Allow: /

User-agent: PerplexityBot Allow: /

User-agent: Google-Extended Allow: /

Sitemap: https://yourdomain.com/sitemap.xml ```

Beyond robots.txt, check for server-level blocks. Cloudflare's "Bot Fight Mode" and many WAF configurations block AI crawlers before they even reach your robots.txt. If you use Cloudflare, go to Security → Bots and verify that legitimate AI crawlers aren't being blocked. Cloudflare now has specific AI bot categories you can manage.

Similarly, some WordPress plugins — particularly security plugins in aggressive mode — maintain their own bot blocklists. Check WordFence, Sucuri, or whatever security plugin you're using for bot management settings.

Fix time: 15–30 minutes to update robots.txt and verify. Server configuration changes may require more time depending on your setup.

Mistake #2: No llms.txt File

Symptom: AI systems find your site but misrepresent your services, call you the wrong type of company, or describe your offerings inaccurately.

If AI bots can access your site but consistently get your description wrong, the issue is often a lack of direct communication about what your site does.

llms.txt is a markdown file at yourdomain.com/llms.txt that tells AI systems what your company does, who you serve, and which pages are most important. When an AI system retrieves your homepage and sees a JavaScript animation and the headline "We build better digital experiences," it has very little to work with. An llms.txt file provides the clear, factual description AI systems need.

The fix is straightforward: create the file. For a complete guide with step-by-step instructions, a format breakdown, and a ready-to-adapt template, read our detailed guide: What Is llms.txt? The New Standard for AI-Ready Websites.

For an immediate stopgap, at minimum include: 1. A two-sentence description of exactly what your company does and who you serve 2. Links to your 3–5 most important service pages with one-sentence descriptions 3. Links to your 3–5 best content pieces (case studies, guides) with descriptions

Fix time: 30–60 minutes to create a solid initial version.

Mistake #3: Content AI Can't Extract

Symptom: You appear sometimes in AI results, but competitors' answers get used instead of yours — even when you've written extensively on the topic.

This is the subtlest and most impactful mistake. Your content might be excellent, but if it's structured in a way that makes extraction difficult, AI systems will consistently prefer your competitors' more extractable content.

AI language models are optimized to find and lift coherent, self-contained answers. Content that requires reading an entire page to understand — with the key point buried halfway through after extensive context-setting — is harder to cite than content that leads with the answer.

What AI can't extract well:

``` Paragraph form, buried answers: "Many professionals in the industry have debated the best approach to this question over the years, with different schools of thought emerging from various research traditions. When we look at the data, we start to see some patterns, and one thing that consistently comes up is that structured content..."

[Key answer buried 200 words in, no clear attribution point] ```

What AI extracts easily:

``` Definition-first structure: Generative Engine Optimization (GEO) is the practice of structuring website content, technical configuration, and authority signals to appear in AI-generated search answers from systems like ChatGPT, Perplexity, and Google AI Overviews.

According to BrightEdge's 2025 AI Search Report, 68% of AI-cited pages used definition-first content structure on their key topic pages. ```

The differences are concrete:

1. Lead with definitions. If your page answers "what is X," put the answer in the first 200 words, formatted as a clear definition.

2. Use numbered steps for processes. "Step 1: Do this. Step 2: Do that." is extracted cleanly. "The first thing to consider is, followed by, and then you should think about..." is not.

3. Include statistics with source names. "Conversion rates improve" can't be cited. "According to [Source], conversion rates improve by X%" can be directly quoted.

4. Write FAQ sections. A heading that reads "Frequently Asked Questions" followed by question-and-answer pairs is one of the most reliably cited content formats across all major AI systems.

5. Use comparison tables. "Option A vs Option B" tables are extracted and cited at high rates. They answer comparative queries directly.

Rewriting your top 5–10 pages to follow these principles — without changing the actual information — typically produces measurable changes in AI citation rates within 4–6 weeks.

Mistake #4: No Author or Authority Signals

Symptom: AI engines mention your topic area but cite academic papers, news articles, or authoritative publications instead of you — even when your content is more practical and current.

AI systems have been trained to prefer content from identifiable, credentialed sources. When they encounter a page with no clear author, no stated credentials, and no verifiable claims, they treat it as lower-authority — even if the content is excellent.

The framework here is Google's E-E-A-T: Experience, Expertise, Authoritativeness, Trustworthiness. AI systems have internalized these signals through their training data and apply them implicitly.

What authoritative content looks like to AI:

  • Named author with a short bio: "Jane Smith, technical SEO consultant with 8 years of agency experience, specializing in B2B SaaS. Previously led SEO at [Company Name]."
  • First-person experience statements: "In our audit of 47 SaaS company sites last year, we found that..."
  • Specific case study data: "Client X, a London-based fintech startup, saw organic traffic increase from 2,400 to 8,100 monthly visits within 5 months of implementing our content restructuring approach."
  • Named client results (with permission, or anonymized but specific)
  • Business credentials: Google Partner badge, years in operation, notable clients

What low-authority content looks like:

  • "Our team" without names
  • "Experts say" without naming the experts
  • "Studies show" without citing which studies
  • Vague claims: "We deliver results" / "Industry-leading approach"

The fix is to audit your key pages and add specificity. Replace every anonymous claim with a named one. Replace every vague result with a specific number. Add author bios to every blog post and guide. This isn't just about AI visibility — it builds genuine trust with human readers too.

Fix time: 1–2 days to update existing pages with proper author information and first-person specificity. Ongoing commitment to writing with this level of detail.

Mistake #5: Missing Schema Markup

Symptom: Competitors consistently answer FAQ and "how to" queries in AI results, but your pages on the same topics don't get cited.

Schema markup is structured data in your page's HTML that explicitly tells AI systems and search engines what type of content they're looking at. Without it, systems must infer content type from structure and language — a process that produces more errors.

The most impactful schema types for AI citation:

FAQPage — If you have a section of question-and-answer pairs, marking it up with FAQPage schema makes it immediately recognizable as FAQ content. AI systems answer questions constantly. FAQPage schema is one of the clearest signals that your content is answer-formatted.

HowTo — For step-by-step guides, HowTo schema marks each step as a discrete unit. AI systems can then extract and cite individual steps, which dramatically increases citation frequency for procedural queries.

Article — Establishes publication date, author, and content type. Freshness is a real signal — AI systems prefer recent, dated content over undated content.

Here's a minimal but functional FAQPage schema:

``json { "@context": "https://schema.org", "@type": "FAQPage", "mainEntity": [ { "@type": "Question", "name": "What is the difference between SEO and GEO?", "acceptedAnswer": { "@type": "Answer", "text": "SEO (Search Engine Optimization) focuses on ranking pages in traditional search results through backlinks, content quality, and technical factors. GEO (Generative Engine Optimization) focuses on getting your content cited by AI answer engines like ChatGPT, Perplexity, and Google AI Overviews. The techniques overlap partially but require different structural approaches." } } ] } ``

Add this to your page inside a <script type="application/ld+json"> tag in the <head>. Validate it at schema.org/validator before publishing.

If you're on WordPress, Yoast SEO Premium and Rank Math Pro both have UI-based schema builders that don't require manual JSON editing. On Next.js, you can create a reusable SchemaMarkup component that accepts props and renders the JSON-LD in the document head.

Fix time: 2–4 hours for a developer to implement basic schema across your key pages, assuming a simple stack. More complex implementations may take longer.

Mistake #6: JavaScript-Only Rendered Content

Symptom: Your site looks excellent visually, but AI bots seem to miss large portions of your content. Your analytics show bot visits but you have low citation rates even for content you know is relevant.

This is a technical issue that's surprisingly common, especially on newer sites built with React, Vue, or Angular without server-side rendering.

When a browser loads a JavaScript-heavy site, here's what happens: the browser receives a near-empty HTML file, downloads several JavaScript bundles, executes that JavaScript, and then renders the actual content. For a human using a modern browser, this happens fast enough to be invisible.

For an AI bot, it looks like this: the bot requests your page and receives an HTML file with barely any text content — just the scaffold for JavaScript to fill in later. Unless the bot waits for JavaScript execution (an expensive operation that most bots don't do), it sees a mostly blank page.

How to check if this is your problem:

Open your site in a browser. Right-click and choose "View Page Source" (not Inspect Element — you want the raw HTML, not the rendered DOM). If your key content — headlines, service descriptions, body text — doesn't appear in the raw HTML, your content is JavaScript-rendered and invisible to most bots.

The fix:

For sites on Next.js, this is straightforward: use Server-Side Rendering (SSR) or Static Site Generation (SSG) for your key pages. In Next.js App Router, components are server-rendered by default — avoid marking key content components as "use client" unless necessary. If they need client-side interactivity, split them into a server-rendered shell with a client-side interactive layer.

For sites on React without Next.js, migrating to Next.js is the cleanest long-term solution. For a faster fix, consider using a prerendering service (Prerender.io, Rendertron) that serves bot-friendly HTML versions of your pages.

For Vue or Angular sites, Nuxt.js and Angular Universal provide SSR solutions equivalent to Next.js.

The performance benefit is a welcome side effect: server-rendered pages load faster, improving Core Web Vitals scores and supporting better traditional SEO rankings at the same time.

Fix time: Ranges from hours (if already on Next.js, just removing unnecessary "use client" directives) to weeks (full migration from a client-rendered React app).

Mistake #7: No External Brand Mentions

Symptom: Newer but more active competitors consistently appear in AI results ahead of you, despite your longer track record and potentially better work.

AI systems don't evaluate your website in isolation. They assess your brand's presence across the broader web — the same way you'd research a company by looking at their website, reading about them on industry sites, checking reviews, and watching their videos. A brand that exists only on its own website is treated with more skepticism than one with external validation.

Perplexity's citation behavior illustrates this clearly. In a 2025 analysis by Authoritas, sites with at least three external mentions in credible publications received AI citations at 2.8x the rate of sites with no external mentions, controlling for content quality.

The channels that matter most:

Reddit. Perplexity treats Reddit content as highly credible. Genuine participation in relevant subreddits — answering questions in r/entrepreneur, r/webdev, r/SEO, or industry-specific communities — builds your brand presence in a context AI systems heavily reference. This requires actual value-add, not promotional posts. A history of helpful comments with your company name mentioned builds citations over months.

YouTube. Video transcripts get indexed. A tutorial video on a topic relevant to your services, with 500 views and a properly structured description, establishes your brand as a practitioner. AI systems frequently reference YouTube content when answering "how to" queries.

Industry publications. A byline article in Search Engine Journal, Marketing Week, a regional business journal, or an industry-specific publication carries significant authority weight. Pitching these isn't as hard as it seems — editors are often looking for practical, experience-based content from practitioners.

LinkedIn. Long-form LinkedIn articles on industry topics build author authority and get indexed. When your name appears across LinkedIn, YouTube, and your own site discussing the same topics, it strengthens the web of signals that AI systems use to assess expertise.

Podcast appearances. Podcasts produce transcripts. Those transcripts get indexed. Even niche podcasts with modest audiences contribute to your external presence.

The compound effect of this activity is what makes it powerful. Six months of consistent external presence — two YouTube videos, ten Reddit answers, one publication byline, regular LinkedIn posts — produces a web of external validation that your competitors who are only working on their own site simply can't match in AI results.

Fix time: This is the longest-term fix. Initial mentions may appear within weeks; consistent, high-confidence AI citations from external validation build over 3–6 months.

Self-Diagnosis Checklist

Run through this table to identify your priority fixes:

MistakeHow to checkSeverityFix time
AI bots blockedView yourdomain.com/robots.txt — look for GPTBot, ClaudeBot, PerplexityBot in DisallowCritical30 min
No llms.txtVisit yourdomain.com/llms.txt — does it exist?High1 hour
Non-extractable contentDo your pages lead with definitions? Have FAQ sections? Use numbered steps?High2–5 days
No author/authority signalsAre author names, credentials, and specific results on your key pages?High1–2 days
Missing schema markupUse Google's Rich Results Test on your key pagesMedium2–4 hours
JavaScript-only renderingView Page Source — is key content in the raw HTML?Medium-HighHours to weeks
No external mentionsGoogle your brand name — any mentions outside your own site?Medium3–6 months

Work through the list from top to bottom. The first two fixes are fast and high-impact. The content and authority fixes take more time but produce the most durable results. The external mentions work should start immediately but doesn't fully pay off for months.

Frequently Asked Questions

We have strong SEO. Why doesn't that translate to AI visibility? Traditional SEO signals (backlinks, domain authority, page rankings) and AI citation signals are different systems. Content extractability, schema markup, and external brand presence matter more for AI citations than traditional authority metrics. A page ranking #1 on Google with unclear content structure often loses to a page ranking #5 with crystal-clear, definition-first content in AI results.

Do I need to optimize for each AI system separately? Mostly no. The same structural improvements — clear definitions, FAQ sections, proper schema, unblocked bot access — work across ChatGPT, Perplexity, and Google AI Overviews. The exception is Perplexity, which weights Reddit and social content more heavily than other systems. If Perplexity is particularly important for your industry, Reddit presence deserves extra attention.

How quickly will I see results after making these fixes? robots.txt and llms.txt changes take effect within days to two weeks (next crawl cycle). Content restructuring changes typically appear in AI results within 4–6 weeks. Schema markup effects appear within 6–12 weeks. External brand presence takes 3–6 months to show consistent impact. Track your results with weekly manual testing across the main AI engines.

My site is on Wix/Squarespace/Webflow. Can I still fix these issues? For most of these mistakes, yes. You can typically edit robots.txt in your platform's settings. You can create an llms.txt by uploading a plain text file. Content restructuring is platform-independent. Schema markup can be added through embed blocks or platform-specific apps. The JavaScript rendering issue is more complex on these platforms — this is one area where Next.js-based sites have a structural advantage.

Should I hire someone for GEO, or can I do it myself? The technical fixes (robots.txt, llms.txt, schema markup) are doable by most site owners with basic technical comfort. The content restructuring work requires solid writing skills and an understanding of how AI extraction works. The authority-building work (external mentions, press) often benefits from professional support. A full GEO audit and implementation is a reasonable professional engagement if you want to move faster or don't have the time to work through it systematically.

Next Steps

Working through this list methodically — starting with the fastest fixes and moving toward the longer-term ones — is how you systematically close the gap between your current AI visibility and where you want to be.

If you'd rather have a professional identify exactly which mistakes apply to your site and prioritize them by impact, we offer GEO audits that cover all seven areas. We'll tell you specifically what's blocking your visibility and give you an actionable implementation plan.

Request a GEO audit →

Related Guides:

Sources:

  • Authoritas — AI Citation Study, 2025
  • BrightEdge — AI Search Report 2025
  • Google Search Central — Structured Data Documentation
  • Perplexity AI — Crawling and Indexing FAQ

Tags

#GEO mistakes#ChatGPT not recommending#Perplexity visibility#AI search optimization#GEO fixes#AI citation#generative engine optimization
Share:

Related Posts

How to Appear in AI Search Results: GEO Strategy Guide 2026
March 28, 2026
18 min read
SEO

How to Appear in AI Search Results: GEO Strategy Guide 2026

You rank well on Google, but ChatGPT never recommends you. The problem isn't your SEO — it's your GEO. Here's the 7-step framework to fix that.

#GEO#generative engine optimization#AI search
Read More

Modern Web Design

Ready to Elevate Your Website?

Schedule a 15-minute strategy session with our team and let's build your digital roadmap together.