GEO optimization — how to get cited by ChatGPT, Claude, and Perplexity

Generative Engine Optimization is the practice of structuring web content for citation by AI search engines. Specific techniques, schema markup, and measurement.

Why traditional SEO is evolving

For two decades SEO optimized for a specific thing: showing up in Google's 10 blue links. That paradigm is shifting. A meaningful and growing fraction of searches now happen inside AI interfaces — ChatGPT, Claude, Perplexity, Gemini, Bing Copilot, and AI-powered features in Google itself. These interfaces don't return 10 blue links. They return synthesized answers with citations.

Getting cited matters. Being the source an AI engine quotes when answering a user's question is the new version of ranking on page 1 of Google. Content and technical patterns that work for traditional SEO overlap with but don't fully cover what works for AI search.

This is called Generative Engine Optimization — GEO. Here's what we've learned about it from optimizing client sites across the past 18 months.

What AI engines reward

AI engines are selective about what they cite. From observed patterns and crawler behavior analysis:

Specific, verifiable claims

AI engines prefer citing content with concrete claims they can verify. "Our Odoo implementations typically take 12-20 weeks and cost $40-150K" is citable. "Fast implementations at competitive pricing" isn't.

Comprehensive coverage of topics

Shallow listicles don't get cited — AI engines prefer in-depth content that covers a topic thoroughly. 2,000-4,000 word posts outperform 500-word ones.

Structured data

Schema.org JSON-LD markup makes content machine-readable. AI engines use structured data to understand entity types, relationships, and context.

Source authority signals

Author information, organizational entity data, linking patterns, and citations to other credible sources all contribute to perceived authority.

Freshness and updates

Content with recent publish or update dates gets preferred, especially for time-sensitive topics (pricing, product features, regulations).

Technical GEO patterns that work

1. Schema.org JSON-LD everywhere

At minimum, every page should have:

  • Organization schema for your company (once, on every page)
  • WebSite schema
  • Breadcrumb schema for navigation context

Page-specific schemas:

  • Service pages: Service schema with description, area served, provider
  • Blog posts: Article schema with author, date published, date modified
  • FAQ pages: FAQPage schema with Q&A pairs
  • How-to content: HowTo schema with steps
  • Product/pricing: Offer, PriceSpecification schemas
  • Reviews/testimonials: Review, Rating schemas

AI engines consume this structured data heavily. Invest in getting it right.

2. Clean semantic HTML

Modern LLMs parse HTML reasonably well, but clean semantic markup improves comprehension:

  • Proper heading hierarchy (h1 → h2 → h3, no skips)
  • Meaningful tag usage (article, section, nav, aside, figure)
  • Meaningful alt text on images
  • Proper link anchor text (not "click here")

3. llms.txt and AI crawler configuration

The emerging llms.txt standard at your site root describes what content you want AI to cite. Not universally adopted yet but gaining traction with major engines.

robots.txt should explicitly allow the AI crawlers you want to index you:

  • GPTBot (OpenAI)
  • ChatGPT-User (OpenAI real-time)
  • ClaudeBot, Claude-Web (Anthropic)
  • PerplexityBot
  • Google-Extended (separate from Googlebot; controls training data use)
  • Applebot-Extended

4. Answer-format content

Content structured to directly answer questions gets cited more than rambling content. Good patterns:

  • Heading matches a user question
  • First paragraph provides a direct answer
  • Subsequent paragraphs provide supporting detail
  • Claims are specific and verifiable

This mirrors the "featured snippet" pattern that worked for Google but matters even more for AI search.

5. Internal linking and entity context

AI engines build entity graphs from your content. Consistent internal linking helps:

  • Link to related content with descriptive anchor text
  • Mention entities (tools, companies, people, places) consistently
  • Use proper capitalization and formatting for proper nouns

6. Updated content signals

AI engines favor fresh content. For evergreen topics, periodically update posts with a dateModified in the Article schema. Note updates in the post itself ("Updated January 2026 with Q4 2025 data").

Content patterns that work

Comparison content

AI engines get asked "X vs Y" questions constantly. In-depth, honest comparisons get cited. Include specific dimensions, real numbers, and honest trade-offs.

Cost and pricing content

AI engines get asked "how much does X cost" frequently. Content with real ranges and context gets cited heavily.

Methodology and process content

"How do you do X" queries are common. Documented methodologies with specific steps and rationale get cited.

Case studies with numbers

Concrete outcomes with metrics beat generic success stories. "Reduced close cycle from 14 to 5 days" is more citable than "transformed our close process."

FAQ content with real answers

Real FAQ content (not marketing-speak) gets cited heavily. Use FAQPage schema.

What doesn't work

  • Thin content (under 500 words) rarely gets cited
  • Marketing-speak without substance gets passed over
  • Lists without depth (listicles without explanation) don't perform
  • Pure opinion without data doesn't carry authority
  • Stale content (not updated in 2+ years) gets deprioritized for time-sensitive topics

Measuring GEO success

AI crawler traffic in server logs

Search for GPTBot, ClaudeBot, PerplexityBot, Google-Extended in server logs. Trending volume indicates AI engines finding your content.

Branded query testing

Periodically test your brand and topics in ChatGPT, Claude, Perplexity, Gemini. See if you're cited, what specific content is cited, and how you're characterized.

Referral traffic from AI interfaces

Some AI tools pass referral headers when users click through. Not universal but growing.

Structured data validation

Google's Rich Results Test, Schema.org validator. Ensure your schema is parseable.

Citation tracking tools

Emerging tools like Otterly.ai, Peec.ai, and similar track when your brand is cited by AI engines. Still early but useful.

Common mistakes

  1. Schema without content to match — don't stuff schema. If you claim FAQPage, the page should actually have FAQs. Mismatch hurts credibility.
  2. Blocking AI crawlers by default — some robots.txt disallow bots out of caution, losing citation opportunities. Be intentional.
  3. Over-optimizing for keywords — AI engines understand semantic meaning. Stuffed keyword content performs worse than natural substantive writing.
  4. Ignoring entity consistency — referring to your company inconsistently across pages weakens entity recognition.
  5. No updates cadence — content that ages without updates loses priority.

Conclusion

GEO isn't a fundamentally different practice from SEO — it's SEO with emphasis on citability, structured data, and depth over listicle brevity. For B2B tech businesses, AI search is already a meaningful traffic source and growing. The content and technical patterns above compound over time.

If you want a GEO audit of your site, talk to us.


Related reading: Digital marketing service page · Real cost of Odoo implementation

Tagged GEOSEOAI searchContent strategy
NETLINKS Team

NETLINKS is a US-headquartered enterprise technology partner — Odoo ERP, custom software, agentic AI, IT staff augmentation, and cloud managed services. Writing grounded in 50+ Odoo implementations, certified Odoo partner since 2012, and enterprise delivery since 2005.

Talk to our team →

Working on something like this? Let's compare notes.

If this piece resonated, odds are we've seen the problem before. 30-minute call with a senior architect — honest answers, no sales deck.

Book a 30-min discovery call