Skip to main content
StackraStackra

Generative Engine Optimization (GEO) Audit Tool: Free AI Visibility Checker

Free GEO and AEO audit. Stackra checks your site signals for bot access, schema readiness, entity clarity, and infrastructure, and tells you exactly where your gaps are.

Free to sign up · Results in ~5 minutes

What is GEO?

Generative Engine Optimization refers to making your website appear as a cited source in AI-generated answers from tools like ChatGPT, Perplexity, and Google AI Overviews.

AI bot access

Whether major AI crawlers are permitted to reach and index your site.

Schema readiness

Whether your content has structured data that AI tools can parse and cite.

Entity clarity

Whether AI tools can identify your business name, location, and type.

Supporting signals

Whether your sitemap and robots.txt are reachable and correctly configured.

The shift

GEO vs SEO: how they differ

Traditional SEO

Optimizing to rank in a list of links. The goal is a high position in search results. The user clicks your link and lands on your site.

GEO

Optimizing to be cited in an AI-generated answer. The goal is to be the source an AI tool quotes when someone asks a relevant question. The user may never click; your brand is named in the answer.

Both matter. But they require different signals. A site that ranks well in Google may score poorly on GEO readiness if its structured data is thin, its entity information is ambiguous, or its content is locked behind JavaScript interactions that AI fetchers cannot see.

The audit

Four signal groups, assessed independently

Most audit tools that mention GEO check one thing: whether your robots.txt blocks Googlebot. Stackra checks four distinct groups and reports them separately so you can see exactly where your gaps are.

AI Visibility + GEO · Stackra scan result
Stackra GEO card showing AI crawler access status for GPTBot, OAI-SearchBot, ClaudeBot, Google-Extended, and PerplexityBot alongside schema readiness and entity clarity signals

AI crawler access in robots.txt (GPTBot, ClaudeBot, Google-Extended)

The most foundational GEO signal, and the one that answers "is my site visible to ChatGPT, Claude, and Perplexity?" Stackra reads and parses your robots.txt and checks access for each major AI crawler independently, not just as a pass or fail on the file as a whole.

GPTBot

OpenAI training + ChatGPT Search

OAI-SearchBot

ChatGPT live web retrieval

ClaudeBot

Anthropic / Claude training

Google-Extended

Google AI Overviews + Gemini

PerplexityBot

Perplexity AI

Informational only. Does not reliably honor robots.txt rules. Treat as advisory.

Naming each AI crawler explicitly in your robots.txt is the same effect as the wildcard User-agent: * rule alone. If you are looking to block specific crawlers, naming them is a good practice.

Schema readiness for AI citation (Article, HowTo, FAQPage, Organization)

Schema markup is the structured signal layer that tells AI tools what your content means, not just what it says. Stackra divides schema readiness into two sub-categories with different GEO functions.

Entity schemas

Establish who you are. AI tools use these to place your site in a knowledge graph and associate it with a real-world entity.

Organization

Primary entity signal for any business. Establishes your name, URL, description, and logo for AI knowledge graphs.

LocalBusiness

Extends Organization with address, phone, and hours. Use for any business with a physical location.

Person

For individual practitioners, consultants, and personal brands. Supports author attribution and expertise signals.

Citability schemas

Establish what you publish. These directly influence whether AI tools cite your pages when generating answers about topics your content covers.

Article / BlogPosting

Primary signal for editorial content. Applies to blog posts, guides, and opinion pieces.

HowTo

Highest citability signal for step-by-step instructional content. AI tools prioritize this for process and tutorial pages.

BreadcrumbList

Signals content hierarchy to AI crawlers. Commonly present on well-cited pages. Detected and stored but excluded from the citability count (subpage schema).

FAQPage

Universal AI citation signal. Google dropped all FAQPage rich results in May 2026, so it no longer produces a SERP feature for any site type. ChatGPT, Perplexity, and other generative tools actively use FAQPage content when generating answers, with no equivalent restriction.

Citability count: Stackra reports a citabilityTypeCount: the number of distinct active citability schema types detected. The count covers Article/BlogPosting, HowTo, and FAQPage, giving a range of 0–3. BreadcrumbList is detected and stored but excluded from the count (subpage schema, always absent on a homepage scan).

Entity clarity (Organization, LocalBusiness, Person schema)

Entity clarity answers whether AI tools can confidently identify who runs your site and where you operate. Detection is fully deterministic; no AI call is made. Stackra derives three signals from up to four layers per signal, falling back from the most structured to the most universal: JSON-LD, microdata itemprop values, geo meta tags, and Open Graph / publisher / author meta tags.

Business name

JSON-LD Organization/LocalBusiness name property → microdata itemprop → og:site_name, publisher tag

Location or service area

JSON-LD address, geo, areaServed → microdata address → geo.region, geo.placename meta tags

Named person

JSON-LD Person schema → microdata author itemprop → <meta name='author'>

Why metadata matters as a fallback. JSON-LD is the cleanest entity signal, but a large slice of the live web never adopted it: older WordPress installs without an SEO plugin, hand-coded sites, and hobbyist Squarespace and Wix builds. Stackra's detection mirrors what AI crawlers like Google-Extended and ClaudeBot already do (fall back to plain meta tags when richer markup is absent), so the entity confidence we report matches what those tools actually see.

Meta tags Stackra reads

  • og:site_name and the publisher meta tag: counted toward business name when no Organization schema is present.
  • meta name="author": counted toward named person when no Person schema is present.
  • geo.region and geo.placename: standard but underused location tags, counted toward location when no LocalBusiness address is present.

Low

0 signals confirmed

Moderate

1 signal confirmed

High

2+ signals confirmed

Supporting infrastructure signals (sitemap, robots.txt reachability)

Two infrastructure signals that determine whether AI crawlers can discover and fully index your site.

Sitemap reachable

Whether your sitemap.xml is present and returns a valid response. AI crawlers use sitemaps the same way Googlebot does, to discover pages they might not find through link crawling alone. WordPress (5.5+), Wix, Shopify, and Squarespace generate sitemaps automatically; custom-built sites and some older setups require manual generation. Submit yours to Google Search Console.

robots.txt reachable

Whether your robots.txt returns a valid response. A missing or unreachable robots.txt means crawlers fall back to permissive defaults, but it also signals a configuration gap that affects how reliably your access rules are enforced.

We scan ourselves

Stackra audits Stackra

Stackra's own site is structured against the same GEO signals it checks. When we run stackra.app through the scanner, the expected output is:

All major AI crawlers explicitly allowed in robots.txt (11 named stanzas)

Organization schema on the homepage with name, URL, description, logo, and sameAs to Product Hunt

Person schema on the About page and blog posts (founder as named entity with LinkedIn sameAs)

BlogPosting schema on every article (citabilityTypeCount: 1/2 at article level)

Entity clarity: high confidence (business name + named person confirmed)

Sitemap and robots.txt both reachable and cross-referenced

We also document what we intentionally omit: LocalBusiness schema (Stackra is a SaaS product, not a physical location) and HowTo schema (our blog content is not currently structured in the step-by-step format HowTo requires). FAQPage schema is included on relevant content pages. Google dropped all FAQPage rich results in May 2026, so it no longer produces a SERP feature, but ChatGPT and Perplexity use FAQPage content actively as a citation signal. Accurate schema is better than inflated schema.

By platform

What you can actually do, by platform

Your platform determines which GEO signals you can control. None of the following require a developer.

WordPress

Full server-level control. Every GEO signal is addressable through free plugins.

Rank Math SEO (free) →

AI bot access

A clean User-agent: * with an empty Disallow: allows all crawlers including AI bots, with no per-bot Allow rules needed. Use explicit bot-specific blocks only if you want to restrict a particular crawler. The most common intentional block is Google-Extended, which opts your content out of Google AI training without affecting standard search indexing.

Organization schema

Rank Math's Knowledge Graph settings output correct JSON-LD automatically from your business name, type, logo, and description. Yoast equivalent is under SEO > Search Appearance > Knowledge Graph & Schema.

Article schema on content pages

Rank Math and Yoast both apply Article or BlogPosting schema to posts automatically. Verify your content type is set correctly in each plugin's schema settings.

JavaScript content visibility

Standard themes render content in initial HTML. If you use Elementor or third-party accordion plugins, verify FAQ sections use native HTML details/summary elements, not JavaScript click-reveal components invisible to AI fetchers.

Wix

Strongest built-in GEO capability of any managed platform. Most signals are configurable without code.

Wix Structured Data Markup →

AI bot access

Wix provides a robots.txt editor at Settings > SEO > robots.txt. The default allows all crawlers; no per-bot Allow rules are needed. Only add explicit bot-specific blocks if you want to restrict a particular crawler, such as adding a Google-Extended block to opt out of Google AI training.

Organization/LocalBusiness schema

Wix auto-generates Organization or LocalBusiness schema from your business information. Complete all fields under Settings > Business Info; this feeds the schema automatically. Verify output with Google's Rich Results Test.

Custom schema

Wix's Structured Data Markup tool (Marketing & SEO > SEO Tools > Structured Data) supports HowTo blocks for step-by-step content. No code required.

Content visibility

Native Wix blocks are generally crawler-readable. If you have embedded third-party JavaScript widgets for accordions or tab content, verify the text is present in the page HTML, not injected after load.

Squarespace

AI bots allowed by default. Schema basics are handled by the platform. Custom schema requires code injection.

AI bot access

Squarespace generates robots.txt automatically. The default allows all crawlers including AI bots. No action required unless you have previously customized it.

Organization schema

Squarespace auto-generates basic Organization markup from your site settings. Complete your business name, description, and contact details under Settings > Business Information.

Custom schema

JSON-LD can be injected via Settings > Advanced > Code Injection. For HowTo pages or step-by-step service guides, this is the only way to add instructional content signals.

Content visibility

Native FAQ and accordion blocks keep content in HTML. Third-party JavaScript components embedded via code blocks may hide text from AI crawlers.

Shopify

Strong defaults on schema and crawl access. Security headers and some structured data are platform-managed.

AI bot access

Shopify's default robots.txt permits all crawlers. The platform manages robots.txt; if you have a custom robots.txt liquid template, verify AI crawlers are not blocked.

Product and Organization schema

Shopify injects Product and Organization schema automatically for product and storefront pages. For additional schema types (HowTo, BlogPosting), use a schema app or manually add JSON-LD via theme code.

Blog and content pages

Shopify's native blog pages support Article-style content. Use the blog for substantive guides and how-to content. AI tools can crawl and cite well-structured Shopify blog posts.

Security header limitation

Shopify blocks custom HTTP security headers on standard plans. When Stackra flags a security header gap on Shopify, it is pointing at a platform ceiling, not something you can add without moving to Shopify Plus.

Common gaps

Five mistakes that hurt GEO readiness

Confusing AI training access with search indexing

Blocking AI training and blocking search indexing are different choices. Google-Extended controls whether Google can use your content for AI model training; it has no effect on Googlebot or standard search ranking. You can block AI training while keeping full search visibility, or allow both. They are independent.

Accidentally blocking content pages

Disallow: /path/ under User-agent: * blocks every crawler from that path, including AI bots. Blocking system paths like /wp-admin/ or /checkout/ is correct and expected. Blocking /blog/, /services/, or /about/ is a crawlability problem. Platforms generate the right path blocks automatically; the risk is custom additions made without realizing the scope.

Click-to-reveal content invisible to AI crawlers

FAQ sections and tabs that only show text after a JavaScript click event are invisible to AI fetchers. Use native HTML <details>/<summary> elements, or ensure the content is present in the initial HTML.

Schema present but mistyped

Subtypes like Restaurant or MedicalClinic are valid LocalBusiness subtypes, but an exact Organization or LocalBusiness type is what Stackra counts for entity presence. Validate your schema output with Google's Rich Results Test.

Entity information incomplete

Organization schema with only a name and URL leaves location and authorship unconfirmed. Each additional confirmed signal (location, named person) raises entity confidence. High confidence is the target for clear AI attribution.

Ignoring sitemap submission

Your sitemap lets AI crawlers discover pages they may not find through link crawling alone. Submit yours to Google Search Console and confirm it is referenced in your robots.txt. WordPress (5.5+), Wix, Shopify, and Squarespace generate sitemaps automatically; custom-built sites and some older setups do not.

Start here

Priority order for GEO readiness

01

Confirm your sitemap exists and is submitted in Google Search Console

WordPress (5.5+), Wix, Shopify, and Squarespace generate one automatically. Custom-built sites need manual creation.

02

Verify major AI crawlers are allowed in your robots.txt

GPTBot, OAI-SearchBot, ClaudeBot, and Google-Extended. Name them explicitly if possible.

03

Complete your business name, location, and description in your platform settings

WordPress (Rank Math Knowledge Graph), Wix (Settings > Business Info), Squarespace (Settings > Business Information). These auto-generate Organization or LocalBusiness schema.

04

Add Article or BlogPosting schema to content pages

Rank Math and Yoast do this automatically on WordPress. Wix applies it to blog posts natively. This is the primary citability signal.

05

Add HowTo schema to any step-by-step guides or service pages

Highest citability signal. Rank Math, Wix Structured Data tool, or Squarespace code injection.

06

Check that FAQ sections use HTML, not JavaScript click-reveal

Content that only appears after a user click is invisible to AI fetchers. Use native details/summary elements.

07

Run a Stackra scan to see your current GEO signal evidence

Bot access per crawler, schema detected, entity confidence level, and supporting signal status, all in one view.

Specific to Google AI Overviews

How to rank in AI Overviews

An AI Overview is Google's AI-generated summary that appears at the top of a Google Search results page on many informational queries, with citations to the source pages underneath. To be cited in an AI Overview, a page must be indexed by Google, accessible to Google-Extended in robots.txt, and structured to answer a specific question clearly enough that the AI can lift a sentence directly. The seven actions below are the concrete steps that move an existing page into AI Overview citation eligibility.

01

Allow Google-Extended in robots.txt

Google's AI training and Gemini crawler. If blocked, the page is excluded from AI Overview generation entirely. This is separate from Googlebot, which continues to handle classic search indexing.

02

Put a 60 to 80 word definition-style answer at the top of the page

AI Overviews lift the most concise, definition-style answer they can find. A clear short paragraph that directly answers the page's target question is the most-cited pattern.

03

Add FAQPage schema for any Q&A sections on the page

FAQPage schema makes Q&A pairs directly liftable. The AI can quote a question and answer almost verbatim into the Overview, with attribution to your URL.

04

Use specific named entities (products, people, places, tools) rather than generic terms

AI Overviews favor pages that name the same entities as the user's query. "Stackra audits 11 named AI crawlers" is more citable than "our tool checks AI bots".

05

Answer the questions Google lists under "People also ask" for your target query

Each related question answered on your page is a separate citation opportunity. One page can be cited in several different Overviews this way.

06

Confirm the page loads quickly and is mobile-readable

Google has stated Core Web Vitals and mobile usability are inputs to AI Overview source selection. Slow or broken pages get filtered out before AI generation.

07

Update both the visible date and the schema dateModified when content materially changes

AI Overviews tend to favor pages with recent freshness signals on time-sensitive topics. Update the schema's dateModified field, not just the displayed date.

AI Overview indexing is not instant. After updating a page, expect 1 to 7 days for Google to re-crawl an established site, then 2 to 6 weeks before the page typically appears in AI Overviews on its target queries.

Further reading

More on Generative Engine Optimization

Background, deep-dives, and platform-specific guidance for being cited by AI search.

See your GEO signals in under five minutes

Enter your URL and Stackra checks all four signal groups: bot access per crawler, schema by function, entity clarity confidence, and infrastructure readiness.