← Back to insights
Guide · #491

How to Audit Your Site for AI-Friendliness in 30 Minutes

Learn to audit your site for AI-friendliness in 30 minutes. Check markup, citations, and parseability. Step-by-step guide for founders.

Filed
April 2, 2026
Read
15 min
Author
The Seoable Team

The Problem Nobody's Talking About

Your site ranks in Google. Traffic looks fine. But ChatGPT can't find you. Perplexity doesn't cite you. Gemini doesn't know you exist.

This is the new SEO problem.

While you've been optimizing for Google's algorithm, AI engines have been building their own visibility requirements. They need different signals. Better structure. Cleaner markup. Parseability that goes beyond traditional on-page SEO.

The brutal truth: most sites fail the AI-friendliness test without knowing it. Missing schema. Broken citations. Unstructured content that AI models can't reliably extract. You're invisible to the tools your audience is using to find answers.

The good news? You can fix this in 30 minutes.

This guide walks you through a practical, founder-friendly audit that surfaces the markup gaps, citation blockers, and structural issues preventing AI engines from citing your brand. No agency required. No expensive tools. Just a step-by-step process that works.

What You'll Need Before You Start

Before diving into the audit, grab these free tools and have them ready:

  • Google Search Console — You need access to your domain. Set up Google Search Console in 10 minutes if you haven't already. This is non-negotiable.
  • Chrome DevTools — Built into Chrome. Press F12 or right-click → Inspect. You'll use this to read your page source and check markup.
  • A text editor — VS Code, Sublime, or even Notepad works. You'll copy-paste HTML snippets here to verify structure.
  • Lighthouse — Also built into Chrome DevTools. Run your first Lighthouse audit to get a baseline performance score.
  • Your robots.txt and sitemap — You should already have these. If not, generate a sitemap.xml and write your first robots.txt file before the audit.
  • A spreadsheet or Google Sheet — To log your findings and track fixes.

That's it. No credit card. No signup walls. Everything here is free.

Step 1: Run Your Site Through the AI-Friendliness Checklist (5 Minutes)

Start with the fundamentals. This step doesn't require tools—just your browser and your site.

Visit your homepage. Ask yourself these questions:

Is your site crawlable? Open Chrome DevTools (F12). Go to the Network tab. Reload the page. Do you see the HTML document load? If you see only JavaScript bundles and no HTML content, you have a rendering problem. AI engines struggle with JavaScript-heavy sites because they can't always wait for the JavaScript to execute. This is a blocker.

Do you have a robots.txt file? Type yoursite.com/robots.txt into your browser. You should see a text file with directives. If you get a 404, create one. Check out the robots.txt guide for a founder-friendly template. AI crawlers respect robots.txt just like Google does.

Do you have a sitemap? Type yoursite.com/sitemap.xml into your browser. You should see an XML file listing your pages. No sitemap? Generate one for your stack—whether you're on Next.js, Webflow, Shopify, or WordPress. Sitemaps help AI engines discover all your content efficiently.

Is your site mobile-friendly? Open your site on a phone. Can you read the text? Are buttons clickable? AI engines penalize mobile-unfriendly sites just like Google does. This matters more than people think.

Can you find your Organization schema? Inspect your homepage source (right-click → View Page Source, then press Ctrl+F and search for "schema.org/Organization"). You should find a JSON-LD block with your company name, logo, and contact info. If it's missing, add Organization schema in 5 minutes. This is the trust signal AI engines use to understand your brand.

Log your findings in your spreadsheet. Mark each as "Present" or "Missing."

Step 2: Check Your Page Structure and Markup (8 Minutes)

Now dig into the technical markup. This is where most sites fail the AI-friendliness test.

Open your homepage in Chrome DevTools. Press F12. Go to the Elements tab. You're looking at your HTML structure.

Check for proper heading hierarchy. Search for <h1> in the source. You should have exactly one H1 on your homepage. Then look for <h2>, <h3>, and so on. They should be in order—no skipping from H1 to H3. AI engines use heading structure to understand content hierarchy. Broken hierarchy confuses them.

Why does this matter for AI? When ChatGPT or Perplexity crawl your site, they parse the heading structure to understand what your page is about. A clear hierarchy signals that your content is well-organized and trustworthy.

Verify your meta description. Search for <meta name="description" in the source. You should see something like:

<meta name="description" content="Your description here">

Your meta description should be 150-160 characters and accurately describe the page. AI engines use this to understand page intent. A vague or missing meta description is a red flag.

Check for open graph tags. Search for <meta property="og:. You should find:

<meta property="og:title" content="...">
<meta property="og:description" content="...">
<meta property="og:image" content="...">

These tags help AI engines (and social platforms) understand your content's context. Missing OG tags make your content harder to cite accurately.

Look for canonical tags. Search for <link rel="canonical". You should see something like:

<link rel="canonical" href="https://yoursite.com/page">

Canonical tags tell AI engines which version of a page is the "official" one. Without them, AI might cite the wrong URL or get confused if you have duplicate content. Learn about canonicals and other common founder mistakes.

Verify your content is in semantic HTML. Look for <article>, <section>, <main>, and <nav> tags. Avoid wrapping everything in <div>. Semantic HTML helps AI engines understand what content is the main article versus navigation or sidebar fluff.

Log each finding. Mark as "Correct," "Needs Fix," or "Missing."

Step 3: Audit Your Schema Markup (7 Minutes)

Schema markup is the language AI engines speak. Without it, they're guessing at your content.

Check for Article schema. If you publish blog posts, search your article page source for "@type": "Article" or schema.org/Article. You should find a JSON-LD block that includes:

{
  "@context": "https://schema.org",
  "@type": "Article",
  "headline": "Your article title",
  "author": {
    "@type": "Person",
    "name": "Author name"
  },
  "datePublished": "2024-01-15",
  "image": "..."
}

AI engines use Article schema to extract the headline, author, publish date, and image. Missing this markup means they have to guess. Guessing leads to inaccurate citations.

Check for FAQPage schema. If you have an FAQ section, search for "@type": "FAQPage". This schema includes questions and answers in a structured format that AI engines can reliably parse. E-commerce sites should check for Product schema to ensure AI engines can cite your products accurately.

Verify your Organization schema. Go back to your homepage. Search for "@type": "Organization". This should include your company name, logo URL, contact email, and social profiles. Add this in 5 minutes if it's missing.

Check for LocalBusiness schema. If you're a local business, search for "@type": "LocalBusiness". This includes your address, phone number, hours, and service area. AI engines use this to understand your location and service range.

Use this AI-friendly websites checklist as a reference for the most critical schema types.

Log each schema type you find or are missing.

Step 4: Test Crawlability with Google Search Console (5 Minutes)

Google Search Console tells you exactly what Google can and can't crawl. AI engines face similar crawlability issues.

Go to Google Search Console. Set it up if you haven't.

Check the Coverage report. This shows you which pages Google can crawl and which it can't. Look for:

  • Excluded pages — Pages Google isn't indexing. Common reasons: noindex tag, blocked by robots.txt, redirect chain, or duplicate content.
  • Error pages — Pages that returned errors when Google tried to crawl them. 404s, 500s, or redirect loops.
  • Valid pages — Pages Google can crawl and index.

AI engines respect the same signals. If Google can't crawl a page, neither can ChatGPT or Perplexity. Fix the errors first.

Use URL Inspection. Pick one of your key pages. Paste the URL into URL Inspection. Click "Test live URL." Google will crawl it right now and show you what it sees. Check:

  • Does the page render correctly?
  • Are all resources (images, CSS, JavaScript) loading?
  • Are there any crawl errors?

If Google sees rendering issues, AI engines will too.

Check your indexing status. Look at the overall coverage summary. What percentage of your pages are indexed? If it's below 80%, you have a problem. Most of your site should be crawlable.

Log your coverage stats and any errors.

Step 5: Check Core Web Vitals and Page Performance (3 Minutes)

AI engines care about page speed. Slow sites get penalized.

Run PageSpeed Insights. Go to pagespeed.web.dev. Enter your homepage URL. Wait for the report.

You'll see three Core Web Vitals:

  • Largest Contentful Paint (LCP) — How fast does your main content load? Should be under 2.5 seconds.
  • Interaction to Next Paint (INP) — How responsive is your page to user input? Should be under 200 milliseconds.
  • Cumulative Layout Shift (CLS) — Does your page jump around while loading? Should be under 0.1.

AI engines crawl thousands of sites. Slow sites get deprioritized. If your Core Web Vitals are in the red, fix them. Learn how to read PageSpeed Insights and fix the three issues that actually move rankings.

Note your scores. Anything in the red is a blocker.

Step 6: Verify Indexing Status (2 Minutes)

You need to confirm Google has actually indexed your key pages. AI engines follow similar patterns.

Use the site: operator. Open Google. Type site:yoursite.com in the search bar. Hit Enter. Google will show you all indexed pages from your domain. Learn the three ways to verify Google indexing instantly.

If you see very few results, you have an indexing problem. Check your robots.txt and meta robots tags for noindex directives.

Check specific pages. Type site:yoursite.com/your-page-slug to verify a specific page is indexed. If it doesn't appear, it's either blocked or too new.

Log which pages are indexed and which aren't.

Step 7: Analyze Your Content for AI Parseability (5 Minutes)

Now look at how AI engines will actually read your content.

Pick your top 3 pages. These are the pages you want AI engines to cite. Usually your homepage, your main product page, and your top blog post.

Copy the main content. Inspect the page source. Find the <main> or <article> tag. Copy all the text content inside it (not navigation, not sidebar). Paste it into a text editor.

Ask yourself: Could an AI extract the key information? Read through the content. Is it clear what your main point is? Are facts separated into distinct paragraphs? Are statistics cited with sources?

AI engines extract information more reliably from:

  • Clear topic sentences — The first sentence of each paragraph should state the main idea.
  • Short paragraphs — Walls of text confuse AI. Keep paragraphs to 3-4 sentences.
  • Lists — Numbered and bulleted lists are easier to parse than prose.
  • Subheadings — They break up content and signal topic changes.
  • Cited facts — If you mention a statistic, link to the source. AI engines value sourced information.

Check for thin content. Pages under 300 words are hard for AI to extract meaningful information from. If you have thin pages, expand them or merge them with related content.

Note which pages need restructuring for better AI parseability.

Step 8: Check for Citation Blockers (3 Minutes)

This is the critical step most founders miss. Even if your content is good, AI engines might not cite you if these blockers exist.

Check your robots.txt for user-agent blocks. Open yoursite.com/robots.txt. Search for lines like:

User-agent: GPTBot
Disallow: /

If you see this, you're explicitly blocking ChatGPT from crawling your site. Same for other AI user-agents: CCBot (for Claude), Bingbot (for Copilot), Googlebot-Extended (for Google's AI services).

If you want AI engines to cite you, remove these blocks. Review your robots.txt setup.

Check your meta robots tag. Search your homepage source for <meta name="robots". You should NOT see noindex, nofollow, or noimageindex. These tags tell crawlers to ignore your site.

Check for X-Robots-Tag headers. This is less common but important. If your server sends X-Robots-Tag: noindex in the HTTP response, it blocks all crawlers. You can check this in Chrome DevTools Network tab → select the HTML document → look at Response Headers.

Verify your sitemap is accessible. AI crawlers use sitemaps to discover content efficiently. Make sure yoursite.com/sitemap.xml returns a 200 status code and is valid XML.

Check for nofollow on internal links. Inspect your homepage source. Search for rel="nofollow". If you see this on internal links, remove it. Nofollow tells crawlers not to follow those links. This blocks AI from discovering your deeper content.

Log any citation blockers you find. These are the highest-priority fixes.

Step 9: Use Free AI Audit Tools (5 Minutes)

Now cross-check your findings with free tools. This validates your manual audit.

Try Seoable's free check-up. Drop your domain at Seoable's free audit tool. You'll see instantly whether ChatGPT, Perplexity, Gemini, and Google can find your brand. No card required. No signup. Just results.

Use Screaming Frog (free tier). Download the free version. Crawl your site. It will show you:

  • Pages with missing meta descriptions
  • Broken internal links
  • Redirect chains
  • Duplicate content
  • Missing schema markup

Check with Lighthouse. Run your first Lighthouse audit in Chrome DevTools. Look at the SEO score. Anything below 90 means there are issues to fix.

Review the top 15 AI SEO audit tools. Check this comprehensive review of AI-powered SEO audit tools to see if there are specific tools that might catch issues your manual audit missed.

Compare these tool findings with your spreadsheet. Anything that shows up in multiple tools is a real problem.

Step 10: Compile Your Findings and Prioritize Fixes (2 Minutes)

You've got a full spreadsheet now. Time to prioritize.

Tier 1 (Fix immediately):

  • Missing or broken robots.txt
  • Noindex tags blocking your site
  • Missing sitemap
  • GPTBot or other AI user-agents explicitly blocked in robots.txt
  • Pages with no schema markup
  • Missing Organization schema
  • Broken internal links
  • Core Web Vitals in the red

Tier 2 (Fix this week):

  • Missing meta descriptions
  • Heading hierarchy issues
  • Missing OG tags
  • Thin content (under 300 words)
  • Pages not indexed by Google
  • Missing Article schema on blog posts

Tier 3 (Fix next sprint):

  • Content restructuring for better parseability
  • Adding more schema types (LocalBusiness, FAQPage, etc.)
  • Expanding thin content
  • Improving mobile responsiveness

Focus on Tier 1. These are the blockers preventing AI engines from finding and citing you.

Pro Tips for Founder Audits

Automate where you can. Use the SEO Pro Extension for on-page audits. Install it, run your first audit in under 5 minutes, and get a free checklist. It catches common issues faster than manual inspection.

Don't get lost in perfection. You don't need every schema type. Focus on the ones that matter for your industry. A SaaS company needs Article and Organization schema. An e-commerce site needs Product and FAQPage schema. A local business needs LocalBusiness schema.

Test after every fix. When you fix something, re-run Lighthouse and PageSpeed Insights. Verify the change actually improved your score. Sometimes fixes break other things.

Check your competitors. Inspect your top competitor's homepage source. What schema do they have? What's their meta description? What's their heading structure? You don't need to copy them, but understanding what works in your space helps.

Set up monitoring. Once you fix your issues, use Google Search Console to monitor your indexing status weekly. Learn how to read the GSC Performance report like a founder. This tells you if your fixes are working.

Common Mistakes Founders Make

Blocking AI crawlers intentionally. Some founders add GPTBot to robots.txt to "protect" their content. This backfires. You can't stop AI from training on your content. But you can control whether they cite you as a source. If you block them, they can't cite you. Remove the blocks.

Ignoring Core Web Vitals. Founders often think page speed doesn't matter if Google ranks them. Wrong. AI engines are increasingly using speed as a ranking signal. A slow site gets deprioritized in AI search results.

Skipping schema markup. Schema feels optional. It's not. Without it, AI engines have to guess at your content. They guess wrong. Schema is the difference between being cited accurately and being misrepresented.

Forgetting to update your robots.txt and sitemap. You add a new page. You forget to update your sitemap. Six months later, you wonder why it's not indexed. Keep your sitemap current. It's the roadmap AI engines use to find your content.

Writing for humans only. Your content needs to work for both humans and AI. Clear headings, short paragraphs, semantic HTML, and proper schema aren't just nice to have. They're table stakes in 2024.

What Happens After the Audit

You've got your findings. You know what's broken. Now what?

Fix Tier 1 issues. This should take a few hours, not days. Most of these are quick wins—adding schema, fixing robots.txt, unblocking AI crawlers.

Re-run the audit in one week. After you fix Tier 1, run through this audit again. You should see improvement in your tool scores and indexing status.

Start thinking about Tier 2. While Tier 1 fixes are live, work on content and meta descriptions. These take longer but have high impact.

Monitor continuously. Use Google Search Console to track your progress. Learn how URL Inspection works—it's the on-demand audit most founders forget exists. Run it weekly on your top pages.

Consider a full SEO roadmap. This 30-minute audit is a snapshot. For sustained growth, you need a keyword roadmap and content strategy. But you can't build that until you fix the foundation. Fix the foundation first.

The Bottom Line

AI-friendliness isn't complicated. It's just discipline.

Clean markup. Proper schema. No blockers. Fast pages. Semantic HTML. Parseability.

These are the fundamentals that let AI engines find, understand, and cite your brand.

Most sites fail because founders don't know these rules exist. Now you do.

Spend 30 minutes running this audit. Log your findings. Fix the blockers. Re-test.

Then watch as ChatGPT, Perplexity, and Gemini start citing you as a source.

That's the new SEO game. Ship or stay invisible.

You know what to do now.

Free weekly newsletter

Get the next one on Sunday.

One short email a week. What is working in SEO right now. Unsubscribe in one click.

Subscribe on Substack →
Keep reading