Technical SEO for AI
A guide to optimizing your technical infrastructure for AI assistant discoverability and citability.
Technical SEO for AI
Technical SEO for AI ensures that AI assistants like ChatGPT, Claude, and Perplexity can discover, access, and accurately cite your content. This guide covers the technical requirements and best practices.
Core Requirements
Before optimizing for AI, ensure these fundamental requirements are met:
1. Server-Side Rendering (SSR)
Why it matters: AI assistants cannot execute JavaScript. Your content must be in the initial HTML response.
Check if you have SSR:
If you can find your content in the raw HTML, you have SSR. If not, see Is Your Site Visible to AI Assistants?
Solutions by framework:
- React: Next.js, Remix, or React Server Components
- Vue: Nuxt.js
- Angular: Angular Universal
- Svelte: SvelteKit
2. Allow AI Crawlers
Why it matters: AI assistants respect robots.txt. If you block them, they can't access your content.
Check your robots.txt:
Allow these user agents:
ChatGPT-User(ChatGPT)Claude-User(Claude)Perplexity-User(Perplexity)Google-Extended(Gemini)
See robots.txt Configuration for details.
3. Fast Load Times
Why it matters: AI crawlers have timeout limits. Slow pages may be abandoned before content loads.
Target metrics:
- Time to First Byte (TTFB): < 600ms
- First Contentful Paint (FCP): < 1.8s
- Largest Contentful Paint (LCP): < 2.5s
Quick wins:
- Enable compression (gzip/brotli)
- Optimize images (WebP, proper sizing)
- Minimize render-blocking resources
- Use CDN for static assets
4. Clean HTML Structure
Why it matters: AI assistants parse HTML to understand content hierarchy and meaning.
Best practices:
- Use semantic HTML5 elements (
<article>,<section>,<header>) - Proper heading hierarchy (h1 → h2 → h3)
- Descriptive link text (not "click here")
- Alt text for all images
- Clear content structure
Structured Data Implementation
Structured data (schema.org markup) helps AI assistants understand your content context.
Priority: Organization Schema
Add to your homepage to define your brand:
What to include:
- Exact brand name (as you want AI to use it)
- Primary domain URL
- High-quality logo (1200x630px recommended)
- One-sentence description of your category
- Social media profiles
Priority: Product/Service Schema
Add to product or service pages:
What to include:
- Product/service name
- Clear category classification
- Current pricing (if public)
- Key features list
- Ratings/reviews if available
Priority: FAQ Schema
Add to pages with Q&A content:
Best practices:
- 50-70 word answers
- Educational tone (not promotional)
- Plain text (no HTML in answer text)
- Natural question phrasing
Other Useful Schema Types
Article Schema (for blog posts):
Breadcrumb Schema (for navigation):
Local Business Schema (for location-based businesses):
Content Formatting for AI
How you format content affects AI comprehension and citability.
Use Semantic HTML
AI assistants understand HTML semantics:
Heading Hierarchy
Use proper heading levels:
Lists and Tables
Use structured elements for organized information:
Descriptive Links
Use meaningful link text:
Image Alt Text
Provide context for images:
Content Writing for AI Citability
Clear, Concise Language
AI assistants prefer straightforward content:
Good:
"Spyglasses analyzes how AI assistants like ChatGPT discover and cite your content. Our reports show your brand consistency score, competitive positioning, and technical accessibility."
Avoid:
"Spyglasses is a revolutionary, cutting-edge solution that empowers businesses to leverage next-generation AI-powered insights for optimizing their digital presence in the emerging landscape of artificial intelligence."
Answer Questions Directly
Structure content around common questions:
Specific Over General
Provide concrete information:
Good:
"Plans start at $49/month for up to 3 properties. Enterprise plans available for larger teams."
Avoid:
"Flexible pricing to fit your needs. Contact us to learn more."
Update Dates and Version Info
Help AI understand content freshness:
Technical Best Practices
robots.txt Configuration
Allow AI crawlers while protecting sensitive areas:
See robots.txt Configuration for complete guide.
Sitemap Optimization
Help AI crawlers discover your content:
Best practices:
- Include all important pages
- Keep lastmod dates current
- Set appropriate priorities
- Update when content changes
- Submit to search engines
Canonical URLs
Prevent duplicate content issues:
Use canonical tags when:
- Same content exists on multiple URLs
- You have pagination
- You have parameter-based URLs
- You syndicate content
Meta Tags
While less important than structured data, good meta tags help:
Content Security and Access
Ensure AI assistants can access content:
Don't:
- Hide content behind login walls (for public pages)
- Use aggressive anti-bot measures
- Require JavaScript for content visibility
- Use CAPTCHA to protect entire public pages
- Rate-limit too aggressively
Do:
- Use authentication only for truly private content
- Allow reasonable crawler access
- Implement server-side rendering
- Trust verified bot user agents
- Monitor crawler activity
Performance Optimization
Image Optimization
Optimize images without sacrificing AI understanding:
Best practices:
- WebP format for modern browsers
- Proper dimensions (no oversized images)
- Lazy loading for below-the-fold content
- Descriptive alt text (critical for AI)
- Compress without losing clarity
Font Loading
Optimize fonts to improve load times:
Or use system fonts:
Minimize JavaScript
Reduce JavaScript that blocks content:
Good:
- Defer non-critical JS
- Async load third-party scripts
- Code splitting
- Server-side rendering
Avoid:
- Blocking scripts in
<head> - Large JavaScript bundles
- Unnecessary third-party scripts
- Client-only rendering
Monitoring and Testing
Test AI Accessibility
Verify AI assistants can access your content:
Validate Structured Data
Use Google's testing tools:
Regular Audits
Schedule quarterly reviews:
- Verify AI crawlers can access key pages
- Test structured data implementation
- Check brand consistency across pages
- Measure page load times
- Review robots.txt configuration
- Update stale content
- Fix broken links
- Validate schema markup
Common Issues and Fixes
Issue: Content Not Accessible
Symptoms:
- AI assistants don't cite your pages
- curl shows empty or minimal HTML
Fixes:
- Implement server-side rendering
- Remove JavaScript dependency for core content
- Pre-render static pages
Issue: Cloudflare Blocking
Symptoms:
- Curl returns challenge page
- AI crawlers blocked in logs
Fixes:
- Whitelist verified bots in Cloudflare
- Adjust Bot Fight Mode settings
- Create custom rules for AI crawlers
Issue: robots.txt Blocking
Symptoms:
- robots.txt disallows AI bots
- Zero AI crawler traffic
Fixes:
- Allow specific AI bot user agents
- Remove wildcard disallow rules
- Keep only necessary restrictions
See robots.txt Guide
Issue: Slow Load Times
Symptoms:
- AI assistants timeout
- Partial content in responses
Fixes:
- Optimize images and assets
- Enable compression
- Use CDN for static files
- Minimize render-blocking resources
Issue: Inconsistent Brand Info
Symptoms:
- AI describes your brand incorrectly
- Outdated information in AI responses
Fixes:
- Audit all pages for consistency
- Update product descriptions
- Ensure pricing is current
- Add/update Organization schema
Next Steps
1. Run an AI Visibility Report
Get your free AI Visibility Report to see:
- Technical accessibility issues
- Brand consistency score
- Structured data coverage
- Competitive positioning
2. Implement Priority Fixes
Start with highest-impact changes:
- Fix technical blockers (SSR, robots.txt, Cloudflare)
- Add Organization schema to homepage
- Add Product/Service schema to key pages
- Optimize page load times
3. Content Optimization
Improve AI understanding:
- Audit brand consistency
- Add FAQ schema
- Improve content structure
- Update stale information
4. Monitor and Iterate
Track progress over time:
- Set up crawler monitoring
- Test key queries monthly
- Run quarterly AI visibility reports
- Measure traffic from AI referrers