Server data from the Official MCP Registry
Audit web pages for AI search readiness (ChatGPT, Claude, Perplexity, Gemini).
Audit web pages for AI search readiness (ChatGPT, Claude, Perplexity, Gemini).
Valid MCP server (4 strong, 3 medium validity signals). 1 code issue detected. No known CVEs in dependencies. Package registry verified. Imported from the Official MCP Registry. 1 finding(s) downgraded by scanner intelligence.
10 files analyzed · 2 issues found
Security scores are indicators to help you make informed decisions, not guarantees. Always review permissions before connecting any MCP server.
Add this to your MCP configuration file:
{
"mcpServers": {
"io-github-agencyenterprise-aiseo-audit": {
"args": [
"-y",
"aiseo-audit"
],
"command": "npx"
}
}
}From the project's GitHub README.
Deterministic CLI that audits web pages for AI search readiness. Think Lighthouse, but for how well AI engines can fetch, extract, understand, and cite your content.
AI SEO measures how reusable your content is for generative engines, not traditional search rankings.
Who is this for? Content teams running pre-publish checks, developers gating deployments in CI/CD, and marketers auditing their own or competitor pages. If your content needs to be cited (not just ranked), this tool tells you where you stand.
Traditional SEO optimizes for ranking in a list of links. AI SEO optimizes for being cited in generated answers. Different goal, different signals.
When someone asks ChatGPT, Claude, Perplexity, or Gemini a question, those engines fetch web content, extract the useful parts, and decide what to cite. AI SEO (also called Generative Engine Optimization or GEO) is the practice of structuring your content so that process works in your favor. The foundational research behind this field comes from Princeton's GEO paper, which identified the specific content traits that increase generative engine citations.
aiseo-audit measures those signals: can the content be extracted? Is it structured for reuse? Does it contain the patterns AI engines actually quote? It runs entirely locally with no AI API calls and no external services.
Most "AI readiness" audits check whether certain files and tags exist. Does the site have llms.txt? Is there a sitemap? Is JSON-LD present? Those are binary checks that tell you very little about whether AI engines will actually use your content.
aiseo-audit goes deeper:
aiseo.config.json. Zero vendor lock-in.# Try it instantly, no install required
npx aiseo-audit https://yoursite.com
# As a project dependency
npm install aiseo-audit
# As a dev dependency
npm install --save-dev aiseo-audit
# Globally
npm install -g aiseo-audit
# Pretty terminal output (default)
aiseo-audit https://example.com
# Quick summary only (top 3 wins + projected score, skip the full breakdown)
aiseo-audit https://example.com --tldr
# Track score changes over time
aiseo-audit https://example.com --diff
# JSON output
aiseo-audit https://example.com --json
# Markdown output
aiseo-audit https://example.com --md
# HTML report (Lighthouse-style)
aiseo-audit https://example.com --html
# Write to a file, format is inferred from the extension automatically
aiseo-audit https://example.com --out report.html
aiseo-audit https://example.com --out report.md
aiseo-audit https://example.com --out report.json
# Explicit format flag still works and takes precedence
aiseo-audit https://example.com --html --out report.html
# CI/CD: fail if score below threshold
aiseo-audit https://example.com --fail-under 70
# Custom timeout
aiseo-audit https://example.com --timeout 30000
# Custom user agent
aiseo-audit https://example.com --user-agent "MyBot/1.0"
# Use config file
aiseo-audit https://example.com --config aiseo.config.json
Use --tldr to get just the top 3 highest-impact fixes and the projected score after applying them. No detailed category breakdown, no long recommendations list. Ideal for CI logs, Slack notifications, and quick pre-publish checks.
aiseo-audit https://example.com --tldr
============================================================
AI SEO Audit
https://example.com
============================================================
Score: 59/100 Grade: F → Top 3 fixes: ~87/100 B+
Quickest wins:
1. +13 pts Answer Capsules (Answerability)
2. +10 pts Author Attribution (Authority Context)
3. +7 pts Image Alt Text (Content Extractability)
--tldr works with every output format (--json, --md, --html) so you can pipe it into any integration. Combine with --out to write a slim summary to a file.
--diff records every audit and shows you what changed since the last run for the same URL.
[!IMPORTANT]
--diffis the only flag that writes files outside of--out. On first use it creates./audits/(or your configuredhistoryDir) and anaiseo.config.jsonin the current directory if one doesn't exist. Both are announced on stderr the moment they're created — nothing is written silently. Add./audits/to.gitignoreif you don't want the history tracked in version control, or commit it to keep a record of AI SEO over time.
# First run — establishes a baseline
aiseo-audit https://example.com --diff
# Subsequent runs — shows the delta
aiseo-audit https://example.com --diff
Output on the second run:
Changes since baseline (2026-04-10 → 2026-04-17)
Overall 59 → 68 (+9)
Answerability 18 → 28 (+10)
Grounding Signals 23 → 26 (+3)
How it works. Each --diff run writes a full JSON audit to ./audits/<slug>-<timestamp>.json and appends an entry to aiseo.config.json under a diff key. Nothing is ever overwritten — every run is preserved so you can walk the history.
{
"historyDir": "./audits",
"diff": {
"https://example.com": [
{
"path": "audits/example-2026-04-10.json",
"timestamp": "2026-04-10T…",
"score": 59
},
{
"path": "audits/example-2026-04-17.json",
"timestamp": "2026-04-17T…",
"score": 68
}
]
}
}
--diff always writes the JSON baseline to historyDir automatically. If you also want a human-friendly report (HTML, Markdown, pretty), pass --out with a file path and the rendered report will be written there:
# Creates two files: the auto-named JSON baseline in ./audits/,
# and a rendered HTML report at ./audits/report.html.
aiseo-audit https://example.com --diff --html --out ./audits/report.html
--out never replaces the baseline. It's purely for the rendered report. If --out points at an existing directory, the audit exits early with a clear error rather than silently writing an invalid file.
Drop the URL and pass --diff --all to see every tracked URL at a glance with a sparkline of its history:
aiseo-audit --diff --all
Audit History (2 URLs tracked, 7 total runs)
https://example.com ▂▅▇▆ 59 → 68 → 72 → 70
https://another-site.com ▇▇ 72 → 72
Works with --html for a shareable timeline page with inline SVG line charts.
If you want to compare against a specific prior JSON result without touching the tracked history, use --baseline:
aiseo-audit https://example.com --baseline ./previous-audit.json
aiseo-audit ships an MCP server so Cursor, Claude Desktop, Windsurf, and any other MCP client can call the audit_url tool inline in a chat.
No install required — the server runs via npx from the published package.
Cursor (~/.cursor/mcp.json):
{
"mcpServers": {
"aiseo-audit": {
"command": "npx",
"args": ["-y", "aiseo-audit-mcp"]
}
}
}
Claude Desktop (~/Library/Application Support/Claude/claude_desktop_config.json on macOS):
{
"mcpServers": {
"aiseo-audit": {
"command": "npx",
"args": ["-y", "aiseo-audit-mcp"]
}
}
}
Once configured, prompt your assistant naturally — e.g. "audit https://mysite.com for AI search readiness" — and it will invoke the tool and return a full audit inline.
aiseo-audit ships an official GitHub Action. Drop it into any workflow to gate PRs on AI search readiness and post a sticky PR comment with the summary:
# .github/workflows/aiseo-audit.yml
name: AI SEO Audit
on:
pull_request:
push:
branches: [main]
jobs:
audit:
runs-on: ubuntu-latest
permissions:
pull-requests: write
steps:
- uses: agencyenterprise/aiseo-audit@v1
with:
url: https://yoursite.com
fail-under: 70
comment-on-pr: true
With comment-on-pr: true, the Action updates a single sticky comment on each PR with the TL;DR score, grade, and top fixes — so reviewers see the impact of content changes inline.
Prefer a one-liner without the Action wrapper? This still works:
steps:
- run: npx aiseo-audit https://yoursite.com --fail-under 70
Using a preview deployment URL? If your CI pipeline produces a dynamic URL (e.g. a Vercel or Netlify preview), capture it from a prior step and pass it in:
jobs:
deploy-and-audit:
runs-on: ubuntu-latest
steps:
- name: Deploy to preview
id: deploy
run: echo "url=https://your-preview-url.vercel.app" >> $GITHUB_OUTPUT
# Replace the above with your actual deploy step that outputs a URL
- name: Run AI SEO Audit
run: npx aiseo-audit ${{ steps.deploy.outputs.url }} --fail-under 70
The --fail-under threshold sets the minimum acceptable score. Exit code 1 is returned when the score falls below it, which GitHub Actions treats as a failed step.
| Option | Description | Default |
|---|---|---|
[url] | URL to audit | - |
--sitemap <url> | Audit all URLs in a sitemap.xml | - |
--signals-base <url> | Base URL to fetch domain signals from (robots.txt, llms.txt, llms-full.txt) | Directory of the URL or sitemap being audited |
--json | Output as JSON | - |
--md | Output as Markdown | - |
--html | Output as HTML | - |
--out <path> | Write output to a file; format is inferred from .html, .md, .json extension | - |
--fail-under <n> | Exit with code 1 if score < threshold | - |
--timeout <ms> | Request timeout in ms | 45000 |
--user-agent <ua> | Custom User-Agent string | AISEOAudit/<version> |
--config <path> | Path to config file | - |
--tldr | Emit only the TL;DR summary (no detailed breakdown) | - |
--diff | Track score over time: record this run, diff against the previous recorded run | - |
--all | With --diff and no URL, render audit history across all tracked URLs | - |
--baseline <path> | Diff against a specific prior JSON result (bypasses history tracking) | - |
Either [url] or --sitemap must be provided (or --diff --all for the cross-URL timeline). If no output flag is given, the default is pretty (color-coded terminal output). The default format can also be set in the config file.
When --out is provided, the format is automatically inferred from the file extension (.html becomes HTML, .md becomes Markdown, .json becomes JSON) so you don't need to pass a separate format flag. An explicit --html, --md, or --json flag takes precedence if provided.
Use --sitemap to audit every URL in a sitemap.xml. Domain signals (robots.txt, llms.txt, llms-full.txt) are fetched once and shared across all URL audits, not re-fetched per page.
By default, domain signals are fetched from the directory that contains the sitemap file. For example, if your sitemap is at https://example.com/projects/sitemap.xml, signals are fetched from https://example.com/projects/ so the tool checks https://example.com/projects/robots.txt, https://example.com/projects/llms.txt, and https://example.com/projects/llms-full.txt. If your signals live at the domain root instead, use --signals-base to specify the correct location explicitly.
# Audit all URLs in a sitemap
aiseo-audit --sitemap https://example.com/sitemap.xml
# With HTML output, format inferred from extension
aiseo-audit --sitemap https://example.com/sitemap.xml --out report.html
# Override where domain signals are fetched from (sitemap)
aiseo-audit --sitemap https://example.com/projects/sitemap.xml --signals-base https://example.com
# Override where domain signals are fetched from (single URL)
aiseo-audit https://example.com/projects/page --signals-base https://example.com
# Fail if average score across all URLs is below threshold
aiseo-audit --sitemap https://example.com/sitemap.xml --fail-under 70
The sitemap report includes:
Sitemap index files (sitemaps that reference other sitemaps) are supported, and all child sitemaps are fetched and flattened automatically. Every report format explicitly shows which URL domain signals were fetched from, so there is no guesswork about where robots.txt, llms.txt, and llms-full.txt were checked.
By default, all HTTP requests (page fetch, robots.txt, llms.txt) are sent with the header User-Agent: AISEOAudit/<version>. This is intentional. If a site blocks unknown bots, that is a meaningful negative signal for AI search readiness, and the audit should surface it as a failing "Fetch Success" score.
The --user-agent flag exists as an escape hatch for cases where you want to bypass bot detection and test the content independently of access policy. It does not change the audit logic, only what the server sees in the request header.
You can run the audit against a local dev server to iterate on your content before deploying:
aiseo-audit http://localhost:3000
The page analysis (content structure, readability, schema markup, answerability, etc.) works identically to a production audit. These factors depend on your HTML output, which is the same locally as it is in production.
The audit also checks for three domain-level files that AI crawlers look for:
robots.txt controls which bots can access your site. AI crawlers (GPTBot, ClaudeBot, etc.) respect this file to determine whether they are allowed to fetch your content.llms.txt is a proposed standard that provides LLMs with a concise summary of your site's purpose, key pages, and preferred citation format.llms-full.txt is the extended version of llms.txt with more comprehensive site documentation.When auditing over HTTP, these files are checked against your local server. If your local server serves them, they will pass. If not, they will show as missing.
[!NOTE] Local audit results may differ slightly from production. Domain signal files (
robots.txt,llms.txt,llms-full.txt) are often configured at the hosting or CDN level and may not be present on your local dev server. Always verify these signals separately against your production domain.
[!NOTE] AI SEO (e.g. GEO: Generative Engine Optimization) is a new and developing field. Methodologies and audit criteria may evolve as research and engine behavior advance.
The audit evaluates 7 categories of AI search readiness (Detailed Breakdown here):
| Category | What It Measures |
|---|---|
| Content Extractability | Can AI engines successfully fetch and extract meaningful text from the page? |
| Content Structure for Reuse | Is the content organized with headings, lists, and tables that engines can segment? |
| Answerability | Does the content provide clear definitions, direct answers, and step-by-step patterns? |
| Entity Clarity | Are named entities (people, orgs, places) clearly present and consistent with the topic? |
| Grounding Signals | Does the content cite external sources, include statistics, and attribute claims? |
| Authority Context | Is there author attribution, organization identity, publish dates, and structured data? |
| Readability for Compression | Is the content written at a readability level that compresses well for AI summarization? |
Grades follow a standard US academic scale with +/- sub-grades for granular tracking across iterations.
| Score | Grade | What It Means |
|---|---|---|
| 93–100 | A | Highly optimized. AI engines can fetch, understand, and cite your content with minimal friction. |
| 90–92 | A- | Near top tier. A handful of polish items stand between you and an A. |
| 87–89 | B+ | Strong foundation with a few high-impact gaps worth closing. |
| 83–86 | B | Good foundation. Targeted improvements will push you into the top tier. |
| 80–82 | B- | Solid, but several category gaps are holding back citation potential. |
| 77–79 | C+ | Above average. Structural or content gaps are limiting citation potential. |
| 73–76 | C | Moderate readiness. Multiple categories need attention. |
| 70–72 | C- | Below the comfort zone. Prioritize high-impact recommendations. |
| 67–69 | D+ | Noticeable gaps across core signals. |
| 63–66 | D | Significant gaps. Core signals (structure, answerability, authority) need attention. |
| 60–62 | D- | On the edge of failing. Most categories need meaningful work. |
| 0–59 | F | Not AI-ready. Start with Content Extractability. If the content cannot be fetched, nothing else matters. |
The per-category breakdown in each report shows exactly where to focus. Start with high-priority recommendations in your lowest-scoring categories.
Color-coded terminal output with scores, factor breakdowns, and recommendations. Where available, recommendations include numbered implementation steps, a ready-to-use code example, and a learn more link. Best for quick checks during development.
Full structured output with all scores, factor details, raw data, and recommendations. Recommendations include optional steps, codeExample, and learnMoreUrl fields where applicable. Best for integrations, CI/CD pipelines, and programmatic consumption.
Structured report with category tables, factor details, and recommendations grouped by category. Recommendations with steps render as numbered lists; code examples render as fenced code blocks. Best for documentation, PRs, and sharing.
Self-contained single-file report with SVG score gauges, color-coded sections, and recommendations grouped by category. Recommendations with steps and code examples render as inline detail sections below each recommendation row. Best for stakeholder reports and visual review.
aiseo-audit https://example.com --out report.html
Tip: Run this against your own site and open the file in a browser to get the most actionable view of where to focus. The HTML report is the closest equivalent to Lighthouse's output.
Create a config file in your project root to customize behavior. The CLI automatically discovers your config by searching from the current directory up to the filesystem root, looking for (in order):
aiseo.config.json.aiseo.config.jsonaiseo-audit.config.jsonYou can also pass an explicit path with --config path/to/config.json.
{
"timeout": 45000,
"format": "pretty",
"failUnder": 50,
"weights": {
"contentExtractability": 1,
"contentStructure": 1,
"answerability": 1,
"entityClarity": 1,
"groundingSignals": 1,
"authorityContext": 1,
"readabilityForCompression": 1
},
"historyDir": "./audits",
"diff": {}
}
Weights are relative. Set a category to 2 to double its importance, or 0 to exclude it entirely from scoring.
historyDir controls where --diff writes audit JSONs (default ./audits). diff is managed by the tool — each --diff run appends an entry keyed by URL. See Tracking AI SEO Over Time.
Example: tuning for a blog or editorial site. Content that needs to be cited by AI engines should be highly answerable and readable. Double those weights and reduce the emphasis on domain-signal files (authorityContext covers schema markup, while groundingSignals covers citations and statistics):
{
"format": "html",
"failUnder": 65,
"weights": {
"contentExtractability": 2,
"contentStructure": 1,
"answerability": 2,
"entityClarity": 1,
"groundingSignals": 2,
"authorityContext": 1,
"readabilityForCompression": 2
}
}
Example: tuning for a product or docs site. Authority and structure matter more when content needs to be trusted and navigable:
{
"format": "html",
"failUnder": 70,
"weights": {
"contentExtractability": 2,
"contentStructure": 2,
"answerability": 1,
"entityClarity": 1,
"groundingSignals": 1,
"authorityContext": 2,
"readabilityForCompression": 1
}
}
import { analyzeUrl, loadConfig, renderReport } from "aiseo-audit";
const config = await loadConfig();
const result = await analyzeUrl({ url: "https://example.com" }, config);
console.log(result.overallScore); // 72
console.log(result.grade); // "B-"
// Render in any format
const html = renderReport(result, { format: "html" });
const md = renderReport(result, { format: "md" });
const json = renderReport(result, { format: "json" });
import type {
AnalyzerResultType,
AnalyzerOptionsType,
AuditResultType,
CategoryNameType,
CategoryResultType,
CategoryDeltaType,
DiffEntryType,
DiffResultType,
FactorResultType,
RecommendationType,
ReportFormatType,
AiseoConfigType,
} from "aiseo-audit";
RecommendationType includes the base fields (category, factor, currentValue, priority, recommendation) plus optional fields populated for high-impact factors:
type RecommendationType = {
category: string;
factor: string;
currentValue: string;
priority: "high" | "medium" | "low";
recommendation: string;
expectedGain?: number; // points you'd recover by resolving this factor
steps?: string[]; // ordered implementation steps
codeExample?: string; // ready-to-use code snippet
learnMoreUrl?: string; // link to canonical spec or docs
};
computeDiff(current, baseline) compares two AnalyzerResultType values and returns a DiffResultType with the overall delta and per-category deltas. Pair with renderDiffReport(result, diff, { format }) to render the diff in any format, or renderHistoryTimeline(diffMap, { format }) to render a cross-URL timeline from a tracked history map.
This tool measures AI search reusability: how well a page's content can be fetched, extracted, understood, and reused by AI engines like ChatGPT, Claude, Perplexity, and Gemini.
It is:
| Code | Meaning |
|---|---|
0 | Success |
1 | Score below --fail-under threshold |
2 | Runtime error (fetch failed, invalid URL, etc.) |
Node.js - Requires Node 20 or later. The engines field in package.json enforces this. Earlier versions will produce runtime errors.
Zod - Uses Zod 4. If you consume the library API and also use Zod in your project, ensure you are on Zod 4+ to avoid type incompatibilities.
CJS bin entries - Both bin/aiseo-audit.js (CLI) and bin/aiseo-audit-mcp.js (MCP server) use require() (CommonJS). This is compatible with all Node 20+ environments regardless of your project's module system. The library exports support both ESM (import) and CJS (require).
Config discovery - When using the programmatic API, loadConfig() searches for config files starting from process.cwd(). If your application's working directory differs from where your config file lives, pass an explicit path:
const config = await loadConfig("/path/to/aiseo.config.json");
See CONTRIBUTING.md for development setup, project structure, and pull request guidelines.
Release notes are published on the GitHub Releases page. A separate CHANGELOG.md is not maintained.
MIT
Be the first to review this server!
by Modelcontextprotocol · Developer Tools
Read, search, and manipulate Git repositories programmatically
by Modelcontextprotocol · Developer Tools
Web content fetching and conversion for efficient LLM usage
by Toleno · Developer Tools
Toleno Network MCP Server — Manage your Toleno mining account with Claude AI using natural language.
by mcp-marketplace · Developer Tools
Create, build, and publish Python MCP servers to PyPI — conversationally.