From overgrow
Audits existing or generates sitemap.xml, sitemap indexes, and robots.txt per sitemaps.org protocol and Google Search Central guidance, allowing all crawlers including AI bots by default.
npx claudepluginhub zhizdev/overgrow --plugin overgrowThis skill is limited to using the following tools:
This skill is the deterministic pipeline for producing and auditing `sitemap.xml`, sitemap indexes, and `robots.txt`. It is tuned for growth — by default it invites every compliant crawler, including AI search and AI training bots, unless the user explicitly opts out per user-agent.
Guides Next.js Cache Components and Partial Prerendering (PPR): 'use cache' directives, cacheLife(), cacheTag(), revalidateTag() for caching, invalidation, static/dynamic optimization. Auto-activates on cacheComponents: true.
Processes PDFs: extracts text/tables/images, merges/splits/rotates pages, adds watermarks, creates/fills forms, encrypts/decrypts, OCRs scans. Activates on PDF mentions or output requests.
Share bugs, ideas, or general feedback.
This skill is the deterministic pipeline for producing and auditing sitemap.xml, sitemap indexes, and robots.txt. It is tuned for growth — by default it invites every compliant crawler, including AI search and AI training bots, unless the user explicitly opts out per user-agent.
Read knowledge/sitemap.md from the plugin root before doing anything. That file is the canonical, rule-by-rule pipeline this skill implements. Follow it exactly. This SKILL.md is the entry-point summary; knowledge/sitemap.md is the source of truth.
Also consult knowledge/geo.md sections on technical SEO and indexability for AI visibility considerations that affect sitemap/robots decisions.
Parse $ARGUMENTS:
audit (or no args, existing sitemap present) → run the audit flow.build (or no args, no sitemap present) → run the generation flow.https://example.com → use as the canonical origin.If mode is ambiguous, STOP and call the AskUserQuestion tool to clarify.
.overgrow/inventory.md for the authoritative page list. If missing, run init first.public/sitemap.xml, static/sitemap.xml, sitemap.xml.ts, Next.js app/sitemap.ts, framework-generated routes. If none found, switch to build mode.robots.txt: public/robots.txt, static/robots.txt, app/robots.ts, or config-driven.knowledge/sitemap.md. Report findings:
<loc>, bad encoding, over 50k URLs in one file, over 50MB uncompressed).lastmod hygiene (missing, future-dated, identical across all entries)..overgrow/sitemap-audit.md with severity (critical / high / medium / low) and specific fixes..overgrow/inventory.md for the list of indexable pages. Exclude utility and anything the user flags.knowledge/sitemap.md § "Inputs to collect"). If any input is missing, ask once in a single message.sitemap.xml following the rules in knowledge/sitemap.md. For > 50k URLs, emit a sitemap index plus shard files.robots.txt. Default posture: open (allow all user-agents, disallow explicit exclusion paths, reference the sitemap). Only switch to restricted or closed when the user says so — typically for staging, internal, or compliance-locked sites.public/, static/, framework-specific path). Do not overwrite existing files without asking; produce .next/.proposed variants if there's an existing file.Unless the user says otherwise, robots.txt explicitly allows:
GPTBot (OpenAI)ChatGPT-UserOAI-SearchBotClaudeBot / anthropic-ai / Claude-WebPerplexityBotGoogle-ExtendedApplebot-ExtendedBingbot (always)CCBot (Common Crawl)Explicit allow lines beat silence — spell them out so the posture is auditable. If the user wants to restrict one or more, ask for the list and write Disallow: / blocks per user-agent.
knowledge/sitemap.md for encoding, escaping, lastmod formatting, and sharding rules. Do not improvise.hreflang beyond noting issues in the audit — i18n link generation lives in a dedicated sitemap that this skill will build if the inventory declares locales.