Expertise in updating local documentation stubs with current online content. Use when the user asks to 'update documentation', 'sync docs with online sources', or 'refresh local docs'.
From interface-designnpx claudepluginhub lwhite702/klutr_primary --plugin interface-designThis skill uses the workspace's default tool permissions.
Provides UI/UX resources: 50+ styles, color palettes, font pairings, guidelines, charts for web/mobile across React, Next.js, Vue, Svelte, Tailwind, React Native, Flutter. Aids planning, building, reviewing interfaces.
Fetches up-to-date documentation from Context7 for libraries and frameworks like React, Next.js, Prisma. Use for setup questions, API references, and code examples.
Builds 3-5 year financial models for startups with cohort revenue projections, cost structures, cash flow, headcount plans, burn rate, runway, and scenario analysis.
You act as a Documentation Automation Engineer, specializing in synchronizing local documentation files with their current online counterparts. You are methodical, respectful of API rate limits, and thorough in tracking changes.
Activate this skill when the user:
Identify the documentation directory
# Find all markdown files with URL stubs
grep -r "Fetch live documentation:" <directory> --include="*.md"
Extract all URLs from stub files
import re
from pathlib import Path
def extract_stub_url(file_path):
with open(file_path, 'r', encoding='utf-8') as f:
content = f.read()
match = re.search(r'Fetch live documentation:\s*(https?://[^\s]+)', content)
return match.group(1) if match else None
Create inventory of files to update
Check if content has changed
import hashlib
import requests
def get_content_hash(content):
return hashlib.md5(content.encode()).hexdigest()
def get_online_content_hash(url):
response = requests.get(url, timeout=10)
return get_content_hash(response.text)
Compare local vs online hashes
Download content from URL
from bs4 import BeautifulSoup
from urllib.parse import urlparse
def download_content_from_url(url):
response = requests.get(url, timeout=10)
soup = BeautifulSoup(response.text, 'html.parser')
# Extract main content
main_content = soup.find('main') or soup.find('article')
if main_content:
content_text = main_content.get_text(separator='\n')
# Extract title
title_tag = soup.find('title')
title = title_tag.get_text().split('|')[0].strip() if title_tag else urlparse(url).path.split('/')[-1]
# Format as markdown
return f"# {title}\n\n{content_text}\n\n---\n\nFetch live documentation: {url}\n"
Update the local file
def update_file(file_path, content):
with open(file_path, 'w', encoding='utf-8') as f:
f.write(content)
Generate summary statistics
Create detailed report
URL unreachable (404/timeout)
Content download fails
File write fails
User: "Update all the documentation in /Volumes/PARA/03_Resources/ai-docs/_kb/@platforms/anthropic/"
Agent Response:
I'll update the documentation files in that directory. Let me start by identifying all stub files.
Found 96 documentation files with online URLs.
Before I proceed with the update, here's the plan:
1. Check each file against its online counterpart
2. Only update files where content has changed
3. Process in batches of 10 to avoid timeouts
4. Rate limit requests to 1 per second
This will take approximately 10-15 minutes. Shall I proceed?
User: "Update just the configuration docs in the anthropic folder"
Agent Response:
I'll update only the configuration documentation files.
Found 5 files in the configuration/ subdirectory:
- configuration/settings.md
- configuration/model-config.md
- configuration/terminal-config.md
- configuration/memory.md
- configuration/statusline.md
Proceeding with update...
After completion, provide a summary like:
════════════════════════════════════════════════
DOCUMENTATION UPDATE SUMMARY
════════════════════════════════════════════════
Files updated: 96
Files skipped (already current): 0
Errors encountered: 0
Total processing time: ~15 minutes
All documentation files have been synchronized with their online sources.
scripts/doc_update.py - Main update scriptreferences/url_patterns.md - Common URL patterns for documentation sitesreferences/error_codes.md - HTTP error code handling guide