Detect AI/LLM-generated text patterns in research writing. Use when: (1) Reviewing manuscript drafts before submission, (2) Pre-commit validation of documentation, (3) Quality assurance checks on research artifacts, (4) Ensuring natural academic writing style, (5) Tracking writing authenticity over time. Analyzes grammar perfection, sentence uniformity, paragraph structure, word frequency (AI-typical words like 'delve', 'leverage', 'robust'), punctuation patterns, and transition word overuse.
Detects AI-generated text patterns in academic writing by analyzing grammar perfection, sentence uniformity, word frequency, and punctuation. Use before manuscript submission, pre-commit validation, or quality assurance checks to ensure authentic writing style.
/plugin marketplace add astoreyai/ai_scientist/plugin install research-assistant@research-assistant-marketplaceThis skill is limited to using the following tools:
Detect patterns typical of LLM-generated text to ensure natural, human-authored academic writing. This skill helps maintain authenticity in research publications, dissertations, and documentation.
What We Check:
Red Flags:
Human Writing Typically Has:
What We Check:
Red Flags:
Human Writing Typically Has:
What We Check:
Red Flags:
Human Writing Typically Has:
High-Risk AI Words (overused by LLMs):
Verbs:
Adjectives:
Transition Words (overused):
Phrases:
Detection Criteria:
What We Check:
Red Flags:
Human Writing Typically Has:
Overall Confidence = Weighted average of:
Each metric scored 0-100, then combined with weights.
Characteristics:
Action: ✅ Writing appears authentic, no changes needed
Characteristics:
Action: ⚠️ Review flagged sections, apply suggestions selectively
Examples of Mixed Writing:
Characteristics:
Action: 🚫 Significant revision needed, rewrite in authentic voice
When running AI-check analysis, generate a comprehensive report:
Overall Confidence Score: 65%
Status: MEDIUM - Possible AI assistance detected
Files Analyzed: 1
Total Words: 3,456
Recommendation: Review flagged sections
Grammar Perfection: 85% (High - suspiciously few errors)
Sentence Uniformity: 72% (High - repetitive structures)
Paragraph Structure: 68% (Medium - some variation)
AI-Typical Words: 58% (Medium - 4.2 per 1000 words)
Punctuation Patterns: 45% (Low - natural variation)
Lines 45-67 (Confidence: 82%)
Pattern: Excessive transition words + uniform sentences
AI Words: "moreover", "furthermore", "leverage", "robust"
Lines 112-134 (Confidence: 76%)
Pattern: Perfect grammar + mechanical structure
AI Words: "delve", "comprehensive", "facilitate"
High-Risk AI Words Found (per 1000 words):
• "delve" (2 occurrences) - RARELY used by humans
• "leverage" (3 occurrences) - Business jargon overuse
• "robust" (4 occurrences) - Technical overuse
• "furthermore" (6 occurrences) - Formal transition overuse
Sentence Uniformity Issues:
• 67% of sentences are 15-25 words (AI sweet spot)
• 82% of paragraphs start with transition words
• Low variation in sentence complexity
Paragraph Structure Issues:
• All paragraphs 4-6 sentences long
• Mechanical topic-sentence pattern throughout
Top AI-Typical Words:
1. "furthermore" - 6x (baseline: 0.5x per 1000 words)
2. "robust" - 4x (baseline: 0.8x per 1000 words)
3. "leverage" - 3x (baseline: 0.3x per 1000 words)
4. "comprehensive" - 3x (baseline: 1.2x per 1000 words)
5. "delve" - 2x (baseline: 0.1x per 1000 words)
Comparison to Human Academic Writing:
Your text: 4.2 AI-typical words per 1000
Human baseline: 1.5 AI-typical words per 1000
Ratio: 2.8x higher than human baseline
Sentence Structure:
Why Better: Simpler words, no transition word, more direct
Word Choice:
Why Better: Active voice, common words, clearer meaning
Paragraph Variation:
Why Better: Natural flow based on content, not formula
Current: 15-25 word sentences consistently
Suggestion: Mix short (5-10), medium (15-20), long (25-35) sentences
Example:
- Short: "The effect was significant."
- Medium: "We observed a 23% increase across all conditions."
- Long: "This finding aligns with previous work showing that..."
Replace → With
- "delve into" → "examine", "explore", "investigate"
- "leverage" → "use", "apply", "employ"
- "utilize" → "use"
- "robust" → "strong", "reliable", "thorough"
- "facilitate" → "enable", "help", "allow"
- "furthermore" → "also", "next", [or remove]
- "moreover" → "additionally", "also", [or use dash]
- "comprehensive" → "complete", "thorough", "full"
- Use contractions in appropriate contexts ("it's", "we'll")
- Include domain-specific jargon naturally
- Allow informal phrasing in methods/procedures
- Use occasional sentence fragments for emphasis
- Add personal observations or interpretations
- Include field-specific colloquialisms
Current: All paragraphs follow topic-support-support-support-conclusion
Suggestion: Vary based on content
- Use single-sentence paragraphs for emphasis
- Combine related ideas into longer paragraphs
- Don't force every paragraph to have 5 sentences
- Let content determine structure, not formula
❌ "Furthermore, the results show... Moreover, the analysis reveals..."
✅ "The results show... The analysis also reveals..." [simpler transitions]
✅ "The results show... Looking closer, the analysis..." [natural bridges]
Automatic checking before git commits
# Configured in .claude/settings.json
"gitPreCommit": {
"command": "python3 hooks/pre-commit-ai-check.py",
"enabled": true
}
Behavior:
.md, .tex, .rst filesgit commit --no-verifyExit Codes:
Part of comprehensive QA workflow
Integrated into code/quality_assurance/qa_manager.py:
Configuration (.ai-check-config.yaml):
qa_integration:
enabled: true
max_confidence_threshold: 0.40
check_manuscripts: true
check_documentation: true
generate_detailed_reports: true
Real-time feedback during writing
Agent checks writing incrementally:
Agent Workflow:
Manual invocation by user or agents
User Invocation:
Please run ai-check on docs/manuscript/discussion.tex and provide detailed feedback.
Agent Invocation:
I'll use the ai-check skill to verify this text before proceeding.
CLI Tool:
python tools/ai_check.py path/to/file.md
python tools/ai_check.py --directory docs/
python tools/ai_check.py --format html --output report.html
Log all AI-check runs to database for evolution tracking:
Database Schema (PostgreSQL via research-database MCP):
CREATE TABLE ai_check_history (
id SERIAL PRIMARY KEY,
file_path TEXT NOT NULL,
git_commit TEXT,
timestamp TIMESTAMP DEFAULT NOW(),
overall_confidence FLOAT,
grammar_score FLOAT,
sentence_score FLOAT,
paragraph_score FLOAT,
word_score FLOAT,
punctuation_score FLOAT,
ai_words_found JSONB,
flagged_sections JSONB
);
Track writing evolution:
File: docs/manuscript/discussion.tex
Version History:
2025-01-15: 78% confidence (HIGH - likely AI)
2025-01-18: 52% confidence (MEDIUM - revision 1)
2025-01-20: 34% confidence (LOW-MEDIUM - revision 2)
2025-01-22: 18% confidence (LOW - authentic writing)
Trend: ✅ Improving toward authentic writing
Use Cases:
.ai-check-config.yaml# AI-Check Skill Configuration
# Pre-Commit Hook Settings
pre_commit:
enabled: true
check_files: [".md", ".tex", ".rst", ".txt"]
check_docstrings: true # Check Python docstrings
block_threshold: 0.70 # Block commit if >= 70%
warn_threshold: 0.30 # Warn if >= 30%
exclude_patterns:
- "*/examples/*"
- "*/tests/*"
- "*/node_modules/*"
- "*/.venv/*"
# Quality Assurance Integration
qa_integration:
enabled: true
max_confidence_threshold: 0.40 # Fail QA if >= 40%
check_manuscripts: true
check_documentation: true
generate_detailed_reports: true
track_history: true
# Detection Parameters
detection:
# Weight each metric (must sum to 1.0)
weights:
grammar_perfection: 0.20
sentence_uniformity: 0.25
paragraph_structure: 0.20
ai_word_frequency: 0.25
punctuation_patterns: 0.10
# AI-typical word lists
ai_words:
high_risk: ["delve", "leverage", "utilize"]
medium_risk: ["robust", "comprehensive", "facilitate"]
transitions: ["furthermore", "moreover", "additionally"]
# Thresholds
ai_words_per_1000_threshold: 3.0
human_baseline_per_1000: 1.5
# Report Generation
reporting:
default_format: "markdown" # markdown, json, html
include_suggestions: true
include_word_frequency: true
include_flagged_sections: true
max_flagged_sections: 10
# Tracking
tracking:
enabled: true
database: "research-database-mcp"
retention_days: 365
Create .ai-check.local.yaml for project-specific settings:
# Project-specific overrides
pre_commit:
block_threshold: 0.60 # More lenient for early drafts
detection:
ai_words:
high_risk: ["delve"] # Only flag worst offenders
Input Text:
Furthermore, this comprehensive study delves into the robust
methodologies utilized to facilitate the implementation of innovative
approaches. Moreover, the analysis demonstrates significant findings
that leverage state-of-the-art techniques. Subsequently, the results
indicate substantial improvements across all metrics. Nevertheless,
additional research is crucial to fully comprehend the implications.
AI-Check Report:
Overall Confidence: 89% (HIGH - Likely AI-generated)
Issues Detected:
- 8 AI-typical words in 60 words (13.3 per 1000 words!)
- Every sentence starts with transition word
- Uniform sentence length (15-18 words each)
- Perfect grammar, zero natural imperfections
- Mechanical paragraph structure
AI Words Found:
- furthermore, comprehensive, delves, robust
- utilized, facilitate, innovative, leverage
- demonstrates, significant, subsequently, substantial
- nevertheless, crucial, comprehend
Recommendation: Complete rewrite recommended
Suggested Revision:
We examined the methods used in this approach. The analysis shows
clear improvements across metrics. However, more research is needed
to understand the full implications.
(23 words, 12% confidence - much more natural)
Input Text:
The experimental design followed standard protocols established in
previous work (Smith et al., 2023). We collected data from 150
participants over six months. Statistical analysis used mixed-effects
models to account for repeated measures. The results showed a
significant main effect of condition (p < 0.001).
AI-Check Report:
Overall Confidence: 35% (MEDIUM - Possible minor AI assistance)
Issues Detected:
- Slightly uniform sentence length (11-15 words)
- One AI-typical word: "significant" (statistical context acceptable)
- Otherwise natural academic writing
Recommendation: Minor revisions optional, writing appears largely authentic
Input Text:
OK so here's what we found. The effect was huge - way bigger than
expected. Participants in the experimental group scored 23% higher
on average. This wasn't just statistically significant; it was
practically meaningful.
We're still not sure why. Maybe it's the timing? Could be the
instructions were clearer. Need to run follow-ups.
AI-Check Report:
Overall Confidence: 8% (LOW - Clearly human writing)
Human Writing Indicators:
- Natural sentence variation (4-19 words)
- Informal elements ("OK so", "way bigger")
- Incomplete thoughts and questions
- Natural uncertainty expressions
- Zero AI-typical words
- Authentic voice throughout
Recommendation: Writing is authentic, no changes needed
Run Before Advisor Meetings
Use During Drafting
Pre-Submission Validation
Establish Team Standards
Code Review Integration
Track Team Writing
Pre-Submission Checklist
Demonstrating Authenticity
Not 100% Accurate
Cannot Detect All AI Usage
Domain Limitations
This skill is a tool, not a replacement for human judgment:
Problem: False positive on authentic writing Solution: Check if writing is overly formal. Consider field-specific norms. Adjust thresholds in config.
Problem: AI text passing with low confidence Solution: Update AI-typical word lists. Check for heavily edited text. Report patterns for skill updates.
Problem: Pre-commit hook too slow Solution: Reduce checked file types. Enable caching. Check only modified sections.
Problem: Disagreement with manual review Solution: Generate detailed report. Review flagged sections specifically. Consider multiple metrics not just overall score.
docs/skills/ai-check-reference.mdLast Updated: 2025-11-09 Version: 1.0.0 License: MIT