Audit AI assistant outputs and pasted text for hallucinations including speculation, ungrounded causality, pseudo-quantification, and completeness overclaims. Receive Pass/Fail results with flagged phrases and rewrite suggestions. Automatically block completions on triggers via stop hooks, inject prevention prompts at session start, and track stats after tool use to enforce evidence-first responses.
npx claudepluginhub jamie-bitflight/claude_skills --plugin hallucination-detectorMatches all tools
Hooks run on every tool call, not just specific ones
Share bugs, ideas, or general feedback.
Own this plugin?
Verify ownership to unlock analytics, metadata editing, and a verified badge.
Sign in to claimOwn this plugin?
Verify ownership to unlock analytics, metadata editing, and a verified badge.
Sign in to claimStop-hook hallucination and speculation-as-diagnosis detector. Audits the last assistant message for speculation, ungrounded causality, pseudo-quantification, and completeness overclaims; blocks stopping to force evidence-first rewrites.
Three-layer verification pipeline for AI output. Extracts claims, finds sources, and flags hallucination risks so humans can verify before acting.
YES.md — PUA says NO, YES says YES. 6-layer AI governance: format → trigger → hooks → anti-slack → gates → memory. Makes AI do things RIGHT with encouragement, not pressure. Available in English, 中文, 日本語.
Force Claude to re-validate when you have doubts (!doubt)
Achieve certain comprehension of AI work — /grasp (κατάληψις: a grasping firmly)
Semantic search for Claude Code conversations. Remember past discussions, decisions, and patterns.
Prevents Claude from finishing tasks with speculation, unverified claims, or invented causality.
Claude regularly delivers responses that sound confident but contain:
These patterns are dangerous precisely because they sound authoritative — indistinguishable from real analysis. This plugin forces Claude to ground every claim in actual observations before it can complete a task.
Use the Vercel Skills CLI to install across any supported agent:
npx skills add bitflight-devops/hallucination-detector
Target a specific agent:
npx skills add bitflight-devops/hallucination-detector -a claude-code
npx skills add bitflight-devops/hallucination-detector -a cursor
npx skills add bitflight-devops/hallucination-detector -a codex
npx skills add bitflight-devops/hallucination-detector -a opencode
To uninstall:
npx skills remove hallucination-detector
In Claude Code, register the marketplace first:
/plugin marketplace add bitflight-devops/hallucination-detector
Then install the plugin:
/plugin install hallucination-detector@hallucination-detector
In Cursor Agent chat, install from marketplace:
/plugin-add hallucination-detector
Tell Codex:
Fetch and follow instructions from https://raw.githubusercontent.com/bitflight-devops/hallucination-detector/refs/heads/main/.codex/INSTALL.md
Detailed docs: .codex/INSTALL.md
Tell OpenCode:
Fetch and follow instructions from https://raw.githubusercontent.com/bitflight-devops/hallucination-detector/refs/heads/main/.opencode/INSTALL.md
Detailed docs: .opencode/INSTALL.md
Start a new session in your chosen platform and write a response containing speculation (e.g., "this is probably caused by..."). The plugin should block the response and require evidence-first rewriting.
For detailed information about how the plugin works, its architecture, configuration, and internal detection mechanisms, see the Architecture Reference.
LLMs like Claude are optimized during training to produce responses that appear helpful and confident. This creates a systematic failure mode:
Speculation as diagnosis - When asked "why did X happen?", Claude draws on training patterns to generate plausible-sounding explanations. These explanations feel authoritative but have no connection to the actual state of your system. Claude hasn't checked logs, read config files, or verified anything — it's pattern-matching from training data.
Invented causality - Causal claims ("X because Y") require evidence showing the relationship. Claude often asserts causality based on what typically causes similar symptoms, not what actually caused this specific instance. The word "because" in Claude's output frequently signals unverified inference.
Fake rigor - Scores and percentages ("8/10 quality", "70% improvement") create an illusion of measurement. Without methodology, sample size, and reproducible criteria, these numbers are meaningless — yet they make responses feel more credible.