From soundcheck
Flags vulnerable patterns in autonomous LLM agents enabling irreversible actions without oversight. Suggests fixes like impact classification, tool allowlists, pre-dispatch auditing, and structured parameters for safe workflows.
npx claudepluginhub thejefflarson/soundcheck --plugin soundcheckThis skill uses the workspace's default tool permissions.
Prevents autonomous agents from taking irreversible or high-impact actions without
Guides building reliable autonomous AI agents with ReAct, Plan-Execute loops, reflection patterns, goal decomposition, and frameworks like LangGraph, CrewAI.
Architects reliable autonomous AI agents with ReAct, Plan-Execute loops, goal decomposition, reflection patterns, guardrails, checkpointing, and production reliability principles.
Implements governance patterns for AI agents: policy-based tool controls, intent classification, trust scoring, audit trails, rate limits. For LangChain, CrewAI, OpenAI Agents.
Share bugs, ideas, or general feedback.
Prevents autonomous agents from taking irreversible or high-impact actions without human oversight. When an LLM can directly write files, send emails, or modify databases, a single compromised or hallucinated step can cause unrecoverable damage.
send_email() or delete_record() immediately on LLM instruction with no confirmationFlag the vulnerable code and explain the risk. Then suggest a fix that establishes these properties:
run_sql that accepts arbitrary queries violates
this; a tool named archive_record(id) that only issues a scoped update does not.Anchor — shape, not implementation:
action = plan_from_llm(task)
require(action.name in TOOL_ALLOWLIST)
if impact(action) == HIGH:
require(human_approves(action)) # blocks until approved
audit_log(action) # before dispatch, not after
dispatch(action)
Confirm these properties hold regardless of language or framework: