From soundcheck
Checks fine-tuning pipelines, dataset loaders, and curation scripts for training data poisoning risks like unverified ingestion, missing validation, duplicates, and poor splits. Suggests OWASP LLM03 fixes.
npx claudepluginhub thejefflarson/soundcheck --plugin soundcheckThis skill uses the workspace's default tool permissions.
Protects against malicious or low-quality examples being introduced into training or
Detects AI/ML security vulnerabilities like unsafe model deserialization in PyTorch/Joblib/NumPy, prompt injection in LLM prompts, and risks in Jupyter notebooks or ML pipelines.
Checks for LLM supply chain vulnerabilities including unverified model downloads, floating version tags, unapproved providers, and unchecked automated updates. Flags risks and suggests pinned SHAs, checksums, org allowlists, and human approval.
Audits AI-generated code and LLM applications for security vulnerabilities, covering OWASP Top 10 for LLMs, secure coding patterns, and AI-specific threat models.
Share bugs, ideas, or general feedback.
Protects against malicious or low-quality examples being introduced into training or fine-tuning datasets. Poisoned data can embed backdoors, degrade accuracy, or skew model behavior in ways that are difficult to detect after training completes.
Flag the vulnerable code and explain the risk. Then suggest a fix that establishes these properties:
ignore previous, <|im_start|>, jailbreak signatures), and
encoding/Unicode sanity. Invalid examples are dropped, not silently used.Anchor — shape, not implementation:
require(sha256(dataset_file) == PINNED_SHA256)
rows = [r for r in parse(dataset_file) if validate(r)] # per-example
unique = dedupe_by_hash(rows)
require(max_class_fraction(unique) < 0.8) # anomaly gate
train, val = split_by_source(unique, val_fraction=0.1)
Confirm the response: