{"name":"vanman2024-llm-evals-plugins-llm-evals","owner":{"name":"ClaudePluginHub"},"plugins":[{"name":"vanman2024-llm-evals-plugins-llm-evals","source":{"source":"github","repo":"vanman2024/ai-dev-marketplace"},"description":"LLM testing and evaluation framework with promptfoo, DeepEval, golden datasets, and Supabase-backed eval tracking","version":"1.0.0","strict":true,"keywords":["llm-testing","evaluation","promptfoo","deepeval","golden-datasets","regression-testing","prompt-testing","metrics","faithfulness","relevance","ci-cd","pytest"],"category":"testing"}]}