{"name":"markac007-evals-plugins-dev-tools-evals","owner":{"name":"ClaudePluginHub"},"plugins":[{"name":"markac007-evals-plugins-dev-tools-evals","source":{"source":"github","repo":"markac007/cg-claude-workspaces-plugins"},"description":"AI agent evaluation framework based on Anthropic best practices. Create use cases, LLM judges, A/B prompt tests, and model comparisons.","version":"1.0.0","strict":true,"keywords":[],"category":"testing"}]}