{"name":"dokimos-dev-dokimos","owner":{"name":"ClaudePluginHub"},"plugins":[{"name":"dokimos-dev-create-evaluator-plugins-create-evaluator","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Scaffold a new Evaluator implementation for the Dokimos LLM evaluation framework. Creates evaluator classes extending BaseEvaluator with the builder pattern, supporting both simple evaluators and LLM-judged evaluators using JudgeLM.","version":"0.1.0","strict":true,"keywords":["dokimos","evaluator","llm-evaluation","scaffold","java","builder-pattern"],"category":"utilities"},{"name":"dokimos-dev-create-dataset-plugins-create-dataset","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Create evaluation datasets for the Dokimos LLM evaluation framework in JSON, CSV, or JSONL format. Supports simple and structured example formats with inputs, expected outputs, and metadata.","version":"0.1.0","strict":true,"keywords":["dokimos","dataset","llm-evaluation","json","csv","jsonl"],"category":"data"},{"name":"dokimos-dev-create-tests-plugins-create-tests","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Scaffold eval-driven tests using dokimos-junit. Creates JUnit parameterized tests with @DatasetSource and Assertions.assertEval() for running Dokimos evaluations as unit tests in CI.","version":"0.1.0","strict":true,"keywords":["dokimos","junit","eval-driven-development","parameterized-test","dataset-source","llm-evaluation"],"category":"development"},{"name":"dokimos-dev-create-experiment-plugins-create-experiment","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Scaffold a Dokimos Experiment that wires together a dataset, task, evaluators, and optional reporter. Supports parallelism, multiple runs for variance reduction, and server-based reporting.","version":"0.1.0","strict":true,"keywords":["dokimos","experiment","llm-evaluation","scaffold","java","eval-pipeline"],"category":"data"},{"name":"dokimos-dev-evaluate-agent-plugins-evaluate-agent","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Set up evaluation of AI agents with tool call validation, correctness checks, task completion, and tool reliability using Dokimos. Framework-agnostic — works with any agent framework.","version":"0.1.0","strict":true,"keywords":["dokimos","agent","tool-call","llm-evaluation","task-completion","tool-validation"],"category":"utilities"},{"name":"dokimos-dev-evaluate-koog-plugins-evaluate-koog","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Set up evaluation of Koog AI agents using Dokimos. Wires Koog agents as the system under test or as LLM judges via KoogSupport utilities, with Kotlin DSL support.","version":"0.1.0","strict":true,"keywords":["dokimos","koog","kotlin","ai-agent","llm-evaluation","agent-testing"],"category":"testing"},{"name":"dokimos-dev-evaluate-langchain4j-plugins-evaluate-langchain4j","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Set up evaluation of LangChain4j applications and RAG pipelines using Dokimos. Provides task and judge creation via LangChain4jSupport, with evaluators for faithfulness, contextual relevance, and hallucination.","version":"0.1.0","strict":true,"keywords":["dokimos","langchain4j","rag","llm-evaluation","faithfulness","hallucination"],"category":"deployment"},{"name":"dokimos-dev-evaluate-spring-ai-plugins-evaluate-spring-ai","source":{"source":"github","repo":"dokimos-dev/dokimos"},"description":"Set up evaluation of Spring AI applications using Dokimos. Provides judge creation and type conversion via SpringAiSupport, with @SpringBootTest integration for evaluations in CI.","version":"0.1.0","strict":true,"keywords":["dokimos","spring-ai","spring-boot","llm-evaluation","chatclient","rag"],"category":"testing"}]}