Evaluate ML model with comprehensive metrics
Evaluate ML models with comprehensive metrics and generate detailed reports. Use this after training to validate performance against specifications before deployment.
/plugin marketplace add anton-abyzov/specweave/plugin install sw-ml@specweaveYou are evaluating an ML model in a SpecWeave increment. Generate a comprehensive evaluation report following ML best practices.
# Load test set from increment
X_test = load_test_data(increment_path)
y_test = load_test_labels(increment_path)
from specweave import ModelEvaluator
evaluator = ModelEvaluator(model, X_test, y_test)
metrics = evaluator.compute_all_metrics()
Create evaluation-report.md in increment folder:
# Model Evaluation Report
## Model: [Model Name]
- Version: [Version]
- Increment: [Increment ID]
- Date: [Evaluation Date]
## Overall Performance
[Metrics table]
## Visualizations
[Embedded plots]
## Cross-Validation
[CV results]
## Comparison to Baseline
[Baseline comparison]
## Statistical Tests
[Significance tests]
## Recommendations
[Deploy/improve/investigate]
After evaluation, report: