From harnessml
Use when generating feature hypotheses from domain knowledge. This is not a one-time pre-work step — return here whenever results surprise you, progress stalls, or a new data source becomes available.
npx claudepluginhub msilverblatt/harness-ml --plugin harnessmlThis skill uses the workspace's default tool permissions.
Use when generating feature hypotheses from domain knowledge. This is not a one-time pre-work step — return here whenever results surprise you, progress stalls, or a new data source becomes available.
Conducts multi-round deep research on GitHub repos via API and web searches, generating markdown reports with executive summaries, timelines, metrics, and Mermaid diagrams.
Dynamically discovers and combines enabled skills into cohesive, unexpected delightful experiences like interactive HTML or themed artifacts. Activates on 'surprise me', inspiration, or boredom cues.
Generates images from structured JSON prompts via Python script execution. Supports reference images and aspect ratios for characters, scenes, products, visuals.
Use when generating feature hypotheses from domain knowledge. This is not a one-time pre-work step — return here whenever results surprise you, progress stalls, or a new data source becomes available.
The biggest ML gains come from features that capture real phenomena, not from hyperparameter tuning or model architecture changes. A feature that encodes domain knowledge — even imperfectly — gives the model information it cannot learn from the raw data alone.
Domain research is how you generate those features.
Before looking at correlations, ask: what does domain expertise say should matter?
Direct predictors — Features that directly measure the outcome driver.
count_of_comorbiditiessquare_footageProxy signals — Indirect indicators when direct measurement is unavailable.
days_payable_outstanding (when cash flow data is missing)pharmacy_visit_frequency (when medical records are incomplete)Interaction effects — Two features weak alone, strong together.
high_leverage * rising_rates — leverage is fine until rates moveis_diabetic * high_bmi — captures a specific high-risk populationConditional effects — A feature that only matters in certain contexts.
marketing_spend only predicts sales for products with existing brand awarenessrainfall only affects crop yield during the growing seasonRegime indicators — Signals that relationships change under different conditions.
vix_above_30 — volatility regime where correlations break downproduct_lifecycle_stage — growth vs maturity dynamics differContrarian signals — Counter-intuitive predictive direction. Often the most valuable.
For each hypothesis:
features(action="discover")
If an existing feature correlates >0.8 with your proposed feature, yours is likely redundant. Either skip it or refine the hypothesis to capture what the existing feature misses.
Maintain a running log. This is the connective tissue between domain knowledge and experiment results.
### Hypothesis: [Name]
- **Domain reasoning**: Why this should be predictive (the mechanism)
- **Source**: Where you found evidence
- **Feature(s)**: Name and formula
- **Expected signal**: Strong / Medium / Weak
- **Result**: What happened when tested
- **Learning**: What this tells us about the domain
- **Follow-up**: Next hypothesis generated by this result
The Follow-up field is the most important. Every tested hypothesis should generate at least one new question.
features(action="auto_search") finds statistical artifacts; domain reasoning finds real signals. Use auto-search as a supplement, not a replacement.