Optuna integration skill for automated hyperparameter optimization with advanced search strategies, pruning, multi-objective optimization, and visualization capabilities.
Optimizes hyperparameters using Optuna with advanced search strategies, pruning, and visualization.
npx claudepluginhub a5c-ai/babysitterThis skill is limited to using the following tools:
README.mdOptimize hyperparameters using Optuna with advanced search strategies, pruning, and visualization.
This skill provides comprehensive capabilities for hyperparameter optimization using Optuna, the state-of-the-art hyperparameter optimization framework. It supports various samplers, pruners, multi-objective optimization, and integration with popular ML frameworks.
pip install optuna>=3.0.0
# Database backends
pip install optuna[mysql] # MySQL support
pip install optuna[postgresql] # PostgreSQL support
# Visualization
pip install optuna-dashboard # Web dashboard
pip install plotly # Interactive plots
# Framework integrations
pip install optuna-integration[sklearn]
pip install optuna-integration[pytorch]
pip install optuna-integration[tensorflow]
import optuna
def objective(trial):
# Suggest hyperparameters
learning_rate = trial.suggest_float('learning_rate', 1e-5, 1e-1, log=True)
n_estimators = trial.suggest_int('n_estimators', 50, 500)
max_depth = trial.suggest_int('max_depth', 3, 15)
subsample = trial.suggest_float('subsample', 0.5, 1.0)
# Train model
model = XGBClassifier(
learning_rate=learning_rate,
n_estimators=n_estimators,
max_depth=max_depth,
subsample=subsample,
random_state=42
)
# Cross-validation
score = cross_val_score(model, X_train, y_train, cv=5, scoring='accuracy').mean()
return score
# Create study
study = optuna.create_study(
direction='maximize',
study_name='xgboost-tuning',
storage='sqlite:///optuna.db',
load_if_exists=True
)
# Optimize
study.optimize(objective, n_trials=100, timeout=3600)
# Best results
print(f"Best trial: {study.best_trial.number}")
print(f"Best value: {study.best_value:.4f}")
print(f"Best params: {study.best_params}")
import optuna
from optuna.pruners import MedianPruner
def objective_with_pruning(trial):
# Suggest hyperparameters
learning_rate = trial.suggest_float('learning_rate', 1e-5, 1e-1, log=True)
n_epochs = trial.suggest_int('n_epochs', 10, 100)
# Create model
model = create_model(learning_rate)
# Training loop with pruning
for epoch in range(n_epochs):
train_loss = train_one_epoch(model)
val_accuracy = evaluate(model)
# Report intermediate value
trial.report(val_accuracy, epoch)
# Prune if unpromising
if trial.should_prune():
raise optuna.TrialPruned()
return val_accuracy
# Create study with pruner
study = optuna.create_study(
direction='maximize',
pruner=MedianPruner(n_startup_trials=5, n_warmup_steps=10)
)
study.optimize(objective_with_pruning, n_trials=100)
import optuna
def multi_objective(trial):
# Hyperparameters
learning_rate = trial.suggest_float('learning_rate', 1e-5, 1e-1, log=True)
model_size = trial.suggest_categorical('model_size', ['small', 'medium', 'large'])
# Train model
model = create_model(learning_rate, model_size)
train(model)
# Multiple objectives
accuracy = evaluate_accuracy(model)
inference_time = measure_inference_time(model)
return accuracy, inference_time # maximize accuracy, minimize time
# Create multi-objective study
study = optuna.create_study(
directions=['maximize', 'minimize'],
study_name='pareto-optimization'
)
study.optimize(multi_objective, n_trials=100)
# Get Pareto front
pareto_front = study.best_trials
for trial in pareto_front:
print(f"Accuracy: {trial.values[0]:.4f}, Time: {trial.values[1]:.4f}")
import optuna
from optuna.integration import OptunaSearchCV
# Define parameter distributions
param_distributions = {
'n_estimators': optuna.distributions.IntDistribution(50, 500),
'max_depth': optuna.distributions.IntDistribution(3, 15),
'learning_rate': optuna.distributions.FloatDistribution(1e-5, 1e-1, log=True),
'subsample': optuna.distributions.FloatDistribution(0.5, 1.0)
}
# Create search
search = OptunaSearchCV(
XGBClassifier(random_state=42),
param_distributions,
n_trials=100,
cv=5,
scoring='accuracy',
study=study, # Optional: use existing study
n_jobs=-1
)
# Fit
search.fit(X_train, y_train)
# Results
print(f"Best score: {search.best_score_:.4f}")
print(f"Best params: {search.best_params_}")
import optuna
from optuna.integration import PyTorchLightningPruningCallback
def objective(trial):
# Hyperparameters
lr = trial.suggest_float('lr', 1e-5, 1e-1, log=True)
hidden_size = trial.suggest_int('hidden_size', 32, 256)
dropout = trial.suggest_float('dropout', 0.1, 0.5)
# Create model
model = LightningModel(
hidden_size=hidden_size,
dropout=dropout,
lr=lr
)
# Create trainer with pruning callback
trainer = pl.Trainer(
max_epochs=100,
callbacks=[
PyTorchLightningPruningCallback(trial, monitor='val_accuracy')
]
)
trainer.fit(model, train_loader, val_loader)
return trainer.callback_metrics['val_accuracy'].item()
import optuna
# Create shared study with database storage
study = optuna.create_study(
study_name='distributed-study',
storage='postgresql://user:pass@host:5432/optuna',
direction='maximize',
load_if_exists=True
)
# Run on multiple workers (each worker runs this)
study.optimize(objective, n_trials=25) # Each worker does 25 trials
# Results are automatically aggregated
print(f"Total trials: {len(study.trials)}")
const hyperparameterTuningTask = defineTask({
name: 'optuna-hyperparameter-tuning',
description: 'Optimize hyperparameters using Optuna',
inputs: {
studyName: { type: 'string', required: true },
direction: { type: 'string', default: 'maximize' },
nTrials: { type: 'number', default: 100 },
timeout: { type: 'number' },
parameterSpace: { type: 'object', required: true },
objectiveScript: { type: 'string', required: true },
sampler: { type: 'string', default: 'tpe' },
pruner: { type: 'string', default: 'median' }
},
outputs: {
bestValue: { type: 'number' },
bestParams: { type: 'object' },
nTrialsCompleted: { type: 'number' },
studyPath: { type: 'string' }
},
async run(inputs, taskCtx) {
return {
kind: 'skill',
title: `Optimize: ${inputs.studyName}`,
skill: {
name: 'optuna-hyperparameter-tuner',
context: {
operation: 'optimize',
studyName: inputs.studyName,
direction: inputs.direction,
nTrials: inputs.nTrials,
timeout: inputs.timeout,
parameterSpace: inputs.parameterSpace,
objectiveScript: inputs.objectiveScript,
sampler: inputs.sampler,
pruner: inputs.pruner
}
},
io: {
inputJsonPath: `tasks/${taskCtx.effectId}/input.json`,
outputJsonPath: `tasks/${taskCtx.effectId}/result.json`
}
};
}
});
{
"mcpServers": {
"optuna": {
"command": "uvx",
"args": ["optuna-mcp"],
"env": {
"OPTUNA_STORAGE": "sqlite:///optuna.db"
}
}
}
}
optuna_create_study - Create new optimization studyoptuna_get_study - Retrieve study informationoptuna_list_studies - List all studiesoptuna_get_best_trial - Get best trial from studyoptuna_get_trials - List trials in studyoptuna_visualize - Generate visualizationoptuna_suggest_params - Get parameter suggestions| Sampler | Use Case | Pros | Cons |
|---|---|---|---|
TPESampler | Default, most cases | Efficient, handles conditionals | May miss global optimum |
CmaEsSampler | Continuous parameters | Good for correlated params | Only continuous |
GridSampler | Small discrete spaces | Exhaustive | Exponential complexity |
RandomSampler | Baseline, parallel | Simple, embarrassingly parallel | Inefficient |
NSGAIISampler | Multi-objective | Pareto optimization | Slower convergence |
QMCSampler | Space exploration | Low discrepancy | Not adaptive |
| Pruner | Use Case | Aggressiveness |
|---|---|---|
MedianPruner | Default, safe | Moderate |
HyperbandPruner | Deep learning | Aggressive |
SuccessiveHalvingPruner | Resource-efficient | High |
PercentilePruner | Configurable threshold | Variable |
NopPruner | No pruning needed | None |
import optuna.visualization as vis
# Optimization history
fig = vis.plot_optimization_history(study)
fig.write_html('optimization_history.html')
# Parameter importance
fig = vis.plot_param_importances(study)
fig.write_html('param_importance.html')
# Parallel coordinate
fig = vis.plot_parallel_coordinate(study)
fig.write_html('parallel_coordinate.html')
# Contour plot (2 params)
fig = vis.plot_contour(study, params=['learning_rate', 'max_depth'])
fig.write_html('contour.html')
# Slice plot
fig = vis.plot_slice(study)
fig.write_html('slice.html')
# Launch dashboard
optuna-dashboard sqlite:///optuna.db
# Access at http://localhost:8080
Activates when the user asks about AI prompts, needs prompt templates, wants to search for prompts, or mentions prompts.chat. Use for discovering, retrieving, and improving prompts.
Search, retrieve, and install Agent Skills from the prompts.chat registry using MCP tools. Use when the user asks to find skills, browse skill catalogs, install a skill for Claude, or extend Claude's capabilities with reusable AI agent components.
This skill should be used when the user asks to "create an agent", "add an agent", "write a subagent", "agent frontmatter", "when to use description", "agent examples", "agent tools", "agent colors", "autonomous agent", or needs guidance on agent structure, system prompts, triggering conditions, or agent development best practices for Claude Code plugins.