Deploy and run ML experiments on local or remote GPU servers. Use when user says "run experiment", "deploy to server", "跑实验", or needs to launch training jobs.
npx claudepluginhub llv22/autoresearchwitheyesThis skill is limited to using the following tools:
Deploy and run ML experiment: $ARGUMENTS
Autonomously runs deep learning experiments 24/7 in a THINK-EXECUTE-REFLECT loop with zero-cost GPU monitoring, Leader-Worker architecture, and constant-size memory.
Sets up ML experiment tracking with MLflow or Weights & Biases: installs packages, initializes tools, and provides logging code for parameters, metrics, and artifacts.
Manages remote GPU clusters via rca CLI: run commands, batch jobs, GPU/node inspection, file sync with mutagen. For training, remote exec, status checks.
Share bugs, ideas, or general feedback.
Deploy and run ML experiment: $ARGUMENTS
Read the project's CLAUDE.md to determine the experiment environment:
If no server info is found in CLAUDE.md, ask the user.
Check GPU availability on the target machine:
Remote:
ssh <server> nvidia-smi --query-gpu=index,memory.used,memory.total --format=csv,noheader
Local:
nvidia-smi --query-gpu=index,memory.used,memory.total --format=csv,noheader
# or for Mac MPS:
python -c "import torch; print('MPS available:', torch.backends.mps.is_available())"
Free GPU = memory.used < 500 MiB.
Only sync necessary files — NOT data, checkpoints, or large files:
rsync -avz --include='*.py' --exclude='*' <local_src>/ <server>:<remote_dst>/
For each experiment, create a dedicated screen session with GPU binding:
ssh <server> "screen -dmS <exp_name> bash -c '\
eval \"\$(<conda_path>/conda shell.bash hook)\" && \
conda activate <env> && \
CUDA_VISIBLE_DEVICES=<gpu_id> python <script> <args> 2>&1 | tee <log_file>'"
# Linux with CUDA
CUDA_VISIBLE_DEVICES=<gpu_id> python <script> <args> 2>&1 | tee <log_file>
# Mac with MPS (PyTorch uses MPS automatically)
python <script> <args> 2>&1 | tee <log_file>
For local long-running jobs, use run_in_background: true to keep the conversation responsive.
Remote:
ssh <server> "screen -ls"
Local: Check process is running and GPU is allocated.
tee to save logs for later inspectionrun_in_background: true to keep conversation responsiveUsers should add their server info to their project's CLAUDE.md:
## Remote Server
- SSH: `ssh my-gpu-server`
- GPU: 4x A100 (80GB each)
- Conda: `eval "$(/opt/conda/bin/conda shell.bash hook)" && conda activate research`
- Code dir: `/home/user/experiments/`
## Local Environment
- Mac MPS / Linux CUDA
- Conda env: `ml` (Python 3.10 + PyTorch)