From autoresearch
Start an autoresearch experiment loop — pre-flight checks, reads program.md, enters autonomous keep-or-revert loop. Specify target: local, server, or auto-detect.
npx claudepluginhub flight505/autoresearch[local|server] [options like 'overnight, aim for 50+ experiments']# Autoresearch Run You are starting an autonomous research session using Karpathy's autoresearch framework. ## 1. Load config Read the autoresearch config to find hardware targets: If `NO_CONFIG`: stop and tell the user to run `/autoresearch:setup` first. ## 2. Determine target If `$ARGUMENTS` specifies a target (`local` or `server`), use that. Otherwise, auto-detect: if the current directory is inside a configured local path, use `local`. If unsure, ask the user. ## 3. Pre-flight checks ### For `local` target: ### For `server` target: Read `ssh_host` and `path` from config, t...
/runAuto-completes all pending agent-foreman tasks in priority order unattended or a specific task by ID. Loops through fetch, implement, verify, done/fail until complete, with summary.
/runOrchestrates post-development tasks for app launch: SEO analysis, screenshots, personas, ads, articles, landing pages in dependency order. Supports init, run, status, and individual tasks.
/runExecutes a single task or batch of tasks from taskmanager SQLite DB with dependency resolution, memory support, and status propagation. Supports task ID, --batch N, and memory flags.
You are starting an autonomous research session using Karpathy's autoresearch framework.
Read the autoresearch config to find hardware targets:
cat "${CLAUDE_PLUGIN_DATA:-${HOME}/.autoresearch}/config.json" 2>/dev/null || echo "NO_CONFIG"
If NO_CONFIG: stop and tell the user to run /autoresearch:setup first.
If $ARGUMENTS specifies a target (local or server), use that.
Otherwise, auto-detect: if the current directory is inside a configured local path, use local. If unsure, ask the user.
local target:# Confirm we're in the right repo
[ -f program.md ] && [ -f train.py ] && [ -f prepare.py ] && echo "repo OK" || echo "FAIL: not in autoresearch repo"
# Detect backend
python3 -c "import mlx; print('backend: mlx')" 2>/dev/null || \
python3 -c "import torch; print('backend: cuda (' + str(torch.cuda.get_device_name(0)) + ')')" 2>/dev/null || \
echo "FAIL: no backend"
# Check data cache
ls ~/.cache/autoresearch/ 2>/dev/null && echo "data OK" || echo "FAIL: run uv run prepare.py first"
server target:Read ssh_host and path from config, then:
# Test SSH connectivity (2s timeout)
ssh -o ConnectTimeout=2 <ssh_host> "echo 'connected'" 2>/dev/null || echo "FAIL: cannot reach server"
# Check repo on server
ssh <ssh_host> "[ -f <path>/program.md ] && [ -f <path>/train.py ] && echo 'repo OK' || echo 'FAIL: repo not found at <path>'"
# Check GPU
ssh <ssh_host> "nvidia-smi --query-gpu=name,memory.total --format=csv,noheader 2>/dev/null || echo 'FAIL: no GPU'"
If any check fails, report the issue and stop.
Read program.md in the target directory. It is your authoritative research protocol.
Also read train.py and prepare.py to understand the training setup.
For server targets: read the files via SSH.
Propose a run tag based on today's date (e.g. mar25). Create the branch:
git checkout -b autoresearch/$(date +%b%d | tr '[:upper:]' '[:lower:]')
Initialize results.tsv with the header:
commit val_bpb memory_gb status description
Follow the experiment loop from program.md exactly:
$ARGUMENTS