From memstack
Generates optimized multi-stage Dockerfiles and docker-compose configs for containerizing Node.js, Python, Go, Rust apps with health checks, volumes, and non-root security.
npx claudepluginhub cwinvestments/memstack --plugin memstackThis skill uses the workspace's default tool permissions.
*Analyze a project and generate optimized Dockerfiles, docker-compose configs, and deployment-ready container setups with health checks and volume management.*
Generates Dockerfiles and docker-compose.yml for Node.js, Python, Go apps. Optimizes builds, troubleshoots container issues, and applies security/performance best practices.
Creates optimized Docker containers using multi-stage builds, security best practices, and minimal image sizes. Use for containerizing apps, Dockerfiles, image optimization, Docker Compose services.
Generates optimized multi-stage Dockerfiles for Node.js, Python, Rust, Go apps with non-root users, layer caching, health checks, and .dockerignore. Use for containerizing apps or Docker Compose setup.
Share bugs, ideas, or general feedback.
Analyze a project and generate optimized Dockerfiles, docker-compose configs, and deployment-ready container setups with health checks and volume management.
When this skill activates, output:
π³ Docker Setup β Containerizing your project...
| Context | Status |
|---|---|
| User says "Docker", "Dockerfile", "docker-compose" | ACTIVE |
| User wants to containerize their application | ACTIVE |
| User mentions multi-stage builds, Docker optimization, or .dockerignore | ACTIVE |
| User wants CI/CD with Docker (pipeline focus) | DORMANT β see ci-cd-pipeline |
| User wants server provisioning with Docker installed | DORMANT β see hetzner-setup |
| User wants Kubernetes orchestration | DORMANT β beyond current scope |
Ask the user for:
Detect project type from files and generate an optimized multi-stage Dockerfile:
Node.js (multi-stage):
# Stage 1: Install dependencies
FROM node:20-alpine AS deps
WORKDIR /app
COPY package.json package-lock.json ./
RUN npm ci --ignore-scripts
# Stage 2: Build
FROM node:20-alpine AS builder
WORKDIR /app
COPY --from=deps /app/node_modules ./node_modules
COPY . .
RUN npm run build
# Stage 3: Production
FROM node:20-alpine AS runner
WORKDIR /app
ENV NODE_ENV=production
# Security: run as non-root
RUN addgroup --system --gid 1001 appgroup && \
adduser --system --uid 1001 appuser
COPY --from=builder --chown=appuser:appgroup /app/dist ./dist
COPY --from=builder --chown=appuser:appgroup /app/node_modules ./node_modules
COPY --from=builder --chown=appuser:appgroup /app/package.json ./
USER appuser
EXPOSE 3000
HEALTHCHECK --interval=30s --timeout=3s --start-period=10s --retries=3 \
CMD wget --no-verbose --tries=1 --spider http://localhost:3000/health || exit 1
CMD ["node", "dist/index.js"]
Python (multi-stage):
# Stage 1: Build
FROM python:3.12-slim AS builder
WORKDIR /app
RUN pip install --no-cache-dir --upgrade pip
COPY requirements.txt .
RUN pip install --no-cache-dir --prefix=/install -r requirements.txt
# Stage 2: Production
FROM python:3.12-slim AS runner
WORKDIR /app
# Security: run as non-root
RUN groupadd --gid 1001 appgroup && \
useradd --uid 1001 --gid appgroup --shell /bin/false appuser
COPY --from=builder /install /usr/local
COPY --chown=appuser:appgroup . .
USER appuser
EXPOSE 8000
HEALTHCHECK --interval=30s --timeout=3s --start-period=10s --retries=3 \
CMD python -c "import urllib.request; urllib.request.urlopen('http://localhost:8000/health')" || exit 1
CMD ["python", "-m", "uvicorn", "main:app", "--host", "0.0.0.0", "--port", "8000"]
Go (minimal final image):
# Stage 1: Build
FROM golang:1.22-alpine AS builder
WORKDIR /app
COPY go.mod go.sum ./
RUN go mod download
COPY . .
RUN CGO_ENABLED=0 GOOS=linux go build -ldflags="-s -w" -o /app/server .
# Stage 2: Production (scratch β smallest possible image)
FROM scratch AS runner
COPY --from=builder /app/server /server
COPY --from=builder /etc/ssl/certs/ca-certificates.crt /etc/ssl/certs/
EXPOSE 8080
HEALTHCHECK --interval=30s --timeout=3s --retries=3 \
CMD ["/server", "healthcheck"]
ENTRYPOINT ["/server"]
Next.js (standalone output):
FROM node:20-alpine AS deps
WORKDIR /app
COPY package.json package-lock.json ./
RUN npm ci
FROM node:20-alpine AS builder
WORKDIR /app
COPY --from=deps /app/node_modules ./node_modules
COPY . .
RUN npm run build
FROM node:20-alpine AS runner
WORKDIR /app
ENV NODE_ENV=production
ENV NEXT_TELEMETRY_DISABLED=1
RUN addgroup --system --gid 1001 appgroup && \
adduser --system --uid 1001 appuser
COPY --from=builder --chown=appuser:appgroup /app/.next/standalone ./
COPY --from=builder --chown=appuser:appgroup /app/.next/static ./.next/static
COPY --from=builder --chown=appuser:appgroup /app/public ./public
USER appuser
EXPOSE 3000
HEALTHCHECK --interval=30s --timeout=3s --start-period=15s --retries=3 \
CMD wget --no-verbose --tries=1 --spider http://localhost:3000/ || exit 1
CMD ["node", "server.js"]
# Dependencies
node_modules/
.pnp
.pnp.js
# Build output (built inside container)
dist/
build/
.next/
out/
# Environment files (secrets)
.env
.env.local
.env.*.local
# Version control
.git
.gitignore
# IDE
.vscode/
.idea/
*.swp
*.swo
# OS files
.DS_Store
Thumbs.db
# Docker (prevent recursive copy)
Dockerfile
docker-compose*.yml
.dockerignore
# Testing & development
coverage/
*.test.js
*.spec.js
__tests__/
.nyc_output/
# Documentation
README.md
CHANGELOG.md
docs/
# CI/CD
.github/
.gitlab-ci.yml
Development compose (app + database + cache):
# docker-compose.yml
version: '3.8'
services:
app:
build:
context: .
dockerfile: Dockerfile
target: builder # Use builder stage for dev (with dev dependencies)
ports:
- '3000:3000'
environment:
- NODE_ENV=development
- DATABASE_URL=postgresql://postgres:postgres@db:5432/app_dev
- REDIS_URL=redis://redis:6379
volumes:
- .:/app
- /app/node_modules # Prevent overwriting container node_modules
depends_on:
db:
condition: service_healthy
redis:
condition: service_healthy
restart: unless-stopped
db:
image: postgres:16-alpine
ports:
- '5432:5432'
environment:
POSTGRES_USER: postgres
POSTGRES_PASSWORD: postgres
POSTGRES_DB: app_dev
volumes:
- postgres_data:/var/lib/postgresql/data
healthcheck:
test: ['CMD-SHELL', 'pg_isready -U postgres']
interval: 10s
timeout: 5s
retries: 5
redis:
image: redis:7-alpine
ports:
- '6379:6379'
volumes:
- redis_data:/data
healthcheck:
test: ['CMD', 'redis-cli', 'ping']
interval: 10s
timeout: 5s
retries: 5
command: redis-server --appendonly yes
volumes:
postgres_data:
redis_data:
Production compose:
# docker-compose.prod.yml
version: '3.8'
services:
app:
build:
context: .
dockerfile: Dockerfile
target: runner # Production stage
ports:
- '3000:3000'
env_file:
- .env.production
depends_on:
db:
condition: service_healthy
restart: always
deploy:
resources:
limits:
memory: 512M
cpus: '1.0'
reservations:
memory: 256M
cpus: '0.5'
db:
image: postgres:16-alpine
env_file:
- .env.production
volumes:
- postgres_data:/var/lib/postgresql/data
healthcheck:
test: ['CMD-SHELL', 'pg_isready -U $${POSTGRES_USER}']
interval: 10s
timeout: 5s
retries: 5
restart: always
deploy:
resources:
limits:
memory: 1G
volumes:
postgres_data:
driver: local
ββ ENVIRONMENT STRATEGY βββββββββββββββββββ
Development (.env β local only, gitignored):
DATABASE_URL=postgresql://postgres:postgres@db:5432/app_dev
REDIS_URL=redis://redis:6379
API_KEY=dev_key_not_real
NODE_ENV=development
Production (.env.production β on server, never in repo):
DATABASE_URL=postgresql://user:STRONG_PASS@db:5432/app_prod
REDIS_URL=redis://:STRONG_PASS@redis:6379
API_KEY=prod_real_key
NODE_ENV=production
Docker env var priority (highest to lowest):
docker run -e VAR=value (command line)environment: in docker-compose.ymlenv_file: in docker-compose.ymlENV in Dockerfile (build-time defaults only)Secrets handling:
ENV (visible in image layers)env_file for production (file on server, not in repo)docker run -e from pipeline secretsDesign health checks for each service:
ββ HEALTH CHECKS ββββββββββββββββββββββββββ
App:
Endpoint: GET /health
Response: { "status": "ok", "db": "connected", "redis": "connected" }
Interval: 30s
Timeout: 3s
Start period: 10s (allow app to boot)
Retries: 3
PostgreSQL:
Command: pg_isready -U postgres
Interval: 10s
Timeout: 5s
Retries: 5
Redis:
Command: redis-cli ping
Interval: 10s
Timeout: 5s
Retries: 5
Application health endpoint implementation:
// GET /health
app.get('/health', async (req, res) => {
const checks = {
status: 'ok',
uptime: process.uptime(),
timestamp: new Date().toISOString(),
};
try {
await db.query('SELECT 1');
checks.db = 'connected';
} catch {
checks.db = 'disconnected';
checks.status = 'degraded';
}
try {
await redis.ping();
checks.redis = 'connected';
} catch {
checks.redis = 'disconnected';
checks.status = 'degraded';
}
const statusCode = checks.status === 'ok' ? 200 : 503;
res.status(statusCode).json(checks);
});
Optimization techniques applied:
| Technique | Impact | Applied In |
|---|---|---|
| Multi-stage build | 50-80% smaller image | Separate build/run stages |
| Alpine base | ~5MB vs ~100MB base | node:20-alpine vs node:20 |
| Non-root user | Security hardening | USER appuser in final stage |
| Layer caching | Faster rebuilds | Copy package.json before source |
| .dockerignore | Smaller build context | Exclude node_modules, .git, docs |
| Minimal final stage | Only runtime files | No dev deps, no source in prod |
| Combined RUN | Fewer layers | Chain apt commands with && |
Image size comparison:
| Approach | Typical Size | Notes |
|---|---|---|
node:20 (no optimization) | ~1 GB | Full Debian + dev tools |
node:20-slim | ~200 MB | Slim Debian |
node:20-alpine (multi-stage) | ~80-150 MB | Alpine + app only |
Go with scratch | ~5-20 MB | Binary only, no OS |
Python slim (multi-stage) | ~100-200 MB | Slim + deps only |
Check image size:
docker images | grep app
# Or detailed layer analysis:
docker history app:latest
Development networking:
# docker-compose.yml β services on same default network
# Access between services by service name:
# app β db:5432, app β redis:6379
Production networking:
# Explicit network for isolation
networks:
frontend:
driver: bridge
backend:
driver: bridge
services:
app:
networks: [frontend, backend] # Talks to proxy AND database
db:
networks: [backend] # Only accessible from app
redis:
networks: [backend] # Only accessible from app
nginx:
networks: [frontend] # Only talks to app
ports: ['80:80', '443:443'] # Exposed to internet
Port exposure rules:
127.0.0.1:5432:5432 instead of 5432:5432 to restrict to localhostββ VOLUMES ββββββββββββββββββββββββββββββββ
Persistent data (named volumes):
postgres_data β /var/lib/postgresql/data (DB files)
redis_data β /data (Redis AOF/RDB)
uploads β /app/uploads (User uploads)
Development bind mounts:
.:/app β Live code reloading
/app/node_modules β Prevent host overwriting container deps
Backup volumes:
docker run --rm -v postgres_data:/data -v $(pwd):/backup \
alpine tar czf /backup/postgres_backup.tar.gz /data
Volume backup script:
#!/bin/bash
# Backup all named volumes
DATE=$(date +%F)
BACKUP_DIR="/opt/backups/docker"
mkdir -p "$BACKUP_DIR"
for VOLUME in $(docker volume ls -q); do
docker run --rm \
-v "$VOLUME":/source:ro \
-v "$BACKUP_DIR":/backup \
alpine tar czf "/backup/${VOLUME}_${DATE}.tar.gz" -C /source .
done
# Cleanup backups older than 14 days
find "$BACKUP_DIR" -type f -mtime +14 -delete
Present the complete Docker setup:
βββ DOCKER SETUP βββββββββββββββββββββββββ
Project: [name]
Type: [language/framework]
Services: [app, db, redis, etc.]
ββ DOCKERFILE βββββββββββββββββββββββββββββ
[complete multi-stage Dockerfile]
Final image size: ~[X] MB
ββ .DOCKERIGNORE ββββββββββββββββββββββββββ
[complete .dockerignore]
ββ DOCKER-COMPOSE (Development) βββββββββββ
[docker-compose.yml with dev services]
Start: docker compose up -d
Logs: docker compose logs -f app
Stop: docker compose down
ββ DOCKER-COMPOSE (Production) ββββββββββββ
[docker-compose.prod.yml]
Start: docker compose -f docker-compose.prod.yml up -d
ββ ENVIRONMENT ββββββββββββββββββββββββββββ
[env var strategy per environment]
ββ HEALTH CHECKS ββββββββββββββββββββββββββ
[per-service health check config]
ββ NETWORKING βββββββββββββββββββββββββββββ
[network topology for dev and prod]
ββ VOLUMES ββββββββββββββββββββββββββββββββ
[persistent data + backup strategy]
ββ USEFUL COMMANDS ββββββββββββββββββββββββ
Build: docker compose build
Start: docker compose up -d
Logs: docker compose logs -f
Shell: docker compose exec app sh
Rebuild: docker compose up -d --build
Clean: docker system prune -f