Configure neuroflow integrations — Miro, Google Workspace, and custom LLM providers. Use when setting up credentials, checking integration status, or guiding a user through connecting external services. Also covers e-INFRA CZ connection for Czech academic researchers.
From neuroflownpx claudepluginhub stanislavjiricek/neuroflow --plugin neuroflowThis skill uses the workspace's default tool permissions.
references/einfra-cc.mdscripts/proxy.mjsDispatches parallel agents to independently tackle 2+ tasks like separate test failures or subsystems without shared state or dependencies.
Executes pre-written implementation plans: critically reviews, follows bite-sized steps exactly, runs verifications, tracks progress with checkpoints, uses git worktrees, stops on blockers.
Guides idea refinement into designs: explores context, asks questions one-by-one, proposes approaches, presents sections for approval, writes/review specs before coding.
Agent-facing knowledge for all neuroflow integrations. Use this skill when a user asks about credentials, integration status, or setting up an external service — without necessarily running the full /setup wizard.
| Integration | Credential required | Key |
|---|---|---|
| PubMed / bioRxiv | ❌ No | — (handled by biorxiv MCP server) |
| Miro | ✅ Yes | MIRO_ACCESS_TOKEN |
| Context7 | ❌ No | — |
Google Workspace CLI (gws) | ✅ Yes | OAuth via gws auth login |
| Custom LLM provider | ✅ Yes (API key) | custom_llm.api_key in integrations.json |
neuroflow supports two credential scopes:
| Scope | Location | When to use |
|---|---|---|
| Global (device-wide) | ~/.neuroflow/integrations.json | Set once — applies to all projects on this machine |
| Per-project | .neuroflow/integrations.json | Overrides global for this project only |
Resolution order: per-project credentials take precedence over global. If a key is not found per-project, fall back to global.
Platform paths for global config:
~/.neuroflow/integrations.json%USERPROFILE%\.neuroflow\integrations.json (e.g. C:\Users\YourName\.neuroflow\integrations.json)When guiding a user, ask:
"Save credentials for this project only, or globally on this machine (all projects)?"
Default recommendation: global, so they don't repeat setup on every new project.
os.platform() or check $OSTYPE / $env:OS).~/.neuroflow/integrations.json (or %USERPROFILE%\.neuroflow\integrations.json on Windows) if it exists..neuroflow/integrations.json if it exists. Per-project keys override global keys.MIRO_ACCESS_TOKEN, etc.).gws: run gws --version 2>/dev/null or which gws (Unix) / where gws (Windows) to detect installation; run gws auth status to check OAuth.integrations.json for a custom_llm key; if present, show provider and base_url.Display a status table:
Integration Status
────────────────────── ──────
PubMed / bioRxiv ✅ no credentials needed
Miro ✅ configured (or ❌ not configured)
Context7 ✅ no credentials needed
Google Workspace CLI ✅ installed (or ❌ not installed)
└─ OAuth credentials ✅ configured (or ❌ not configured)
Custom LLM ✅ configured (provider: einfra) (or ❌ not configured)
Full schema including all integrations:
{
"miro": {
"MIRO_ACCESS_TOKEN": "eyJ..."
},
"google_workspace": {
"GOOGLE_WORKSPACE_CLI_CREDENTIALS_FILE": "/home/user/.config/gws/client_secret.json",
"GOOGLE_WORKSPACE_CLI_CLIENT_ID": "<optional>",
"GOOGLE_WORKSPACE_CLI_CLIENT_SECRET": "<optional>"
},
"custom_llm": {
"provider": "einfra",
"base_url": "https://llm.ai.e-infra.cz/v1",
"api_key": "<YOUR_API_KEY>",
"model": "qwen3.5-122b",
"proxy_port": 3456
}
}
.neuroflow/integrations.json is gitignored — credentials never leave the local machine.api_key fields are always local-only, even if other non-secret fields are synced to a flowie profile.| Integration | Validation |
|---|---|
| Miro | Non-empty, at least 20 characters; eyJ prefix is a good sign (JWT) |
| Google Workspace | File must exist at platform path, or env vars must be set |
| Custom LLM | API key: non-empty; base URL: must start with http; model: optional |
The /setup Step 5 allows configuring any OpenAI-compatible API endpoint as a replacement for Anthropic's API in Claude Code. This is how it works:
ANTHROPIC_BASE_URL to the custom endpoint (e.g. https://llm.ai.e-infra.cz/v1).ANTHROPIC_API_KEY to the custom provider's API key.Model selection limitation: Claude Code always sends claude-* model names in requests. In direct mode, the custom API must accept these names (some do, routing to a default model). To select a specific model, use the proxy script (skills/setup/scripts/proxy.mjs), which intercepts requests and replaces the model name before forwarding.
⚠️ Claude model name in responses: Claude Code validates that the model field in the API response matches the claude-* name it originally sent. If the custom API returns a response with a non-Claude model name (e.g. qwen3.5-122b), Claude Code will error with an "unexpected model" message. The proxy script handles this automatically — it saves the original claude-* model name from the request, then replaces the custom model name back in every response chunk before it reaches Claude Code.
Setting env vars on Windows (PowerShell):
# Direct mode — set for current session
$env:ANTHROPIC_BASE_URL = "https://llm.ai.e-infra.cz/v1"
$env:ANTHROPIC_API_KEY = "<YOUR_API_KEY>"
claude
# Proxy mode
$env:ANTHROPIC_BASE_URL = "http://localhost:3456"
$env:ANTHROPIC_API_KEY = "any"
claude
For persistence on Windows, add these to your PowerShell profile ($PROFILE) or set them as System/User environment variables via Settings → System → Environment Variables.
For the Czech-specific e-INFRA CZ example — including available models, direct mode, proxy mode, and full terminal workflow for both Unix and Windows — read skills/setup/references/einfra-cc.md.
integrations.json (global or per-project), which is gitignored. Never write credentials to any other file.~/.neuroflow/integrations.json) is in the home directory — not inside any repo — so it can never be accidentally committed..neuroflow/integrations.json) is gitignored via the project's .gitignore. Double-check this is in place before any git push..neuroflow/flowie/sync.json), non-secret custom_llm settings can be written to .neuroflow/flowie/integrations.json and pushed to the user's private GitHub repo for cross-machine sync. The api_key field is always excluded from this sync.The full interactive wizard is /neuroflow:setup. It covers all integrations step by step. This skill provides agent-level knowledge so Claude can guide credential setup without running the wizard — for example, answering "how do I configure Miro?" or surfacing the right reference when a user mentions e-INFRA.