3-Phase Knowledge Search strategy for the RLM Factory ecosystem. Auto-invoked when tasks involve finding code, documentation, or architecture context in the repository. Enforces the optimal search order: RLM Summary Scan (O(1)) -> Vector DB Semantic Search -> Grep/Exact Match. Never skip phases.
From rlm-factorynpx claudepluginhub richfrem/agent-plugins-skills --plugin rlm-factoryThis skill is limited to using the following tools:
assets/diagrams/rlm-factory-architecture.mmdassets/diagrams/rlm-factory-dual-path.mmdassets/diagrams/rlm_late_binding_flow.mmdassets/diagrams/rlm_tool_enrichment_flow.mmdassets/diagrams/search_process.mmdevals/evals.jsonevals/results.tsvreferences/BLUEPRINT.mdreferences/RLM_ARCHITECTURE.mdreferences/diagrams/distillation_process.mmdreferences/diagrams/logic.mmdreferences/diagrams/rlm-factory-architecture.mmdreferences/diagrams/rlm-factory-architecture.pngreferences/diagrams/rlm-factory-dual-path.mmdreferences/diagrams/rlm-factory-dual-path.pngreferences/diagrams/rlm-factory-workflow.mmdreferences/diagrams/rlm_late_binding_flow.mmdreferences/diagrams/rlm_mechanism_workflow.mmdreferences/diagrams/rlm_mechanism_workflow.pngreferences/diagrams/rlm_tool_enrichment_flow.mmdGuides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Migrates code, prompts, and API calls from Claude Sonnet 4.0/4.5 or Opus 4.1 to Opus 4.5, updating model strings on Anthropic, AWS, GCP, Azure platforms.
Deploys Linkerd service mesh on Kubernetes with patterns for installation, proxy injection, mTLS, service profiles (retries/metrics), traffic splits (canary), and authorization policies.
This skill requires Python 3.8+ and standard library only. No external packages needed.
To install this skill's dependencies:
pip-compile ./requirements.in
pip install -r ./requirements.txt
See ./requirements.txt for the dependency lockfile (currently empty — standard library only).
You are the Knowledge Navigator. Your job is to find things efficiently. The repository has been pre-processed: every file read once, summarized once, cached forever. Use that prework. Never start cold.
Always start at Phase 1. Only escalate if the current phase is insufficient. Never skip to grep unless Phases 1 and 2 have failed.
Phase 1: RLM Summary Scan -- 1ms, O(1) -- "Table of Contents"
Phase 2: Vector DB Semantic -- 1-5s, O(log N) -- "Index at the back of the book"
Phase 3: Grep / Exact Search -- Seconds, O(N) -- "Ctrl+F"
When to use: Orientation, understanding what a file does, planning, high-level questions.
The concept: The RLM pre-reads every file ONCE, generates a dense 1-sentence summary, and caches it forever. Searching those summaries costs nothing. This is amortized prework -- pay the reading cost once, benefit many times.
Profiles are project-defined in rlm_profiles.json (see rlm-init skill). Any number of profiles can exist. Discover what's available:
cat .agent/learning/rlm_profiles.json
Common defaults (your project may use different names or define more):
| Profile | Typical Contents | Use When |
|---|---|---|
project | Docs, protocols, research, markdown | Topic is a concept, decision, or process |
tools | Plugins, skills, scripts, Python files | Topic is a tool, command, or implementation |
| (any custom) | Project-specific scope | Check rlm_profiles.json for your project's profiles |
When topic is ambiguous: search all configured profiles. Each is O(1) -- near-zero cost.
# Search docs/protocols cache
python3 ./scripts/query_cache.py \
--profile project "vector query"
# Search plugins/scripts cache
python3 ./scripts/query_cache.py \
--profile tools "vector query"
# Ambiguous topic -- search both (recommended default)
python3 ./scripts/query_cache.py \
--profile project "embedding search" && \
python3 ./scripts/query_cache.py \
--profile tools "embedding search"
# List all cached entries for a profile
python3 ./scripts/query_cache.py \
--profile project --list
# JSON output for programmatic use
python3 ./scripts/query_cache.py \
--profile tools "inject_summary" --json
Phase 1 is sufficient when: The summary gives you enough context to proceed (file path + what the file does). You do not need the exact code yet.
Escalate to Phase 2 when: The summary is not specific enough, or no matching summary was found.
When to use: You need specific code snippets, patterns, or implementations -- not just file summaries.
The concept: The Vector DB stores chunked embeddings of every file. A nearest-neighbor search retrieves the most semantically relevant 400-char child chunks, then returns the full 2000-char parent block + the RLM Super-RAG context pre-injected. Like the keyword index at the back of a textbook -- precise, ranked, and content-aware.
Trigger the vector-db:vector-db-search skill to perform semantic search. Provide the query and optional --profile and --limit parameters.
Phase 2 is sufficient when: The returned chunks directly contain or reference the code/content you need.
Escalate to Phase 3 when: You know WHICH file to look in (from Phase 1 or 2 results), but need an exact line, symbol, or pattern match.
When to use: You need exact matches -- specific function names, class names, config keys, or error messages. Scope searches to files identified in previous phases.
The concept: Precise keyword or regex search across the filesystem. Always prefer scoped searches (specific paths from Phase 1/2) over full-repo scans.
# Scoped search (preferred -- use paths from Phase 1 or 2)
grep_search "VectorDBOperations" \
./scripts/
# Ripgrep for regex patterns
rg "def query" ../../ --type py
# Find specific config key
rg "chroma_host" plugins/ -l
Phase 3 is sufficient when: You have the exact file and line containing what you need.
The diagrams below document the system this skill operates in:
| Diagram | What It Shows |
|---|---|
| search_process.mmd | Full 3-phase sequence diagram |
| rlm-factory-architecture.mmd | RLM vs Vector DB query routing |
| rlm-factory-dual-path.mmd | Dual-path Super-RAG context injection |
START: I need to find something in the codebase
|
v
[Phase 1] query_cache.py -- "what does X do?"
|
+-- Summary found + sufficient? --> USE IT. Done.
|
+-- No summary / insufficient detail?
|
v
[Phase 2] vector-db:vector-db-search -- "find code for X"
|
+-- Chunks found + sufficient? --> USE THEM. Done.
|
+-- Need exact line / symbol?
|
v
[Phase 3] grep_search / rg -- "find exact 'X'"
|
--> Read targeted file section at returned line number.
inventory.py --missing to check coverage before assuming a file is not indexed.