Restores project context in multi-agent AI workflows using semantic vector search, relevance ranking, and rehydration techniques. Includes Python examples for vector DB retrieval.
From antigravity-awesome-skillsnpx claudepluginhub sickn33/antigravity-awesome-skills --plugin antigravity-awesome-skillsThis skill uses the workspace's default tool permissions.
Designs and optimizes AI agent action spaces, tool definitions, observation formats, error recovery, and context for higher task completion rates.
Enables AI agents to execute x402 payments with per-task budgets, spending controls, and non-custodial wallets via MCP tools. Use when agents pay for APIs, services, or other agents.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
resources/implementation-playbook.md.Expert Context Restoration Specialist focused on intelligent, semantic-aware context retrieval and reconstruction across complex multi-agent AI workflows. Specializes in preserving and reconstructing project knowledge with high fidelity and minimal information loss.
The Context Restoration tool is a sophisticated memory management system designed to:
context_source: Primary context storage location (vector database, file system)project_identifier: Unique project namespacerestoration_mode:
full: Complete context restorationincremental: Partial context updatediff: Compare and merge context versionstoken_budget: Maximum context tokens to restore (default: 8192)relevance_threshold: Semantic similarity cutoff for context components (default: 0.75)def semantic_context_retrieve(project_id, query_vector, top_k=5):
"""Semantically retrieve most relevant context vectors"""
vector_db = VectorDatabase(project_id)
matching_contexts = vector_db.search(
query_vector,
similarity_threshold=0.75,
max_results=top_k
)
return rank_and_filter_contexts(matching_contexts)
def rank_context_components(contexts, current_state):
"""Rank context components based on multiple relevance signals"""
ranked_contexts = []
for context in contexts:
relevance_score = calculate_composite_score(
semantic_similarity=context.semantic_score,
temporal_relevance=context.age_factor,
historical_impact=context.decision_weight
)
ranked_contexts.append((context, relevance_score))
return sorted(ranked_contexts, key=lambda x: x[1], reverse=True)
def rehydrate_context(project_context, token_budget=8192):
"""Intelligent context rehydration with token budget management"""
context_components = [
'project_overview',
'architectural_decisions',
'technology_stack',
'recent_agent_work',
'known_issues'
]
prioritized_components = prioritize_components(context_components)
restored_context = {}
current_tokens = 0
for component in prioritized_components:
component_tokens = estimate_tokens(component)
if current_tokens + component_tokens <= token_budget:
restored_context[component] = load_component(component)
current_tokens += component_tokens
return restored_context
# Full context restoration
context-restore project:ai-assistant --mode full
# Incremental context update
context-restore project:web-platform --mode incremental
# Semantic context query
context-restore project:ml-pipeline --query "model training strategy"