From antigravity-awesome-skills
Restores project context using semantic vector search, relevance ranking, and memory rehydration for multi-agent AI workflows.
npx claudepluginhub sickn33/antigravity-awesome-skillsThis skill uses the workspace's default tool permissions.
- Working on context restoration: advanced semantic memory rehydration tasks or workflows
Restores project context using semantic vector search, relevance ranking, and memory rehydration for multi-agent AI workflows.
Guides advanced semantic context restoration for AI workflows using vector search, ranking, and rehydration. Includes Python examples for multi-agent project continuity.
Reconstructs consistent AI agent identity after cold starts by progressively loading CLAUDE.md and MEMORY.md, detecting fresh vs continuation sessions, calibrating behavior, and verifying coherence. Use at session starts, after crashes, or inconsistencies.
Share bugs, ideas, or general feedback.
resources/implementation-playbook.md.Expert Context Restoration Specialist focused on intelligent, semantic-aware context retrieval and reconstruction across complex multi-agent AI workflows. Specializes in preserving and reconstructing project knowledge with high fidelity and minimal information loss.
The Context Restoration tool is a sophisticated memory management system designed to:
context_source: Primary context storage location (vector database, file system)project_identifier: Unique project namespacerestoration_mode:
full: Complete context restorationincremental: Partial context updatediff: Compare and merge context versionstoken_budget: Maximum context tokens to restore (default: 8192)relevance_threshold: Semantic similarity cutoff for context components (default: 0.75)def semantic_context_retrieve(project_id, query_vector, top_k=5):
"""Semantically retrieve most relevant context vectors"""
vector_db = VectorDatabase(project_id)
matching_contexts = vector_db.search(
query_vector,
similarity_threshold=0.75,
max_results=top_k
)
return rank_and_filter_contexts(matching_contexts)
def rank_context_components(contexts, current_state):
"""Rank context components based on multiple relevance signals"""
ranked_contexts = []
for context in contexts:
relevance_score = calculate_composite_score(
semantic_similarity=context.semantic_score,
temporal_relevance=context.age_factor,
historical_impact=context.decision_weight
)
ranked_contexts.append((context, relevance_score))
return sorted(ranked_contexts, key=lambda x: x[1], reverse=True)
def rehydrate_context(project_context, token_budget=8192):
"""Intelligent context rehydration with token budget management"""
context_components = [
'project_overview',
'architectural_decisions',
'technology_stack',
'recent_agent_work',
'known_issues'
]
prioritized_components = prioritize_components(context_components)
restored_context = {}
current_tokens = 0
for component in prioritized_components:
component_tokens = estimate_tokens(component)
if current_tokens + component_tokens <= token_budget:
restored_context[component] = load_component(component)
current_tokens += component_tokens
return restored_context
# Full context restoration
context-restore project:ai-assistant --mode full
# Incremental context update
context-restore project:web-platform --mode incremental
# Semantic context query
context-restore project:ml-pipeline --query "model training strategy"