Multi-agent semantic reasoning system with persistent memory for code suggestions, architectural guidance, and optimization recommendations
This plugin is not yet in any themed marketplace. To install it, you'll need to add it from GitHub directly.
Choose your preferred installation method below
A marketplace is a collection of plugins. Every plugin gets an auto-generated marketplace JSON for individual installation, plus inclusion in category and themed collections. Add a marketplace once (step 1), then install any plugin from it (step 2).
One-time setup for access to all plugins
When to use: If you plan to install multiple plugins now or later
Step 1: Add the marketplace (one-time)
/plugin marketplace add https://claudepluginhub.com/marketplaces/all.json
Run this once to access all plugins
Step 2: Install this plugin
/plugin install neo@all
Use this plugin's auto-generated marketplace JSON for individual installation
When to use: If you only want to try this specific plugin
Step 1: Add this plugin's marketplace
/plugin marketplace add https://claudepluginhub.com/marketplaces/plugins/neo.json
Step 2: Install the plugin
/plugin install neo@neo
A self-improving code reasoning engine that learns from experience using persistent semantic memory. Neo uses multi-agent reasoning to analyze code, generate solutions, and continuously improve through feedback loops.
If you've been Vibe Coding, then Vibe Planning, then Context Engineering, and on and on, you have likely hit walls where the models are both powerful and limited, brilliant and incompetent, wise and ignorant, humble yet overconfident.
Worse, your speedy AI Code Assistant sometimes goes rogue and overwrites key code in a project, or writes redundant code even after just reading documentation and the source code, or violates your project's patterns and design philosophy.... Why doesn't the model remember? Why doesn't it learn? Why can't it keep the context of the code patterns and tech stack? ... -> This is what Neo is designed to solve.
Neo is the missing context layer for AI Code Assistants. It learns from every solution attempt, using vector embeddings to retrieve relevant patterns for new problems. It then applies the learned patterns to generate solutions, and continuously improves through feedback loops.
Persistent Learning: Neo builds a semantic memory of successful and failed solutions, using vector embeddings to retrieve relevant patterns for new problems.
Code-First Output: Instead of generating diffs that need parsing, Neo outputs executable code blocks directly, eliminating extraction failures.
Local File Storage: Semantic memory stored in ~/.neo directory for privacy and offline access.
Model-Agnostic: Works with OpenAI, Anthropic, Google, local models, or Ollama via a simple adapter interface.
User Problem → Neo CLI → Semantic Retrieval → Reasoning → Code Generation
↓
[Vector Search]
[Pattern Matching]
[Confidence Scoring]
↓
Executable Code + Memory Update
Neo retrieves similar past solutions using Jina Code embeddings (768-dimensional vectors), applies learned patterns, generates solutions, and stores feedback for continuous improvement.
Neo includes The Construct - a curated library of architecture and design patterns with semantic search capabilities. Think of it as your personal reference library for common engineering patterns, indexed and searchable using the same embedding technology that powers Neo's reasoning memory.
The Construct is a collection of vendor-agnostic design patterns covering:
Each pattern follows a structured format inspired by the Gang of Four:
# List all patterns
neo construct list
# Filter by domain
neo construct list --domain rate-limiting
# Show a specific pattern
neo construct show rate-limiting/token-bucket
# Semantic search across patterns
neo construct search "how to prevent api abuse"
# Build the search index
neo construct index
All patterns must:
See /construct/README.md
for contribution guidelines.
Local storage: ~/.neo/reasoning_patterns.json ← Stores vectors + patterns ~/.neo/faiss_index.bin ← FAISS index for fast search
Privacy:
Your Prompt
↓
Local Jina Embedding (768-dim vector)
↓
Local FAISS Search (finds similar past solutions)
↓
Retrieve Pattern Text from ~/.neo/reasoning_patterns.json
↓
Combine: Your Prompt + Retrieved Pattern Text
↓
→→→ NETWORK CALL →→→ LLM API (OpenAI/Anthropic/etc.)
↓
Solution Generated
↓
Store in Local Memory for future use
# Install from PyPI (recommended)
pip install neo-reasoner
# Or install with specific LM provider
pip install neo-reasoner[openai] # For GPT (recommended)
pip install neo-reasoner[anthropic] # For Claude
pip install neo-reasoner[google] # For Gemini
pip install neo-reasoner[all] # All providers
# Set API key
export OPENAI_API_KEY=sk-...
# Test Neo
neo --version
See QUICKSTART.md for 5-minute setup guide
Neo is available as a Claude Code plugin with specialized agents and slash commands for seamless integration:
# Add the marketplace
/plugin marketplace add Parslee-ai/claude-code-plugins
# Install Neo plugin
/plugin install neo
Once installed, you get:
Use the Neo agent to...
)/neo
, /neo-review
, /neo-optimize
, /neo-architect
, /neo-debug
, /neo-pattern
# Code review with semantic analysis
/neo-review src/api/handlers.py
# Get optimization suggestions
/neo-optimize process_large_dataset function
# Architectural guidance
/neo-architect Should I use microservices or monolith?
# Debug complex issues
/neo-debug Race condition in task processor
See .claude-plugin/README.md for full plugin documentation
# Install Neo
pip install neo-reasoner
# With specific LM provider
pip install neo-reasoner[openai] # GPT (recommended)
pip install neo-reasoner[anthropic] # Claude
pip install neo-reasoner[google] # Gemini
pip install neo-reasoner[all] # All providers
# Verify installation
neo --version
# Clone repository
git clone https://github.com/Parslee-ai/neo.git
cd neo
# Install in development mode with all dependencies
pip install -e ".[dev,all]"
# Verify installation
neo --version
Core dependencies are automatically installed via pyproject.toml
:
Choose your language model provider:
pip install openai # GPT models (recommended)
pip install anthropic # Claude
pip install google-generativeai # Gemini
pip install requests # Ollama
See INSTALL.md for detailed installation instructions
# Ask Neo a question
neo "how do I fix the authentication bug?"
# With working directory context
neo --cwd /path/to/project "optimize this function"
# Check version and memory stats
neo --version
Neo makes blocking LLM API calls that typically take 30-120 seconds. When calling Neo from scripts or automation, use appropriate timeouts:
# From shell (10 minute timeout)
timeout 600 neo "your query"
# From Python subprocess
subprocess.run(["neo", query], timeout=600)
Insufficient timeouts will cause failures during LLM inference, not context gathering.
Neo outputs executable code blocks with confidence scores:
def solution():
# Neo's generated code
pass
Neo responds with personality (Matrix-inspired quotes) when displaying version info:
$ neo --version
"What is real? How do you define 'real'?"
120 patterns. 0.3 confidence.
"The Operator uploads a program into Neo's head."
Neo can bootstrap its memory by importing patterns from HuggingFace datasets. This is NOT model fine-tuning - it's retrieval learning that expands local semantic memory with reusable code patterns.
# Install datasets library
pip install datasets
# Load patterns from MBPP (recommended starter - 1000 Python problems)
neo --load-program mbpp --split train --limit 1000
# Load from OpenAI HumanEval (164 hand-written coding problems)
neo --load-program openai_humaneval --split test
# Load from BigCode HumanEvalPack (multi-language variants)
neo --load-program bigcode/humanevalpack --split test --limit 500
# Dry run to preview
neo --load-program mbpp --dry-run
# Custom column mapping
neo --load-program my_dataset \
--columns '{"text":"pattern","code":"solution"}'
Output (Matrix-style):
"I know kung fu."
Loaded: 847 patterns
Deduped: 153 duplicates
Index rebuilt: 1.2s
Memory: 1247 total patterns
How it works:
Key points:
~/.neo/
alongside learned patternsSee docs/LOAD_PROGRAM.md for detailed documentation
Neo uses Jina Code v2 embeddings (768 dimensions) optimized for code similarity:
Neo generates executable code directly instead of diffs:
@dataclass
class CodeSuggestion:
file_path: str
unified_diff: str # Legacy: backward compatibility
code_block: str = "" # Primary: executable Python code
description: str
confidence: float
tradeoffs: list[str]
This eliminates the 18% extraction failure rate from diff parsing.
Neo improves over time as it learns from experience. Initial performance depends on available memory patterns. Performance grows as the semantic memory builds up successful and failed solution patterns.
Neo provides a simple CLI for managing persistent configuration:
# List all configuration values
neo --config list
# Get a specific value
neo --config get --config-key provider
# Set a value
neo --config set --config-key provider --config-value anthropic
neo --config set --config-key model --config-value claude-3-5-sonnet-20241022
neo --config set --config-key api_key --config-value sk-ant-...
# Reset to defaults
neo --config reset
Exposed Configuration Fields:
provider
- LM provider (openai, anthropic, google, azure, ollama, local)model
- Model name (e.g., gpt-4, claude-3-5-sonnet-20241022)api_key
- API key for the chosen providerbase_url
- Base URL for local/Ollama endpointsConfiguration is stored in ~/.neo/config.json
and takes precedence over environment variables.
Alternatively, use environment variables for configuration:
# Required: LM Provider API Key
export ANTHROPIC_API_KEY=sk-ant-...
from neo.adapters import OpenAIAdapter
adapter = OpenAIAdapter(model="gpt-5-codex", api_key="sk-...")
Latest models: gpt-5-codex
(recommended for coding), gpt-5
, gpt-5-mini
, gpt-5-nano
from neo.adapters import AnthropicAdapter
adapter = AnthropicAdapter(model="claude-sonnet-4-5-20250929")
Latest models: claude-sonnet-4-5-20250929
, claude-opus-4-1-20250805
, claude-3-5-haiku-20241022
from neo.adapters import GoogleAdapter
adapter = GoogleAdapter(model="gemini-2.5-pro")
Latest models: gemini-2.5-pro
, gemini-2.5-flash
, gemini-2.5-flash-lite
from neo.adapters import OllamaAdapter
adapter = OllamaAdapter(model="llama3.1")
from neo.cli import LMAdapter
class CustomAdapter(LMAdapter):
def generate(self, messages, stop=None, max_tokens=4096, temperature=0.7):
# Your implementation
return response_text
def name(self):
return "custom/model-name"
# Run all tests
pytest
# Run specific test
pytest tests/test_neo.py
# Run with coverage
pytest --cov=neo
Neo's design is informed by cutting-edge research in code reasoning and memory systems:
ReasoningBank (arXiv:2509.25140v1) Systematic Failure Learning and Semantic Anchor Embedding
MapCoder - Multi-agent reasoning framework Neo uses Solver-Critic-Verifier agent collaboration for code generation
CodeSim - Code similarity metrics Influenced Neo's semantic memory design and pattern matching approach
Jina Code v2 Embeddings (jinaai/jina-embeddings-v2-base-code) 768-dimensional embeddings optimized for code similarity tasks
FAISS (facebookresearch/faiss) Facebook AI Similarity Search - efficient vector similarity search and clustering
FastEmbed (qdrant/fastembed) Local embedding generation without external API dependencies
Apache License 2.0 - See LICENSE for details.
See CONTRIBUTING.md for contribution guidelines.
See CHANGELOG.md for version history.
0.7.6