Hardware-accelerated persistent memory for AI agents. Local-first. No cloud. One-time payment. 66.9% on LoCoMo benchmark (adjusted). Under 1ms retrieval. Zero cloud dependency. Retrieval pipeline rebuilt from scratch. - bge-small-en-v1.5 bi-encoder + ms-marco cross-encoder reranker (spec-decode architecture) - BM25 + Porter-stemmed BM25 + named entity injection, fused via RRF - MAGMA graph layer —
Add this skill
npx mdskills install Vektor-Memory/vektor-memoryHardware-accelerated local memory system with comprehensive MCP tools, strong benchmarks, and excellent documentation
Hardware-accelerated persistent memory for AI agents. Local-first. No cloud. One-time payment.
66.9% on LoCoMo benchmark (adjusted). Under 1ms retrieval. Zero cloud dependency.
npm install vektor-slipstream
npx vektor setup
const { createMemory } = require('vektor-slipstream');
const memory = await createMemory({
agentId: 'my-agent',
licenceKey: process.env.VEKTOR_LICENCE_KEY,
});
// Store a memory
await memory.remember('User prefers TypeScript over JavaScript');
// Recall by semantic similarity — sub-1ms, fully local
const results = await memory.recall('coding preferences', 5);
// → [{ content, score, id }]
// Traverse the MAGMA graph
const graph = await memory.graph('TypeScript', { hops: 2 });
// What changed in 7 days?
const delta = await memory.delta('project decisions', 7);
// Morning briefing
const brief = await memory.briefing();
// Graph stats
const stats = memory.graphStats();
// → { nodes, edges, entities }
Retrieval pipeline rebuilt from scratch.
vektor_entities) for guaranteed named-entity recallLoCoMo benchmark results (conv 0, 154 valid questions):
| Category | Judge Accuracy |
|---|---|
| Multi-hop | 79.1% |
| Adversarial | 70.4% |
| Temporal | 46.2% |
| Single-hop | 51.6% |
| Adjusted total | 66.9% |
#Under 1ms retrieval latency with zero cloud API calls at query time.
Chat with any LLM with full memory across every session. Zero configuration.
npx vektor chat # start chat (auto-detects Ollama)
npx vektor chat --provider claude # use Anthropic Claude
npx vektor chat --provider groq --model llama-3.3-70b-versatile
npx vektor chat --provider gemini
npx vektor chat --provider openai
| Provider | Details |
|---|---|
ollama | Default — free, local, no API key. Auto-detects best installed model. |
claude | Anthropic Claude — set ANTHROPIC_API_KEY |
openai | OpenAI GPT — set OPENAI_API_KEY |
groq | Groq LLaMA — set GROQ_API_KEY (free tier available) |
gemini | Google Gemini — set GEMINI_API_KEY |
Set a permanent default:
# Windows
$env:VEKTOR_PROVIDER = "claude"
# macOS/Linux
export VEKTOR_PROVIDER=claude
Type / to see available commands with autocomplete. Tab to select, arrow keys to navigate.
| Command | Action |
|---|---|
/recall | Search MAGMA memory mid-conversation |
/stats | Show memory node count, edges, pinned |
/briefing | Generate memory briefing inline |
/exit | Exit chat (Ctrl+C also works) |
# Store a fact
npx vektor remember "I prefer TypeScript over JavaScript"
npx vektor remember "deadline is Friday" --importance 5
# Pipe support
cat meeting-notes.txt | npx vektor remember
# One-shot recall + LLM answer
npx vektor ask "what stack am I using?"
npx vektor ask "what did we decide about the database?"
# Autonomous goal executor
npx vektor agent "summarise everything I know about project Alpha"
npx vektor agent "research AI memory tools" --steps 15 --provider groq
VEKTOR queries http://localhost:11434/api/tags and picks the best available model:
qwen3 → qwen2 → llama → mistral → first available.
Override:
$env:OLLAMA_MODEL = "qwen3.5:4b"
export OLLAMA_MODEL=qwen3.5:4b
npx vektor setup # First-run wizard — licence, hardware, integrations
npx vektor activate # Activate licence key on this machine
npx vektor test # Test memory engine with progress bar
npx vektor status # System health check
npx vektor mcp # Start Claude Desktop MCP server
npx vektor rem # Run REM dream cycle
npx vektor chat # Persistent memory chat (all LLMs)
npx vektor remember # Store a fact
npx vektor ask # Query memory + LLM answer
npx vektor agent # Autonomous goal executor
npx vektor help # All commands
Install the .dxt extension for zero-config memory in every Claude Desktop session.
Install: drag vektor-slipstream.dxt onto the Claude Desktop Extensions page.
Once installed, Claude automatically:
All 44 tools are available in Claude Desktop — no configuration needed beyond your licence key.
User config fields:
| Field | Purpose |
|---|---|
licence_key | Your Polar licence key (required) |
db_path | Memory DB path (defaults to ~/vektor-slipstream-memory.db) |
project_path | Default path for cloak_cortex project scanning (optional) |
Download the latest .dxt from vektormemory.com/docs/dxt.
| Tool | Function |
|---|---|
vektor_recall | Semantic + BM25 + graph search across MAGMA memory |
vektor_recall_rrf | BM25+RRF dual-channel recall with cross-encoder rerank |
vektor_store | Store memory with importance score |
vektor_ingest | Batch ingest conversation turns with session date |
vektor_graph | Traverse associative memory graph |
vektor_delta | See what changed on a topic over time |
vektor_briefing | Generate morning briefing from recent memories |
vektor_stats | Memory DB stats — node count, edges, entities |
vektor_graph_stats | MAGMA graph node/edge/entity counts |
vektor_timeline | Query memories by date range |
| Tool | Function |
|---|---|
cloak_fetch | Stealth headless browser fetch via Playwright |
cloak_fetch_smart | Checks llms.txt first, falls back to stealth browser |
cloak_render | Full CSS/DOM layout sensor |
cloak_diff | Semantic diff of URL since last fetch |
cloak_diff_text | Structural diff between two text blobs |
cloak_passport | AES-256-GCM credential vault (get/set/delete/list) |
cloak_ssh_exec | Execute commands on remote server via SSH |
cloak_ssh_upload | Upload file to remote server via SFTP |
tokens_saved | Token efficiency ROI calculator |
| Tool | Function |
|---|---|
cloak_identity_create | Create persistent browser fingerprint identity |
cloak_identity_use | Apply saved identity to a fetch call |
cloak_identity_list | List all saved identities with trust summary |
| Tool | Function |
|---|---|
cloak_inject_behaviour | Human mouse/scroll injection for reCAPTCHA/Cloudflare bypass |
cloak_behaviour_stats | List available patterns and categories |
cloak_load_pattern | Load custom recorded behaviour pattern |
cloak_pattern_stats | Self-improving pattern store tier breakdown |
cloak_pattern_list | List patterns with scores and tier |
cloak_pattern_prune | Remove stale/low-scoring patterns |
cloak_pattern_seed | Seed store with built-in patterns |
| Tool | Function |
|---|---|
cloak_detect_captcha | Detect CAPTCHA type and sitekey |
cloak_solve_captcha | Solve via vision AI (Claude/GPT-4o/2captcha) |
| Tool | Function |
|---|---|
turbo_quant_compress | PolarQuant vector compression (~75% smaller) |
turbo_quant_stats | Compression ratio and savings stats |
cloak_cortex | Scan project directory into MAGMA entity graph |
cloak_cortex_anatomy | Get cached file anatomy without rescanning |
| Tool | Function |
|---|---|
vektor_text | Text generation across providers (OpenAI/Claude/Groq/Gemini/NVIDIA NIM) |
vektor_image | Image generation (DALL-E, Stability, NVIDIA) |
vektor_vision | Image understanding and analysis |
vektor_speech | Text-to-speech and transcription |
vektor_search | Web search with memory integration |
vektor_providers | List available multimodal providers and status |
| Tool | Function |
|---|---|
vektor_agent_run | Run autonomous goal executor with memory |
vektor_swarm | Launch multi-agent swarm task |
vektor_watch | File system watcher — auto-ingest on change |
Add to .claude/settings.json in your project:
{
"mcpServers": {
"vektor": {
"command": "node",
"args": ["/path/to/node_modules/vektor-slipstream/index.js"],
"env": {
"VEKTOR_LICENCE_KEY": "your-licence-key",
"CLOAK_PROJECT_PATH": "/path/to/your/project"
}
}
}
}
All 44 tools are available in Claude Code via this config.
vektor-magma-bridge.js)vektor-embedder.js)chat, remember, ask, agent commands| Metric | Value |
|---|---|
| Recall latency | sub-1ms (local SQLite + ONNX) |
| Embedding cost | $0 — fully local ONNX |
| Embedding latency | ~10ms GPU / ~25ms CPU |
| LoCoMo benchmark | 66.9% adjusted judge accuracy |
| vs Mem0 | beats Mem0 old algorithm (62.47%) |
| First run | ~2 min (downloads ~25MB model once) |
| Subsequent boots | <100ms |
Zero config. VEKTOR detects and uses the best available accelerator:
| Variable | Default | Purpose |
|---|---|---|
VEKTOR_SUMMARIZE | false | Enable LLM session summarization on ingest |
VEKTOR_TRIPLES | true | Enable batch triple extraction on ingest |
VEKTOR_FORESIGHT | true | Extract future-tense foresight signals |
VEKTOR_TEMPORAL | true | Enable temporal index and date boosting |
VEKTOR_CONTRADICT | true | Enable ADD-only contradiction detection |
VEKTOR_DEBUG | — | Enable verbose retrieval debug output |
VEKTOR_MODEL | Xenova/bge-small-en-v1.5 | Swap embedding model (e.g. bge-large for higher accuracy) |
VEKTOR_RERANK | true | Enable cross-encoder reranking |
Commercial licence granted. Monthly fee - all updates included
Solo $9/mo → 3 licences | Team $35/mo → 5 licences | Studio $59/mo → 10 licences | Enterprise $99/mo → 25 licences |
Purchase: vektormemory.com/product#pricing Docs: vektormemory.com/docs Support: hello@vektormemory.com
Built on peer-reviewed research:
Install via CLI
npx mdskills install Vektor-Memory/vektor-memoryVektor Memory is a free, open-source AI agent skill. Hardware-accelerated persistent memory for AI agents. Local-first. No cloud. One-time payment. 66.9% on LoCoMo benchmark (adjusted). Under 1ms retrieval. Zero cloud dependency. Retrieval pipeline rebuilt from scratch. - bge-small-en-v1.5 bi-encoder + ms-marco cross-encoder reranker (spec-decode architecture) - BM25 + Porter-stemmed BM25 + named entity injection, fused via RRF - MAGMA graph layer —
Install Vektor Memory with a single command:
npx mdskills install Vektor-Memory/vektor-memoryThis downloads the skill files into your project and your AI agent picks them up automatically.
Vektor Memory works with Claude Code, Claude Desktop, Cursor, Vscode Copilot, Windsurf, Continue Dev, Gemini Cli, Amp, Roo Code, Goose. Skills use the open SKILL.md format which is compatible with any AI coding agent that reads markdown instructions.