OpenAI's Agents SDK is brilliant for execution — but it's stateless by default. Every agent run starts from zero. VEKTOR adds persistent, local-first memory in three lines of code. No Pinecone. No Neo4j. No monthly bill.
The OpenAI Agents SDK gives you execution primitives: tools, handoffs, guardrails. What it doesn't give you is memory. By default, every agent run is isolated. The agent doesn't know what it decided last time. It doesn't remember the user's preferences. It has no concept of project history. You either manage context manually — which scales poorly — or you pay for a proprietary cloud memory solution that puts your data off-premises.
VEKTOR is the third option: local-first, one-time-purchase, zero-cloud persistent memory that integrates in three lines. Your agent gets a permanent, growing brain. Your data stays on your server. Your context window stays clean.
import { createMemory } from 'vektor-memory'; const memory = await createMemory({ provider: 'openai' }); await memory.remember("User wants to deploy on Vercel.");
That's it for the baseline. But the real power comes from wiring VEKTOR into your agent's tool loop — so it remembers and recalls automatically, without any manual context management.
import { Agent, tool } from 'openai-agents'; import { createMemory } from 'vektor-memory'; const memory = await createMemory({ provider: 'openai' }); // Give the agent memory tools const rememberTool = tool({ name: 'remember', description: 'Save important information to long-term memory', parameters: { content: 'string', importance: 'number' }, execute: async ({ content, importance }) => { await memory.remember(content, { importance }); return 'Remembered.'; } }); const recallTool = tool({ name: 'recall', description: 'Retrieve relevant memories for the current task', parameters: { query: 'string' }, execute: async ({ query }) => { const memories = await memory.recall(query, { topK: 5 }); return memories.map(m => m.content).join('\n'); } }); const agent = new Agent({ name: 'persistent-agent', model: 'gpt-4o', tools: [rememberTool, recallTool], instructions: 'You have persistent memory. Always recall context before responding. Save important decisions.' });
Most memory solutions require you to call an embedding API for every write and recall. At scale, this is a hidden cost that compounds quickly — 10,000 memory operations per month can cost $50–200 in embedding API calls alone.
VEKTOR generates embeddings locally using Transformers.js — running the embedding model directly on your hardware via WebAssembly. First run downloads the model (~80MB). Every subsequent embedding is free, instant, and private.