Stackme
The context layer for every AI.
Stackme is a free, open-source memory layer for AI. It stores what matters about you, retrieves relevant context before every query, and injects it into any AI β ChatGPT, Claude, Copilot, Gemini, Ollama, or any AI via API.
No server. No subscription. No data leaves your machine.
Install
pip install stackme
Or install from source:
pip install git+https://github.com/my-ai-stack/stackme
Quick Start
from stackme import Context
ctx = Context()
# Add facts about yourself
ctx.add_fact("I run a fintech B2B SaaS, launched March 2024")
ctx.add_fact("Q3 goal: 10K paying customers")
ctx.add_fact("Users are 25-40, income $50-100K")
# Add a user message β facts are auto-extracted
ctx.add_user_message("I'm building a B2B SaaS, targeting fintech companies")
# Ask any AI β Stackme retrieves your context first
context = ctx.get_relevant("What pricing should we use?")
# β "## Facts\n- I run a fintech B2B SaaS...\n- Q3 goal: 10K paying customers..."
# Your AI gets the full picture every time.
print(context)
How It Works
You: "What pricing should we use?"
Stackme retrieves:
- I run a fintech B2B SaaS
- Q3 goal: 10K paying customers
- Users are 25-40, income $50-100K
Enriched prompt β ChatGPT / Claude / any AI
Architecture
~/.stackme/
βββ memory.sqlite β all memories
βββ vectors.faiss β semantic index
βββ facts.graph β structured knowledge graph
- Session Memory β current conversation, in-process
- Short-Term Memory β last 24h, SQLite
- Long-Term Memory β permanent, SQLite + vector search
- Knowledge Graph β structured facts, extracted from your prompts
CLI
# Add facts
stackme add-fact "I run a fintech startup"
stackme add "I'm building a B2B SaaS" # also auto-extracts facts
# Retrieve context
stackme get "what pricing?"
# Search all memories
stackme search "fintech"
# List all facts
stackme facts
# Knowledge graph
stackme graph
# Session history
stackme history --last 10
# Export all data
stackme export
# Stats
stackme count
API Reference
Context
Main class. All data stored locally in ~/.stackme/.
ctx = Context(user_id="default") # multi-user support
Adding Memories
ctx.add_fact("I run a fintech startup") # structured fact
ctx.add_prompt("User asked about pricing") # user prompt
ctx.add_context("AI responded with...") # AI response
ctx.add_user_message("I'm building a B2B SaaS") # both + auto-extract
ctx.add_ai_message("Here's my recommendation...")
Retrieving
ctx.get_relevant("pricing strategy?", top_k=5) # β formatted string
ctx.search("fintech", top_k=10) # β list of strings
ctx.get_facts() # β all facts
ctx.get_graph(subject="User") # β GraphFact list
Session
ctx.get_session_history(last_n=20) # in-memory conversation
ctx.clear_session() # wipe session only
Utilities
ctx.export() # full JSON backup
ctx.count() # total items
ctx.clear_all() # β οΈ wipe everything
Privacy
Everything stays on your machine. Your memories are yours. We never see, store, or transmit your data. No account required.
License
Apache 2.0 β free for commercial and personal use.
Built by Stack AI Β· GitHub Β· HuggingFace
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support