Stackme

The context layer for every AI.

Stackme is a free, open-source memory layer for AI. It stores what matters about you, retrieves relevant context before every query, and injects it into any AI β€” ChatGPT, Claude, Copilot, Gemini, Ollama, or any AI via API.

No server. No subscription. No data leaves your machine.


Install

pip install stackme

Or install from source:

pip install git+https://github.com/my-ai-stack/stackme

Quick Start

from stackme import Context

ctx = Context()

# Add facts about yourself
ctx.add_fact("I run a fintech B2B SaaS, launched March 2024")
ctx.add_fact("Q3 goal: 10K paying customers")
ctx.add_fact("Users are 25-40, income $50-100K")

# Add a user message β€” facts are auto-extracted
ctx.add_user_message("I'm building a B2B SaaS, targeting fintech companies")

# Ask any AI β€” Stackme retrieves your context first
context = ctx.get_relevant("What pricing should we use?")
# β†’ "## Facts\n- I run a fintech B2B SaaS...\n- Q3 goal: 10K paying customers..."

# Your AI gets the full picture every time.
print(context)

How It Works

You: "What pricing should we use?"

Stackme retrieves:
  - I run a fintech B2B SaaS
  - Q3 goal: 10K paying customers
  - Users are 25-40, income $50-100K

Enriched prompt β†’ ChatGPT / Claude / any AI

Architecture

~/.stackme/
β”œβ”€β”€ memory.sqlite     ← all memories
β”œβ”€β”€ vectors.faiss     ← semantic index
└── facts.graph       ← structured knowledge graph
  • Session Memory β€” current conversation, in-process
  • Short-Term Memory β€” last 24h, SQLite
  • Long-Term Memory β€” permanent, SQLite + vector search
  • Knowledge Graph β€” structured facts, extracted from your prompts

CLI

# Add facts
stackme add-fact "I run a fintech startup"
stackme add "I'm building a B2B SaaS"   # also auto-extracts facts

# Retrieve context
stackme get "what pricing?"

# Search all memories
stackme search "fintech"

# List all facts
stackme facts

# Knowledge graph
stackme graph

# Session history
stackme history --last 10

# Export all data
stackme export

# Stats
stackme count

API Reference

Context

Main class. All data stored locally in ~/.stackme/.

ctx = Context(user_id="default")  # multi-user support

Adding Memories

ctx.add_fact("I run a fintech startup")       # structured fact
ctx.add_prompt("User asked about pricing")    # user prompt
ctx.add_context("AI responded with...")      # AI response
ctx.add_user_message("I'm building a B2B SaaS")  # both + auto-extract
ctx.add_ai_message("Here's my recommendation...")

Retrieving

ctx.get_relevant("pricing strategy?", top_k=5)  # β†’ formatted string
ctx.search("fintech", top_k=10)                  # β†’ list of strings
ctx.get_facts()                                   # β†’ all facts
ctx.get_graph(subject="User")                     # β†’ GraphFact list

Session

ctx.get_session_history(last_n=20)  # in-memory conversation
ctx.clear_session()                # wipe session only

Utilities

ctx.export()      # full JSON backup
ctx.count()       # total items
ctx.clear_all()   # ⚠️ wipe everything

Privacy

Everything stays on your machine. Your memories are yours. We never see, store, or transmit your data. No account required.


License

Apache 2.0 β€” free for commercial and personal use.


Built by Stack AI Β· GitHub Β· HuggingFace

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Space using my-ai-stack/stackme 1