Package Exports
This package does not declare an exports field, so the exports above have been automatically detected and optimized by JSPM instead. If any package subpath is missing, it is recommended to post an issue to the original package (pi-memctx) to support the "exports" field. If that is not possible, create a JSPM override to customize the exports field for this package.
Readme
pi-memctx
Stop making Pi rediscover your repo every session.
Quickstart • Why • Learning • Benchmark • How it works • Packs • Commands • Development
Stop making your coding agent rediscover your repo.
pi-memctx gives Pi a local, durable, Markdown-native memory layer. It searches project memory before each prompt, injects only compact/relevant context, lets Pi inspect the repo normally when memory is not enough, and learns durable discoveries after turns.
No database server. No hosted memory vendor. No hidden black box. Just Markdown memory packs you can read, edit, grep, version, and sync.
Quickstart
1. Install
pi install npm:pi-memctxAlready installed? Update with:
pi update npm:pi-memctxOr install from GitHub:
pi install git:github.com/weauratech/pi-memctxYou can also verify the local installation:
npx pi-memctx doctor2. Start Pi with the extension
cd /path/to/your/workspace
pi -e pi-memctx3. Generate your first memory pack
Inside Pi:
/memctx-pack-generatepi-memctx scans the workspace and creates a structured Markdown memory pack with context, decisions, runbooks, and indexes.
4. Ask normally
How do I deploy this service to production?The user experience stays the same: prompt, wait, get the answer. The Memory Gateway works behind the scenes.
Why it feels like magic
A normal coding agent starts every session cold:
User: How do I deploy the gateway service to production?
Agent: I'll inspect the repo...
[tool] list files
[tool] read workflows
[tool] inspect Helm charts
[tool] search docs
...pi-memctx turns that into:
User: How do I deploy the gateway service to production?
Agent: Merge to main, GitHub Actions builds Docker and pushes to ECR,
Helm values are updated, ArgoCD syncs to Kubernetes, staging is
automatic, and production is manual via ArgoCD approval/sync.The difference is not a better prompt. It is memory arriving before the model starts reasoning.
Automatic learning
After a rich planning, debugging, or repository-discovery turn, pi-memctx can persist multiple linked Markdown notes instead of one shallow summary:
memctx: learned 5 memories:
- context: [[packs/my-pack/20-context/payment-api|Payment API]] (updated)
- observation: [[packs/my-pack/60-observations/deploy-patterns|Deploy patterns]] (created)
- runbook: [[packs/my-pack/70-runbooks/deploy-payment-api|Deploy Payment API]] (created)
- action: [[packs/my-pack/40-actions/2026-05-02-prepared-rollout|Prepared rollout]] (created)
- session: [[packs/my-pack/80-sessions/rich-persistence-payment-api|Rich planning snapshot]] (created)Learned notes are cross-linked with [[wikilinks]], so future searches can recover the whole discovery: context, observations, runbooks, actions, decisions, and rich session snapshots.
Local and inspectable by design
- Memories are Markdown files on your machine.
- No hosted memory service or external vector database is required.
- You can inspect, edit, delete, commit, or sync packs yourself.
- Secret-looking values are blocked/redacted before persistence.
- When memory is insufficient or stale, Pi falls back to normal repo inspection.
Benchmark
Latest local benchmark from the synthetic NovaPay fixture, 5 tasks, 1 repeat:
QMD_PATH=/tmp/pi-memctx-qmd/node_modules/.bin/qmd \
BENCH_REPEATS=1 \
BENCH_PROFILES="baseline gateway" \
BENCH_TIMEOUT=120 \
bash benchmark/run.sh /tmp/pi-memctx-benchmark-gateway-final| Profile | Avg latency | Provider tokens/task | Visible tokens/task | Tool calls/task | Failed tools/task | Quality |
|---|---|---|---|---|---|---|
| baseline | 24.2s | 2,315 | 594 | 5.4 | 0.2 | 12/22 |
| gateway | 5.18s | 2,016 | 238 | 0.0 | 0.0 | 21/22 |
Compared with baseline:
| Metric | Gateway vs baseline |
|---|---|
| Latency | 78.6% faster |
| Visible tokens | 59.9% fewer |
| Tool calls | 100% fewer |
| Quality | +9 facts |
Benchmarks are intentionally local and reproducible. Run them on your own projects:
bash benchmark/setup.sh /tmp/pi-memctx-benchmark
QMD_PATH=$(pwd)/node_modules/.bin/qmd \
BENCH_PROFILES="baseline gateway" \
bash benchmark/run.sh /tmp/pi-memctx-benchmarkHow it works
pi-memctx is a Memory Gateway between the user and the main LLM.
User prompt
│
▼
Memory Gateway
├─ detects the active pack
├─ retrieves relevant Markdown memories with qmd or grep fallback
├─ ranks candidates with cheap semantic coverage
├─ builds a compact local memory summary
└─ injects only useful context
│
▼
Pi agent answers normally
├─ uses memory when sufficient
├─ inspects the repo when memory is partial or stale
└─ after the turn, pi-memctx learns durable context back to Markdown
├─ context
├─ observations
├─ runbooks
├─ decisions
├─ actions
└─ rich session snapshotsThe gateway does not replace the main LLM. It does the boring part first: finding the right project memory, compressing it, and preventing redundant tool exploration when the answer is already known.
Profile
pi-memctx now has one recommended runtime profile: gateway.
| Profile | Best for | Behavior |
|---|---|---|
gateway |
Fast daily use | Conservative local judge, compact context, qmd/grep retrieval, zero redundant memory searches when memory is sufficient. |
Inspect or re-apply the profile inside Pi:
/memctx-profile status
/memctx-profile gatewayOld profile names such as gateway-lite, gateway-full, qmd-economy, low, balanced, auto, and full are compatibility-mapped to gateway.
Memory packs
A pack is a directory of Markdown files with frontmatter. You can edit it with any editor, commit it, review it, or open it in Obsidian.
packs/my-project/
00-system/
pi-agent/
memory-manifest.md
resource-map.md
indexes/
context-index.md
decision-index.md
runbook-index.md
20-context/
api.md
web.md
infra.md
40-actions/
2026-05-02-prepared-rollout.md
50-decisions/
001-hexagonal-architecture.md
002-use-pgx.md
60-observations/
deploy-patterns.md
70-runbooks/
deploy.md
terraform.md
80-sessions/
rich-persistence-payment-api.mdRecommended note types:
| Type | Directory | Use it for |
|---|---|---|
context |
20-context/ |
Stack, services, repositories, conventions, environments. |
decision |
50-decisions/ |
Architecture and technical decisions with rationale. |
observation |
60-observations/ |
Durable facts, requirements, caveats, structural discoveries. |
runbook |
70-runbooks/ |
Repeatable operational procedures. |
action |
40-actions/ |
Completed work, migrations, deploys, incident notes. |
session |
80-sessions/ |
Sanitized rich planning/discovery snapshots for future retrieval. |
Commands
Most users only need the daily commands:
| Command | Purpose |
|---|---|
/memctx-pack-generate |
Create a memory pack from the current workspace. If a model is selected, deep LLM enrichment starts in the background by default; use --no-deep to skip it. |
/memctx-pack |
Select or show the active pack. |
/memctx-profile gateway |
Re-apply the recommended profile. |
/memctx-doctor |
Diagnose qmd, packs, and configuration. |
Advanced commands
| Command | Purpose |
|---|---|
/memctx-pack-status |
Show active pack and retrieval status. |
/memctx-config |
Show current config. |
/memctx-retrieval |
Configure retrieval policy. |
/memctx-autosave |
Configure automatic learning behavior. The default gateway profile uses conservative auto. |
/memctx-save-queue |
Review queued lower-confidence memory candidates. |
/memctx-pack-enrich |
Enrich a pack with deterministic repository inventory and optional LLM synthesis. Runs in the background. |
Deprecated aliases such as /pack and /pack-generate are still registered for compatibility.
Tools
memctx_search
Search the active memory pack:
Use memctx_search to find the deploy runbook.Parameters:
| Parameter | Values | Default |
|---|---|---|
query |
string | required |
mode |
keyword, semantic, deep |
keyword |
limit |
number | 5 |
When the Memory Gateway already injected sufficient memory, the agent is instructed not to call memctx_search again. That keeps answers fast and prevents duplicate context retrieval.
memctx_save
Save durable knowledge into the active pack:
Remember that production deploys require ArgoCD manual approval.The tool supports:
observationdecisionactionrunbookcontextsession
Secret-looking content is blocked.
Status overlay
pi-memctx keeps a small status overlay in Pi:
🧠 my-pack · memory ready · 3 memory hits · search:qmd · profile:gateway · learn autoThis tells you which pack is active, whether memory was useful, which search backend is being used, and whether automatic learning is enabled.
Best use cases
pi-memctx is especially useful when:
- you work on large repos or monorepos;
- Pi keeps rediscovering architecture and deploy flows;
- your team has many services with repeated conventions;
- you want local memory without a hosted vector database;
- you want agent memory you can read, edit, grep, and version;
- you want planning/discovery sessions to become durable project notes.
Why not just RAG?
Most RAG setups retrieve documents at query time. pi-memctx is different:
- it is local-first and file-based;
- it stores durable memories as Markdown;
- it learns after turns, not only before prompts;
- it links related memories together;
- it can fall back to repo inspection when memory is stale;
- it does not require a hosted vector database.
qmd integration
pi-memctx uses @tobilu/qmd when available for fast memory retrieval. To keep pi install npm:pi-memctx clean and warning-free, qmd is not installed automatically. If qmd is not available, pi-memctx falls back to grep-based search.
Resolution order:
QMD_PATHorMEMCTX_QMD_BIN- an already-installed local
.binor bundled/vendor path qmdonPATH- grep fallback
Optional qmd install example:
npm install -g @tobilu/qmdCheck your setup:
npx pi-memctx doctorSafety model
pi-memctx is local-first and intentionally boring about sensitive data.
It will not save:
- API keys
- tokens
- passwords
- private keys
- customer data
- payment card data
- sensitive payloads
Memory is Markdown on disk. You can inspect every byte.
Configuration
Most users should start with the defaults. Advanced users can configure behavior with environment variables or /memctx-profile.
Common environment variables:
| Variable | Purpose |
|---|---|
MEMCTX_CONTEXT_TOKEN_BUDGET |
Approximate injected context budget. |
MEMCTX_CONTEXT_MAX_ITEMS |
Maximum memory items to include. |
MEMCTX_RETRIEVAL |
Retrieval policy: fast, balanced, deep, strict, auto. |
MEMCTX_GATEWAY_JUDGE |
Gateway judge mode: off, conservative, auto, main-llm. |
MEMCTX_AUTOSAVE |
Autosave mode: off, suggest, confirm, auto. |
QMD_PATH / MEMCTX_QMD_BIN |
Explicit qmd binary path. |
Development
Requirements:
- Node.js 20+
- Bun for tests
- Pi coding agent
Install dependencies:
npm installRun checks:
npm run typecheck
npm test
npm run test:e2eRun all CI checks:
npm run ciRun benchmark:
bash benchmark/setup.sh /tmp/pi-memctx-benchmark
QMD_PATH=/tmp/pi-memctx-qmd/node_modules/.bin/qmd \
BENCH_REPEATS=1 \
BENCH_PROFILES="baseline gateway" \
BENCH_TIMEOUT=120 \
bash benchmark/run.sh /tmp/pi-memctx-benchmarkContributing
Issues and pull requests are welcome.
Good contributions include:
- better language-agnostic retrieval heuristics
- safer memory extraction
- clearer benchmark fixtures
- docs and examples
- profile tuning with reproducible numbers
Please do not commit private company memory packs, customer data, secrets, or benchmark fixtures derived from private repositories.
License
MIT