Begin adding AI memory
This commit is contained in:
13
artifacts/plans/15-simplify-task-settings-and-more.md
Normal file
13
artifacts/plans/15-simplify-task-settings-and-more.md
Normal file
@@ -0,0 +1,13 @@
|
||||
|
||||
No Tasks Yet
|
||||
|
||||
This group has no derived tasks yet. To start populating this view:
|
||||
|
||||
Open Task Settings and confirm this chat is mapped under Group Mapping.
|
||||
Send task-like messages in this group, for example: task: ship v1, todo: write tests, please review PR.
|
||||
Mark completion explicitly with a phrase + reference, for example: done #12, completed #12, fixed #12.
|
||||
Refresh this page; new derived tasks and events should appear automatically.
|
||||
|
||||
|
||||
|
||||
task settings sound complicated, make them simpler
|
||||
34
artifacts/plans/16-agent-knowledge-memory-foundation.md
Normal file
34
artifacts/plans/16-agent-knowledge-memory-foundation.md
Normal file
@@ -0,0 +1,34 @@
|
||||
# Feature Plan: Agent Knowledge Memory Foundation (Pre-11/12)
|
||||
|
||||
## Goal
|
||||
Establish a scalable, queryable memory substrate so wiki and MCP features can rely on fast retrieval instead of markdown-file scans.
|
||||
|
||||
## Why This Comes Before 11/12
|
||||
- Plan 11 (personal memory) needs performant retrieval and indexing guarantees.
|
||||
- Plan 12 (MCP wiki/tools) needs a stable backend abstraction independent of UI and tool transport.
|
||||
|
||||
## Scope
|
||||
- Pluggable memory search backend interface.
|
||||
- Default Django backend for zero-infra operation.
|
||||
- Optional Manticore backend for scalable full-text/vector-ready indexing.
|
||||
- Reindex + query operational commands.
|
||||
- System diagnostics endpoints for backend status and query inspection.
|
||||
|
||||
## Implementation Slice
|
||||
1. Add `core/memory/search_backend.py` abstraction and backends.
|
||||
2. Add `memory_search_reindex` and `memory_search_query` management commands.
|
||||
3. Add system APIs:
|
||||
- backend status
|
||||
- memory query
|
||||
4. Add lightweight Podman utility script for Manticore runtime.
|
||||
5. Add tests for diagnostics and query behavior.
|
||||
|
||||
## Acceptance Criteria
|
||||
- Memory retrieval works with `MEMORY_SEARCH_BACKEND=django` out of the box.
|
||||
- Switching to `MEMORY_SEARCH_BACKEND=manticore` requires only env/config + container startup.
|
||||
- Operators can verify backend health and query output from system settings.
|
||||
|
||||
## Out of Scope
|
||||
- Full wiki article model/UI.
|
||||
- Full MCP server process/tooling.
|
||||
- Embedding generation pipeline (next slice after backend foundation).
|
||||
25
artifacts/plans/16-memory-backend-evaluation.md
Normal file
25
artifacts/plans/16-memory-backend-evaluation.md
Normal file
@@ -0,0 +1,25 @@
|
||||
# Memory Backend Evaluation: Manticore vs Alternatives
|
||||
|
||||
## Decision Summary
|
||||
- **Recommended now:** Manticore for indexed text retrieval and future vector layering.
|
||||
- **Default fallback:** Django/ORM backend for zero-infra environments.
|
||||
- **Revisit later:** dedicated vector DB only if recall quality or ANN latency requires it.
|
||||
|
||||
## Why Manticore Fits This Stage
|
||||
- Already present in adjacent infra and codebase history.
|
||||
- Runs well as a small standalone container with low operational complexity.
|
||||
- Supports SQL-like querying and fast full-text retrieval for agent memory/wiki content.
|
||||
- Lets us keep one retrieval abstraction while deferring embedding complexity.
|
||||
|
||||
## Tradeoff Notes
|
||||
- Manticore-first gives immediate performance over markdown scans.
|
||||
- For advanced ANN/vector-only workloads, Qdrant/pgvector/Weaviate may outperform with less custom shaping.
|
||||
- A hybrid approach remains possible:
|
||||
- Manticore for lexical + metadata filtering,
|
||||
- optional vector store for semantic recall.
|
||||
|
||||
## Practical Rollout
|
||||
1. Start with `MEMORY_SEARCH_BACKEND=django` and verify API/command workflows.
|
||||
2. Start Manticore container and switch to `MEMORY_SEARCH_BACKEND=manticore`.
|
||||
3. Run reindex and validate query latency/quality on real agent workflows.
|
||||
4. Add embedding pipeline only after baseline lexical retrieval is stable.
|
||||
Reference in New Issue
Block a user