Four shipped features and a PRD realignment, all measured end-to-end:
HNSW trial system (Phase 15 horizon item → complete)
- vectord: EmbeddingCache, harness (eval sets + brute-force ground truth),
TrialJournal, parameterized HnswConfig on build_index_with_config
- /vectors/hnsw/trial, /hnsw/trials/{idx}, /hnsw/trials/{idx}/best,
/hnsw/evals/{name}/autogen, /hnsw/cache/stats
- Measured on resumes_100k_v2 (100K × 768d): brute-force 44ms -> HNSW 873us
at 100% recall@10. ec=80 es=30 locked as HnswConfig::default()
- Lower ec values trade recall for build time: 20/30 = 0.96 recall in 8s,
80/30 = 1.00 recall in 230s
Catalog manifest repair
- catalogd: resync_from_parquet reads parquet footers to restore row_count
and columns on drifted manifests
- POST /catalog/datasets/{name}/resync + POST /catalog/resync-missing
- All 7 staffing tables recovered to PRD-matching 2,469,278 rows
Federation foundation (ADR-017)
- shared::secrets: SecretsProvider trait + FileSecretsProvider (reads
/etc/lakehouse/secrets.toml, enforces 0600 perms)
- storaged::registry::BucketRegistry — multi-bucket resolution with
rescue_bucket read fallback and reachability probing
- storaged::error_journal — bucket op failures visible in one HTTP call
- storaged::append_log — write-once batched append pattern (fixes the RMW
anti-pattern llms3.com calls out; errors and trial journals both use it)
- /storage/buckets, /storage/errors, /storage/bucket-health,
/storage/errors/{flush,compact}
- Bucket-aware I/O at /storage/buckets/{bucket}/objects/{*key} with
X-Lakehouse-Rescue-Used observability headers on fallback
Postgres streaming ingest
- ingestd::pg_stream: DSN parser, batched ORDER BY + LIMIT/OFFSET pagination
into ArrowWriter, lineage redacts password
- POST /ingest/db — verified against live knowledge_base.team_runs
(586 rows × 13 cols, 6 batches, 196ms end-to-end)
PRD realignment (2026-04-16)
- Dual use case: staffing analytics + local LLM knowledge substrate
- Removed "multi-tenancy (single-owner system)" from non-goals
- Added invariants 8-11: indexes hot-swappable, per-reader profiles,
trials-as-data, operational failures findable in one HTTP call
- New phases 16 (hot-swap generations), 17 (model profiles + dataset
bindings), 18 (Lance vs Parquet+sidecar evaluation)
- Known ceilings table documents the 5M vector wall and escape hatches
- ADR-017 (federation), ADR-018 (append-log pattern) added
- EXECUTION_PLAN.md sequences phases B-E with success gates and
decision rules
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
163 lines
8.2 KiB
Markdown
163 lines
8.2 KiB
Markdown
# Phase Tracker
|
|
|
|
## Phase 0: Bootstrap ✅
|
|
- [x] Cargo workspace with all crate stubs compiling
|
|
- [x] `shared` crate: error types, ObjectRef, DatasetId
|
|
- [x] `gateway` with Axum: GET /health → 200
|
|
- [x] tracing + tracing-subscriber wired in gateway
|
|
- [x] justfile with build, test, run recipes
|
|
- [x] docs committed to git
|
|
|
|
## Phase 1: Storage + Catalog ✅
|
|
- [x] storaged: object_store backend init (LocalFileSystem)
|
|
- [x] storaged: Axum endpoints (PUT/GET/DELETE/LIST)
|
|
- [x] shared/arrow_helpers.rs: RecordBatch ↔ Parquet + schema fingerprinting
|
|
- [x] catalogd/registry.rs: in-memory index + manifest persistence
|
|
- [x] catalogd service: POST/GET /datasets + by-name
|
|
- [x] gateway routes wired
|
|
|
|
## Phase 2: Query Engine ✅
|
|
- [x] queryd: SessionContext + object_store config
|
|
- [x] queryd: ListingTable from catalog ObjectRefs
|
|
- [x] queryd service: POST /query/sql → JSON
|
|
- [x] queryd → catalogd wiring
|
|
- [x] gateway routes /query
|
|
|
|
## Phase 3: AI Integration ✅
|
|
- [x] Python sidecar: FastAPI + Ollama (embed/generate/rerank)
|
|
- [x] Dockerfile for sidecar
|
|
- [x] aibridge/client.rs: HTTP client
|
|
- [x] aibridge service: Axum proxy endpoints
|
|
- [x] Model config via env vars
|
|
|
|
## Phase 4: Frontend ✅
|
|
- [x] Dioxus scaffold, WASM build
|
|
- [x] Ask tab: natural language → AI SQL → results
|
|
- [x] Explore tab: dataset browser + AI summary
|
|
- [x] SQL tab: raw DataFusion editor
|
|
- [x] System tab: health checks for all services
|
|
|
|
## Phase 5: Hardening ✅
|
|
- [x] Proto definitions (lakehouse.proto)
|
|
- [x] Internal gRPC: CatalogService on :3101
|
|
- [x] OpenTelemetry tracing: stdout exporter
|
|
- [x] Auth middleware: X-API-Key (toggleable)
|
|
- [x] Config-driven startup: lakehouse.toml
|
|
|
|
## Phase 6: Ingest Pipeline ✅
|
|
- [x] CSV ingest with auto schema detection
|
|
- [x] JSON ingest (array + newline-delimited, nested flattening)
|
|
- [x] PDF text extraction (lopdf)
|
|
- [x] Text/SMS file ingest
|
|
- [x] Content hash dedup (SHA-256)
|
|
- [x] POST /ingest/file multipart upload
|
|
- [x] 12 unit tests
|
|
|
|
## Phase 7: Vector Index + RAG ✅
|
|
- [x] chunker: configurable size + overlap, sentence-boundary aware
|
|
- [x] store: embeddings as Parquet (binary f32 vectors)
|
|
- [x] search: brute-force cosine similarity
|
|
- [x] rag: embed → search → retrieve → LLM answer with citations
|
|
- [x] POST /vectors/index, /search, /rag
|
|
- [x] Background job system with progress tracking
|
|
- [x] Dual-pipeline supervisor with checkpointing + retry
|
|
- [x] 100K embeddings: 177/sec on A4000, zero failures
|
|
- [x] 6 unit tests
|
|
|
|
## Phase 8: Hot Cache + Incremental Updates ✅
|
|
- [x] MemTable hot cache: LRU, configurable max (16GB)
|
|
- [x] POST /query/cache/pin, /cache/evict, GET /cache/stats
|
|
- [x] Delta store: append-only delta Parquet files
|
|
- [x] Merge-on-read: queries combine base + deltas
|
|
- [x] Compaction: POST /query/compact
|
|
- [x] Benchmarked: 9.8x speedup (1M rows: 942ms → 96ms)
|
|
|
|
## Phase 8.5: Agent Workspaces ✅
|
|
- [x] WorkspaceManager with daily/weekly/monthly/pinned tiers
|
|
- [x] Saved searches, shortlists, activity logs per workspace
|
|
- [x] Instant zero-copy handoff between agents
|
|
- [x] Persistence to object storage, rebuild on startup
|
|
|
|
## Phase 9: Event Journal ✅
|
|
- [x] journald crate: append-only mutation log
|
|
- [x] Event schema: entity, field, old/new value, actor, source, workspace
|
|
- [x] In-memory buffer with auto-flush to Parquet
|
|
- [x] GET /journal/history/{entity_id}, /recent, /stats
|
|
- [x] POST /journal/event, /update, /flush
|
|
|
|
## Phase 10: Rich Catalog v2 ✅
|
|
- [x] DatasetManifest: description, owner, sensitivity, columns, lineage, freshness, tags
|
|
- [x] PII auto-detection: email, phone, SSN, salary, address, medical
|
|
- [x] Column-level metadata with sensitivity flags
|
|
- [x] Lineage tracking: source_system → ingest_job → dataset
|
|
- [x] PATCH /catalog/datasets/by-name/{name}/metadata
|
|
- [x] Backward compatible (serde default)
|
|
|
|
## Phase 11: Embedding Versioning ✅
|
|
- [x] IndexRegistry: model_name, model_version, dimensions per index
|
|
- [x] Index metadata persisted as JSON, rebuilt on startup
|
|
- [x] GET /vectors/indexes — list all (filter by source/model)
|
|
- [x] GET /vectors/indexes/{name} — metadata
|
|
- [x] Background jobs auto-register metadata on completion
|
|
|
|
## Phase 12: Tool Registry ✅
|
|
- [x] 6 built-in staffing tools (search_candidates, get_candidate, revenue_by_client, recruiter_performance, cold_leads, open_jobs)
|
|
- [x] Parameter validation + SQL template substitution
|
|
- [x] Permission levels: read / write / admin
|
|
- [x] Full audit trail per invocation
|
|
- [x] GET /tools, GET /tools/{name}, POST /tools/{name}/call, GET /tools/audit
|
|
|
|
## Phase 13: Security & Access Control ✅
|
|
- [x] Role-based access: admin, recruiter, analyst, agent
|
|
- [x] Field-level sensitivity enforcement
|
|
- [x] Column masking determination per agent
|
|
- [x] Query audit logging
|
|
- [x] GET/POST /access/roles, GET /access/audit, POST /access/check
|
|
|
|
## Phase 14: Schema Evolution ✅
|
|
- [x] Schema diff detection (added, removed, type changed, renamed)
|
|
- [x] Fuzzy rename detection (shared word parts)
|
|
- [x] Auto-generated migration rules with confidence scores
|
|
- [x] AI migration prompt builder for complex cases
|
|
- [x] 5 unit tests
|
|
|
|
## Phase 15+: Horizon
|
|
- [x] HNSW vector index with iteration-friendly trial system (2026-04-16)
|
|
- `HnswStore.build_index_with_config` — parameterized ef_construction, ef_search, seed
|
|
- `EmbeddingCache` — pins 100K vectors in memory, shared across trials
|
|
- `harness::EvalSet` — named query sets with brute-force ground truth
|
|
- `TrialJournal` — append-only JSONL at `_hnsw_trials/{index}.jsonl`
|
|
- Endpoints: `/vectors/hnsw/trial`, `/hnsw/trials/{idx}`, `/hnsw/trials/{idx}/best?metric={recall|latency|pareto}`, `/hnsw/evals`, `/hnsw/evals/{name}/autogen`, `/hnsw/cache/stats`
|
|
- Measured on 100K resumes: **brute-force 44-54ms → HNSW 509us-1830us**, recall 0.92-1.00 depending on `ef_construction`. Sweet spot: ec=80 es=30 → p50=873us recall=1.00 — locked in as `HnswConfig::default()`
|
|
- [x] Catalog manifest repair — `POST /catalog/resync-missing` restores row_count and columns from parquet footers (2026-04-16). All 7 staffing tables recovered to PRD-matching 2.47M rows.
|
|
- [~] Federated multi-bucket query — **foundation complete 2026-04-16**, see ADR-017
|
|
- [x] `StorageConfig.buckets` + `rescue_bucket` + `profile_root` config shape
|
|
- [x] `SecretsProvider` trait + `FileSecretsProvider` (reads /etc/lakehouse/secrets.toml, checks 0600 perms)
|
|
- [x] `storaged::BucketRegistry` — multi-backend, rescue-aware, reachability probes
|
|
- [x] `storaged::error_journal::ErrorJournal` — append-only JSONL at `primary://_errors/bucket_errors.jsonl`
|
|
- [x] Endpoints: `GET /storage/buckets`, `GET /storage/errors`, `GET /storage/bucket-health`
|
|
- [x] Bucket-aware I/O: `PUT/GET /storage/buckets/{bucket}/objects/{*key}` with rescue fallback + `X-Lakehouse-Rescue-Used` observability headers
|
|
- [x] Backward compat: empty `[[storage.buckets]]` synthesizes a `primary` from legacy `root`
|
|
- [x] Three-bucket test (primary + rescue + testing) verified: normal reads, rescue fallback with headers, hard-fail missing, write to unknown bucket 503, error journal + health summary
|
|
- [ ] `X-Lakehouse-Bucket` header middleware on ingest/query/catalog endpoints
|
|
- [ ] Catalog migration: set `bucket = "primary"` on every legacy ObjectRef
|
|
- [ ] `queryd` registers every bucket with DataFusion for cross-bucket SQL
|
|
- [ ] Profile hot-load endpoints: `POST /profile/{user}/activate|deactivate`
|
|
- [ ] `vectord` bucket-scoped paths (trial journals, eval sets per-bucket)
|
|
- [x] Database connector ingest (Postgres first) — 2026-04-16
|
|
- `pg_stream::stream_table_to_parquet` — ORDER BY + LIMIT/OFFSET pagination, configurable batch_size
|
|
- `parse_dsn` — postgresql:// and postgres:// URL scheme, user/password/host/port/db
|
|
- `POST /ingest/db` endpoint: `{dsn, table, dataset_name?, batch_size?, order_by?, limit?}` → streams to Parquet, registers in catalog with PII detection + redacted-password lineage
|
|
- Existing `POST /ingest/postgres/import` (structured config) preserved alongside
|
|
- 4 DSN-parser unit tests + live end-to-end test against `knowledge_base.team_runs` (586 rows, 13 cols, 6 batches, 196ms)
|
|
- [ ] Database connector ingest (Postgres/MySQL)
|
|
- [ ] PDF OCR (Tesseract)
|
|
- [ ] Scheduled ingest (cron)
|
|
- [ ] Fine-tuned domain models
|
|
- [ ] Multi-node query distribution
|
|
|
|
---
|
|
|
|
**30 unit tests | 11 crates | 16 ADRs | 2.47M rows | 100K vectors | All built 2026-03-27**
|
|
**HNSW trial system: 2026-04-16**
|