# Architecture Decision Records ## ADR-001: Object storage as source of truth **Date:** 2026-03-27 **Decision:** All data lives in S3-compatible object storage. No traditional database. **Rationale:** Eliminates DB operational overhead, enables infinite scale at storage tier, forces clean separation of data and metadata. ## ADR-002: Catalog metadata persistence **Date:** 2026-03-27 **Decision:** catalogd persists manifests as Parquet files in object storage. In-memory index rebuilt on startup. **Rationale:** No external DB dependency. Storage is already the source of truth. Write-ahead pattern ensures consistency. ## ADR-003: Real models only (no mocks) **Date:** 2026-03-27 **Decision:** AI sidecar hits Ollama with real models from Phase 3 onward. No stub/mock endpoints. **Rationale:** Local Ollama instance available with nomic-embed-text, qwen2.5, mistral, gemma2, llama3.2. Mocks hide integration bugs. ## ADR-004: Python sidecar as Ollama adapter **Date:** 2026-03-27 **Decision:** Python FastAPI sidecar is a thin HTTP adapter over Ollama's API. No model loading in Python. **Rationale:** Ollama handles model lifecycle, GPU scheduling, caching. Sidecar stays stateless and lightweight — no torch/transformers deps. ## ADR-005: HTTP-first, gRPC later **Date:** 2026-03-27 **Decision:** All inter-service communication uses HTTP through Phase 4. gRPC migration in Phase 5. **Rationale:** HTTP is simpler to debug, test, and iterate on. gRPC adds protobuf compilation and streaming complexity before APIs stabilize. ## ADR-006: TOML config over environment variables **Date:** 2026-03-27 **Decision:** System configuration via lakehouse.toml file, with sane defaults for all values. **Rationale:** Config files are versionable, self-documenting, and support structured data. Env vars remain available as overrides. System must be restartable from repo + config alone. ## ADR-007: Dual HTTP+gRPC on gateway **Date:** 2026-03-27 **Decision:** Gateway serves HTTP on :3100 (external) and gRPC on :3101 (internal). Both run in the same process. **Rationale:** Single binary simplifies deployment. HTTP stays for browser/curl access. gRPC provides typed contracts for service-to-service calls. No premature microservice split. ## ADR-008: Embeddings stored as Parquet, not a proprietary vector DB **Date:** 2026-03-27 **Decision:** Vector embeddings stored as Parquet files (doc_id, chunk_text, vector columns). Vector index (HNSW) serialized as a sidecar file. **Rationale:** Keeps all data in one portable format. No vendor lock-in to Pinecone/Weaviate/Qdrant. Vectors are queryable via DataFusion like any other data. Trade-off: brute-force search is fine up to ~100K vectors; HNSW needed beyond that. ## ADR-009: Incremental updates via delta files, not Delta Lake **Date:** 2026-03-27 **Decision:** Updates append to delta Parquet files. Queries merge base + deltas at read time. Periodic compaction merges deltas into base. Single-writer model (no concurrent writers). **Rationale:** Full ACID over Parquet (Delta Lake/Iceberg) is a multi-year project. Our use case is single-writer (one ingest pipeline) with read-heavy workloads. Merge-on-read with compaction is sufficient and dramatically simpler. ## ADR-010: Schema detection defaults to string **Date:** 2026-03-27 **Decision:** Ingest pipeline infers column types from data. When ambiguous or mixed, defaults to String rather than failing. **Rationale:** Legacy data is messy. A column with "123", "N/A", and "" is a string, not an integer. Downstream queries can CAST as needed. Better to ingest everything than reject on type errors. ## ADR-011: This is not a CRM replacement **Date:** 2026-03-27 **Decision:** The lakehouse is the analytical layer BEHIND operational systems. It ingests exports, not live data. CRM/ATS stays for daily operations. **Rationale:** Operational systems need single-record CRUD, permissions, UI workflows. The lakehouse answers cross-cutting questions that no single operational system can. They complement, not compete.