Engine: - Chains modes in sequence: extract → research → validate → debate → synthesize - Each stage feeds its output to the next as input - Runs same pipeline with different model sets (one model per iteration) - Auto-scores final output using judge model (1-10) - Keeps best result across all iterations - All stage results + final outputs saved to meta_runs table 4 preset pipelines: 1. Security Deep Dive — security logs through 5-stage analysis 2. Run History Insights — team run data through 4-stage extraction 3. Threat Intel Enrichment — profiled IPs through 5-stage analysis 4. Cross-Report Synthesis — past self-reports through 4-stage debate Database: - meta_pipelines: name, source, stages, status, best_score, iterations - meta_runs: per-iteration stage results, final output, score, models API: - POST /api/meta-pipeline — create pipeline from preset - POST /api/meta-pipeline/:id/start — run in background - POST /api/meta-pipeline/:id/stop — halt execution - GET /api/meta-pipelines — list all with live status - GET /api/meta-pipeline/:id — full detail with all iteration results UI (Lab page): - Magenta-bordered Meta-Pipeline card with 4 clickable presets - Click preset → creates + auto-starts pipeline - Pipeline list with live status dots, progress, scores - Click pipeline → drill-down with per-iteration results - Each stage expandable (click to show output) - Best output highlighted in green border - Auto-refreshes every 5 seconds during runs Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Description
LLM Team UI - Full-stack local AI orchestration platform
Languages
Python
97.4%
Shell
2.6%