root 3eaac413e6 auditor: route kimi_architect through ollama_cloud/kimi-k2.6 (TOS-clean primary)
Two changes:

1. Default provider now ollama_cloud/kimi-k2.6 (env-overridable via
   LH_AUDITOR_KIMI_PROVIDER + LH_AUDITOR_KIMI_MODEL). Ollama Cloud Pro
   exposes kimi-k2.6 legitimately, so we no longer need the User-Agent-
   spoof path through api.kimi.com. Smoke test 2026-04-27:
     api.kimi.com    368s  8 findings   8/8 grounded
     ollama_cloud    54s   10 findings  10/10 grounded
   The kimi.rs adapter (provider=kimi) stays wired as a fallback when
   Ollama Cloud is upstream-broken.

2. Switch HTTP transport from Bun's native fetch to curl via Bun.spawn.
   Bun fetch has an undocumented ~300s ceiling that AbortController +
   setTimeout cannot override; curl honors -m for end-to-end max
   transfer time without a hard intrinsic limit. Required for Kimi's
   reasoning-heavy responses on big audit prompts.

3. Bug fix Kimi caught in this very file (turtles all the way down):
   Number(process.env.LH_AUDITOR_KIMI_MAX_TOKENS ?? 128_000) yields 0
   when env is set to empty string — `??` only catches null/undefined.
   Switched to Number(env) || 128_000 so empty/0/NaN all fall back.
   Same pattern probably exists in other files; future audit pass.

4. Bumped MAX_TOKENS default 12K -> 128K. Kimi K2.6's reasoning_content
   counts against this budget but isn't surfaced in OpenAI-shape content;
   12K silently produced finish_reason=length with empty content when
   reasoning consumed the budget.

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-27 06:14:16 -05:00
2026-04-22 02:41:15 -05:00
2026-03-27 05:54:09 -05:00
2026-04-22 02:41:15 -05:00
Description
Rust-first object storage system
6.3 GiB
Languages
TypeScript 38.4%
Rust 35.8%
HTML 13.9%
Python 7.8%
Shell 2.1%
Other 2%