scrum: cloud-default models — swap mistral:latest for ollama_cloud::gpt-oss:120b #1
Loading…
x
Reference in New Issue
Block a user
No description provided.
Delete Branch "scrum/cloud-default-models"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Summary
Surfaced by the lakehouse scrum-master pipeline (run 2026-04-24) pointed at this repos source. The scrum found three hardcoded
mistral:latestdefaults in the meta-pipeline orchestrator paths. Perfeedback_no_mistral.md(lakehouse project memory), mistral 7B has decoder-level JSON malformation (0/5 fill on structured-output A/B) and is unreliable in paths that consume structured output.Change: swap to
ollama_cloud::gpt-oss:120b(Phase 20 T3 cloud tier, proven workhorse).Patch locations (3)
llm_team_ui.py:9959—model_setsdefault for meta-pipeline stagesllm_team_ui.py:10084— fallback when Ollama/api/tagsprobe throwsllm_team_ui.py:11835— defaultworkerslist for orchestrator modeAll three are DEFAULTS — callers passing explicit
config.model_sets/config.modelsare unaffected.⚠ This PR carries 4 additional pre-existing local commits
Base
main(local) was 4 commits ahead oforigin/mainbefore this branch was cut — those commits were already committed locally, just unpushed. They are included here as prerequisites (specificallyfa6ccff Ollama Cloud provider + model browser + OpenRouter key fixprovides theollama_cloud::routing prefix this PR relies on):fa6ccffOllama Cloud provider + model browser + OpenRouter key fix98bda6eOpenRouter: show all 343 models (free + paid) with pricing and filter34ee12eFix adaptive mode: model list + synthesizer dropdown were never populated205eff6Deep Analysis mode + token tracking for all runs (including public)12ab391scrum: swap mistral:latest defaults to ollama_cloud::gpt-oss:120b ← this PRReviewer options: (a) merge as-is, (b) reject and first push
mainto origin, then rebase this onto updated origin/main (will leave only commit12ab391), or (c) cherry-pick just12ab391onto a smaller branch that explicitly pulls infa6ccffonly.Runtime activation
Ollama Cloud requires
OLLAMA_CLOUD_API_KEYenv var or a key saved via the Admin UI (providers.ollama_cloud.api_key). This PR does not change credential behavior, only the default model list. Without a key configured, the orchestrator will fail on the cloud default and the user should set explicitconfig.model_setswith local models.Not in this PR (other scrum findings, not acted on)
Scrum also surfaced: (F2) no 3-tier access middleware, (F3) no sentinel loop, (F4) no cloud-determinism consensus, (F5)
saveProviderno bearer validation. After grep verification: F2, F3 turned out to be false positives from tree-split distillation signal loss (rate_limited + is_allowlisted + sentinel loop are all implemented; reviewer couldn’t see them in the 209-shard scratchpad). F4 and F5 are partial / inconclusive and would need a forensic-preamble rerun before being actionable.Three default model lists hardcoded mistral:latest as the fallback when config.get("model_sets" / "models") returns nothing. Per feedback_no_mistral.md, mistral 7B has decoder-level JSON malformation issues (0/5 fill rate on A/B) and is a liability in any path that depends on structured output from the model. Swapping to ollama_cloud::gpt-oss:120b (Phase 20 T3 cloud tier) keeps the defaults reliable for the meta-pipeline orchestrator (line 9959), the fallback model list for empty Ollama (10084), and the worker pool default (11835). All three are DEFAULTS — any caller passing explicit config.model_sets / config.models is unaffected. Routing works because query_model's "::" provider prefix already resolves ollama_cloud via commit fa6ccff. Activation requires OLLAMA_CLOUD_API_KEY or a key saved via the Admin UI; this PR does not change credential behavior, only the default model list. Surfaced by lakehouse scrum-master pipeline run 2026-04-24, findings confirmed by grep verification against the live code.Checkout
From your project repository, check out a new branch and test the changes.