A deterministic state machine. Every box maps to a real agent on GitHub. Human gates highlighted. Memory feedback loop visible. No prompt-wrapper magic.
One linear pipeline with two parallel fan-outs (implementation + review) and two human gates. Memory feedback loop runs on the right. Total wall-time for a small feature: ~45 min. Total LLM cost: ~$0.80–$2.40 depending on archetype.
Stack detection runs in your shell (no LLM call). The architect drafts ARCH.md + ADR + cost estimate. You see a diff before anyone touches code.
~$0.15–0.40 · 2 minApprove scope, reject, or edit. Without your green light, no implementation starts. The gate writes to a local file — auditable, version-controlled.
~30 s · 1 clickPM decomposes into beads tasks with explicit dependencies. Independent tasks run in parallel worktrees. Each senior-dev follows TDD (RED → GREEN → IMPROVE).
~$0.60–1.80 · 10–25 min5 reviewers run concurrently: QA, security, performance, archetype-specific (PCI / HIPAA / GDPR / EU AI Act / OWASP LLM), and 12-angle code review.
~$0.75–2.00 · 5–8 minYou see the full review verdict — APPROVED / BLOCKED / FAIL with rationale per reviewer. Approve to deploy, or send back with comments. Same audit trail.
~30 s · 1 clickCI runs your existing pipeline. devops watches health checks, error rates, p99 latency. Auto-rollback if SLO burn rate spikes within 10 min.
~$0.05–0.15 · 3–5 minMonitors Grafana / Sentry / Cloudflare. P0 → immediate triage + postmortem. P1/P2 → beads task with severity tag. All decisions logged.
~$0.10–0.40 / incidentAt session end (or every 50 commits), extracts repeatable patterns. Promotes to ~/.great_cto/decisions.md after ≥3 occurrences. Injects into agent Step 0 next iteration.
Memory is the difference between each iteration starting from zero and each iteration starting from everything you've already learned. Stored locally, git-backed, never sent to a vendor.
.great_cto/lessons.md. Survives session restarts because it's plain markdown in git.~/.great_cto/decisions.md. Yours, not the agent's.Retrieval: at agent Step 0, the kernel grep-ranks lessons by pattern_hash similarity to the current task, picks top-N (default 3), and injects them into the agent's context. Latency: ~50ms. Misses are logged and analyzed weekly.
No black boxes. Click any node above to jump to its source. Want the full agent index?