Field Notes6 min readVersão em português
Where Humans Belong in Your AI Loop
Every AI feature has humans somewhere in the loop. Most teams put them in the wrong place. Four placement modes — and a four-question test for picking the right one.
"Should we have a human review the output?" is the wrong question. The right one is: which human, doing what, at which step?
Every AI feature in production has humans in the loop somewhere. Even the ones that look fully autonomous have a human on call when the model breaks, a human writing the eval set, a human auditing the logs weekly. The question is never whether humans are involved; it's where they sit, and most teams pick the wrong spot.
Four placement modes
1. Reviewer — human before action
Model proposes, human approves, then the action happens. Slow, expensive, but recoverable: a bad output never reaches the user.
Use when: the action is irreversible or expensive (sending a customer email, executing a refund, posting publicly). The model is new and you don't trust it yet. Volume is low enough that a person can keep up.
2. Editor — human starts from the model output
Model drafts, human edits before publishing or sending. The human is doing the work — the model is removing the blank page.
Use when: the task is creative or judgment-heavy (drafting a proposal, writing a customer reply, summarizing a meeting). Quality matters more than speed. The cost of a wrong-but-unedited output is high. This is often the sweet spot — biggest leverage with lowest risk.
3. Auditor — sampled review after action
Model acts, human checks a sample (say, 5%) after the fact. Mistakes happen and are not always recoverable, but you catch systematic issues before they compound.
Use when: the action is reversible or low-stakes (categorizing tickets, tagging content, routing to a queue). Volume is too high for per-item review. You have the discipline to actually run the audit on schedule.
4. Escalator — human only when the model says "I don't know"
Model handles what it's confident about, hands off to a human when it isn't. The hardest of the four to get right because it requires the model to be honestly calibrated about its uncertainty.
Use when: there's a long tail of edge cases and a well-defined "easy path." You can measure how often the model escalates and tune the threshold. You have a human team standing by — escalation to a queue nobody watches is worse than no escalation at all.
The four-question test
Before you pick a mode, answer these four about the action your AI takes:
- Reversibility. If the output is wrong, can you undo it cheaply? Sending an email is irreversible; a draft in a queue is.
- Volume. How many of these per day? 50 means a reviewer can handle it. 50,000 means they can't.
- Skill ceiling. Does a competent reviewer take 30 seconds or 30 minutes per item? This sets your throughput math.
- Visibility of failure. If the output is wrong, who notices? The user immediately? Nobody until next quarter? "Nobody" means you need pre-action review, not post-hoc audit.
Reversible + high volume + low skill + visible failure → Auditor works.
Irreversible + low volume + high skill + invisible failure → Reviewer or Editor.
Mixed traffic with a clear "easy path" → Escalator, only if you can measure confidence.
Three failure modes we keep seeing
- Rubber-stamping. The reviewer approves so consistently that they're not actually reviewing. Either the model is good enough to remove them, or the review needs to be redesigned (different reviewer, harder check, randomized samples).
- Alert fatigue. The escalator queue gets 200 items a day and nobody can keep up. Within two weeks the team is closing items without reading them. Either tighten the model's confidence threshold or accept a higher error rate downstream.
- Missing audit. "We'll spot-check it later" never becomes a real cadence. If the audit isn't on someone's weekly calendar with a named owner, it doesn't exist.
The trap
The most expensive design we see: a reviewer in front of every action, plus an auditor sampling afterward, plus an escalator queue — because nobody felt comfortable removing any of them. The result is human cost that exceeds the AI savings, and the AI was supposed to make this cheaper.
Pick one mode. Measure it. Add a second only when the first proves insufficient. Over-engineering the human loop is how teams ship slowly and still feel unsafe.
Close
Humans aren't a fallback for bad AI. They're a load-bearing part of every AI system that works. Putting them in the right place — not too late to catch failures, not so early they bottleneck the system — is most of the design work.