- Published on
AI Agents: what to automate and what to keep human
People ask me if AI agents should run everything. My answer is no — and that’s usually the whole point.
Agents are strongest at execution loops. Humans are strongest at intent and judgment.
The division that scales
I split agent tasks into three buckets:
- Automate fully: repetitive, deterministic, low risk,
- Automate with review: medium impact, needs validation,
- Human-only: legal, trust, architecture tradeoffs, and safety-sensitive decisions.
This split keeps speed up and prevents silent bad behavior.
Example: reliable automations
Good agent loops:
- file format normalization,
- routine lint/build/test checks,
- periodic issue summaries,
- and post-incident notes from logs.
You want these to be boringly repeatable.
Example: partial automation
Mixed mode works for:
- patch generation,
- migration suggestions,
- content updates with style constraints,
- and deployment prep.
I let agents draft, then enforce a short human review before apply.
Example: keep humans in control
Do not automate without human sign-off when:
- data privacy boundaries are at stake,
- release risk is high,
- or a decision requires product-level judgment.
I have seen too many polished mistakes come from this gap.
Good prompt habits for agent reliability
- define scope,
- define success criteria,
- define hard constraints,
- define a rollback path.
That’s often enough to get from noisy suggestions to safe output.
Practical checklist before each run
- Is this task reversible?
- Is there a test to prove success?
- What is the blast radius?
- Which file paths are allowed?
If those aren’t clear, I pause automation and clarify first.
AI agents are not here to remove responsibility. They’re here to offload the repetitive load so humans can do the decisions that matter.