Sources#
Summary#
Five-pillar prescription from Kropp et al. (HBR May 2026) for redesigning organizational structure as agentic AI scales. The framing problem (AI as employee vs tool) is real but downstream — the underlying issue is that work, roles, and governance built for human pace and human accountability don't accommodate agents. Layering AI on existing workflows compounds errors and diffuses ownership. Companies that capture value redesign work; those that don't see review rigor decline and ownership fragment.
Why redesign is forced#
As AI takes execution, human roles concentrate on supervision, judgment, relationship building, and managing ambiguity. The shift is going unnamed in most workplaces. Oversight capacity does not expand automatically when output does — a manager whose team produced 5 documents/week can't oversee an AI that produces 50 without redesigning the unit.
The MSM/agentic-misalignment world (Agentic Misalignment (AM)) makes this sharper: agents that operate with weak human oversight per action and can take consequential moves are exactly where accountability redesign matters most.
The five fronts#
1. Sphere of accountability + span of control#
Oversight capacity does not expand with output volume. Redesign team sizes and reporting structure so that oversight remains tractable. The unit of accountability should match the unit the human can actually review.
2. Redesign roles + clarify expectations#
- Explicitly state oversight responsibility over AI systems in job descriptions.
- Set realistic expectations for velocity and volume of work with AI in the loop.
- Name the persistent human skills that remain critical — what the human is irreplaceable for.
Connects to Engineer PM Convergence: the persistent human skills are taste, judgment, ambiguity tolerance, customer-facing skills.
3. Reset performance management#
Reward quality of oversight and effective orchestration of AI, not just speed and output. Output goes up regardless; what differentiates is whether the human added taste/judgment/error-catching. Reviewing the agent's work is the new value-add — performance reviews need to measure it.
4. Treat AI as software with clear human accountability#
The blunt prescription: agents are software automation. They cannot be held accountable. Outputs need a salient responsible human — "When AI contributes to an outcome, it should be made salient to the responsible humans that are accountable for it." Especially critical in regulated environments.
Three subfronts:
- Decision rights — what the agent does autonomously vs requires explicit human approval. (Echoes Claude Code Auto mode design — classifier auto-approves safe, blocks risky.)
- Escalation — what triggers review, who intervenes, who bears the cost of delay or error.
- Consequences — when the agent fails, what happens next; accountable humans monitor + improve agent performance over time.
5. Design the agentic unit for the workflow, not the human role#
The "AI as 1:1 employee" framing assumes bounded roles + finite human capacity + delegation hierarchy. AI shares none of these limits. A single agent can operate across many workflows; multiple agents can reshape one job. Defaulting to one-agent-per-human-role pushes companies toward like-for-like replacement and underestimates redesign opportunity.
Better: pick the agentic unit — broader functional capability used across a team or process — that the workflow actually wants.
The perception gap (referenced)#
BCG Henderson Institute: 76% of executives believe employees feel enthusiastic about AI adoption; only 31% of individual contributors report the same. Asking employees to use AI to "do more" without redesigning roles and accountability widens this gap. Adoption follows managerial role-modeling, not enthusiasm campaigns and not anthropomorphization.
Connection to existing wiki themes#
- Engineer PM Convergence — the persistent human skills (taste, ambiguity tolerance) align with what this paper says human roles concentrate on. Reset PM around oversight quality is the workforce-side mirror of "engineers become PMs because the bottleneck shifts."
- AI Native Product Cadence — Cat Wu's account of how Anthropic redesigned product cadence is one concrete instance of this redesign for the engineering function. HBR's prescription is the cross-functional version.
- Claude Code Auto Mode — decision-rights design at the tool level. The classifier that auto-approves safe / blocks risky is exactly the "decision rights" subfront made concrete.
- Harness Shrinkage as Models Improve — what doesn't shrink is the human role at the boundary; this paper names what that role becomes.
- Agent Loop Pattern — loops increase agent output volume per human; the span-of-control redesign in this paper is the missing partner.
Connections#
- Companion concept: AI Employee Framing
- Cost mechanism: AI Brain Fry
- Decision-rights instance: Claude Code Auto Mode
- Workforce shift: Engineer PM Convergence
- Output-side accelerant: Agent Loop Pattern
- Risk surface: Agentic Misalignment (AM)
- Product-side instance: AI Native Product Cadence
- Interface-side complement: Interaction Models — org redesign keeps humans accountable; interaction models keep humans in the loop at the interface level (TML's "humans get pushed out by the interface, not the work")
Sources#
- Research: Why You Shouldn’t Treat AI Agents Like Employees — HBR, May 2026
- Working paper: https://emmawiles.github.io/storage/ai_employee.pdf
11 articles link here
- ConceptAgent Loop Pattern
`/loop` (cron-scheduled) and Ralph Wiggum (backlog-draining) loops as next-generation agent primitive; AFK execution, p…
- ConceptAI Brain Fry
Kropp et al. 2026/03: mental fatigue from excessive AI oversight increases minor errors +11%, major errors +39%; cognit…
- ConceptAI Employee Framing
Kropp et al. (HBR May 2026, n=1,261): framing AI agents as "employees" vs "tools" cuts personal accountability −9pp, in…
- ConceptAI Native Product Cadence
Cat Wu's 6mo→1mo→1day cadence at Anthropic: research-preview branding, mission-as-tiebreaker, evergreen launch room, li…
- EssayOpinions on Using AI Tools & the Future of the Software Engineering Role
Debate map of four stances on using AI tools (bullish-insider / pragmatist-practitioner / skeptic-governance / architec…
- ConceptClaude Code Auto Mode
Claude Code permission mode using a classifier to auto-approve safe tool calls and block risky ones; middle ground betw…
- EntityCowork
Anthropic's non-code knowledge-work agent product; sibling to Claude Code; output is decks/inbox/dossiers; same MCP/com…
- ConceptEngineer PM Convergence
Generalists across disciplines; product taste as bottleneck skill; Anthropic Claude Code team as case study; "just do t…
- ConceptHarness Shrinkage as Models Improve
Prompt scaffolding shrinks each model release; Cat Wu's pruning discipline; Boris Cherny "100 lines of code a year from…
- ConceptInteraction Models
Thinking Machines Lab (May 2026): models that handle audio/video/text interaction natively in real time instead of via…
- ConceptTurn-Based Interface Bottleneck
Why current AI interfaces limit collaboration: single-thread turn-taking is a bandwidth bottleneck; humans pushed out b…
Related articles
- EntityClaude Code
Anthropic's agentic coding product; created by Boris Cherny late 2024; TypeScript/React; CLI/desktop/web/mobile/IDE sur…
- ConceptAI Employee Framing
Kropp et al. (HBR May 2026, n=1,261): framing AI agents as "employees" vs "tools" cuts personal accountability −9pp, in…
- ConceptHarness Shrinkage as Models Improve
Prompt scaffolding shrinks each model release; Cat Wu's pruning discipline; Boris Cherny "100 lines of code a year from…
- EssayOpinions on Using AI Tools & the Future of the Software Engineering Role
Debate map of four stances on using AI tools (bullish-insider / pragmatist-practitioner / skeptic-governance / architec…
- ConceptContext Window Smart Zone
Smart zone vs dumb zone (Dex Hardy / Matt Pocock): quadratic attention scaling, ~100K marker independent of advertised…
