The People
AI-Native Teams
Hierarchy exists because human information-processing capacity is bounded. When agents carry the information processing, the natural team size collapses to where deep trust and improvisation are possible — seven or eight people rather than fifty. The AI-native organization takes a different shape on a different substrate.
AI-Native Teams
Hierarchy exists because a single human can oversee roughly seven people before attention breaks. That span-of-control limit, combined with the cognitive ceiling of the people inside it, gives the thousand-person company its four-layer shape. Jay Galbraith named this directly in a 1974 Interfaces paper still cited by organization-design academics today: structure follows the information-processing required to reduce task uncertainty, and hierarchy is one of four coping strategies for organizations whose information-processing needs exceed what individual managers can carry. Fifty years later, the four strategies still work. What has changed is that agents now carry the information processing, and the natural team size collapses to where deep trust, shared context, and real-time improvisation are possible — the seven-or-eight-person team rather than the fifty-person department.
The AI-native organization takes a different shape on a different substrate. A flat structure built without the substrate underneath is the most common early failure pattern, and it collapses back to ad-hoc hierarchy inside a year. The new shape is small teams with distributed authority. The substrate is a company-wide intelligence layer that every human and every agent queries against. The two must arrive together, and the order matters — substrate first, team redesign second. Flattening before the intelligence layer exists removes the coordination mechanism the hierarchy provided and leaves nothing in its place; the firm rebuilds its hierarchy informally inside a year. Legibility without the team change is a different stall: the firm pays the substrate cost while middle management persists alongside the new intelligence layer, capturing none of the team-size benefit. Both patterns are common; the combination — substrate and redesign arriving together — is what compounds.
Every hierarchy layer compresses the signal above and below it
Every hierarchy layer compresses information. A customer complaint that starts as a five-minute call from an unhappy enterprise buyer gets summarized by the customer-success rep, re-summarized by the CS lead for the team's weekly meeting, re-framed by the manager in a status memo, reduced to a line item in the director's monthly rollup, and aggregated into a single metric in the VP's quarterly review. By the time the engineer who could fix the issue hears about it, the complaint has lost the specific pricing objection, the competitor name, and the internal champion. Each compression step is rational at its layer; the aggregate is structurally lossy. A small fraction of the original signal reaches the executing layer through a four-layer company, and the fraction collapses further as volume grows. Reorg-and-rehire cycles fight the symptoms. All-hands meetings do not fix the root cause, because the compression happens in-between the layers, not at the top or bottom.
What changes with AI is not the number of layers but the substrate underneath them. An intelligence layer preserves the originating signal with timestamps, speaker attribution, and source pointers. The engineer who would have seen the complaint as a rolled-up line item can now query the exact call, hear the objection, see the competitor mentioned, and reach the champion directly. Jack Dorsey's own framing of this move, from the From Hierarchy to Intelligence discussion with Roelof Botha in March 2026, is the cleanest public statement of what legibility buys: every artifact a remote-first company generates — Slack messages, email threads, pull requests, Google documents, recorded meetings — carries information about how the company is actually working, failing, and deciding, and that information has historically moved through a chain of managers summarizing upward. Putting an intelligence on top of the artifacts lets anyone in the company query the company itself, without the intermediating layers.
Organizations that ingest the artifacts but leave the org chart unchanged keep paying the compression tax. The layers still decide what gets forwarded, and they still forward a small fraction. Removing layers without the substrate underneath is the other direction of the same error: every flat-org experiment without a machine-readable information layer collapses back into ad-hoc hierarchy within a year because no shared context exists for the team to coordinate against. What the substrate buys is the possibility of a team redesign that would otherwise destabilize the company.
Small cross-trained teams distribute authority without a conductor
Two canonical small-team analogies converge on the same underlying mechanism. A seven-piece jazz band has no conductor, yet every player knows where the solo is going and when to pass it. A SEAL platoon in Naval Special Warfare doctrine consists of sixteen operators, subdivided for operations into two 8-man squads, four 4-man fire teams, or eight 2-man sniper pairs, with warfare-skill specialties distributed across the platoon — combat diving, demolitions, air operations, communications, and small-boat handling. Most land-warfare missions run at an 8-man squad. Every operator is cross-trained on multiple specialties, so any member can cover another role when the mission shape shifts. Authority passes in real time based on context, not pre-assigned by rank. The 2026 business discourse has picked up the analogy directly: Jamie Dimon's April 2026 annual letter to JPMorgan shareholders explicitly framed small, cross-trained "Navy SEAL"-like teams as his preferred structure for competitive battles.
Both analogies work because of three mechanisms the small team has that the large team does not. The first is shared context: every jazz player hears the same music; every SEAL operator has the mission brief and the current situational read. The second is real-time legibility: each player sees what the others are doing as they do it; each operator sees position, status, ammunition, line-of-sight in real time. The third is agreed protocols — chord changes and tempo for jazz, rules of engagement and hand signals for SEAL teams — that let any player hand off or take the lead without explicit coordination.
The AI-native team runs on the same three mechanisms at company scale. Shared context is the token-metabolism substrate — every Slack message, every commit, every customer call, every artifact ingested into the intelligence layer and made queryable by every agent and every human. Real-time legibility is the shared operational surface where work happens: Slack-native agent interfaces, shared workspaces, live dashboards. Agreed protocols are the skill files, the governance tiers, the handoff schemas that constrain how agents and humans interact. The team consults the substrate rather than the manager. The "manager" role as work-distributor disappears because work distribution becomes a protocol question — the substrate answers the question that a human previously did.
A seven-person pod can span the functional surface that a fifty-person department used to cover, because specialist depth lives in skills rather than headcount. The pod composition question stops being "do we have an engineer, a designer, a marketer, a finance analyst?" and becomes "are the four responsibilities (below) covered, with no one purely executing?" The answer determines whether the pod is functional or whether it is going to fail in a specific, diagnosable way.
Make the company legible, then compress what matters
Legibility is the substrate; on top of it sits the practical discipline of what format actually gets used to communicate. Dorsey's framing of the legibility goal is direct: imagine the company entirely legible, every artifact queryable against an intelligence layer that preserves the information that humans have historically compressed on the way up the hierarchy. Legibility alone does not buy the team redesign, but the redesign is impossible without it.
Once legibility is in place, a second discipline decides what gets compressed and how. The practical hierarchy of communication formats, ordered from worst to best by information-per-unit-time:
- A meeting. Four person-hours of attention spent on two people; output rarely captured; decisions often not recorded in any artifact a later reader can consult.
- A memo. One hour to write, often skimmed. Better than a meeting for the reader, equal in writer-cost.
- A presentation. Slower than a memo for the reader but more visual; the preparation itself forces specificity.
- A working artifact. A functioning prototype, a filled-in financial model, a tested feature with real user data. Shows exactly what is meant because the reader can interact with it.
- A single number. Maximum compression. "We tested the new onboarding flow on one percent of traffic and conversion went up thirty percent." Forces specificity; leaves no room for rhetorical cover; makes the decision obvious.
The operating discipline: default to artifacts and numbers over memos and meetings. When the product question is "should we ship the new onboarding flow?", a working prototype in Slack beats a slide deck beats a status meeting. The companies that become legible but keep compressing through meetings pay both the substrate cost and the broken-telephone cost — doubly expensive, worse than the pre-AI baseline.
The human role splits into four responsibilities
When agents execute, the human role is not one rung up the hierarchy from the agent. It is four distinct responsibilities that every role — engineer, marketer, customer-success lead, legal counsel, finance analyst — now holds simultaneously.
Architect and designer of the system. Every role writes specs. The marketer writes the brand-style file that constrains the content agent. The CS lead writes the escalation skill that decides when a ticket needs a human. The finance analyst writes the variance-analysis skill. Specifying what good output looks like is the unit of knowledge work no agent does. Engineers used to hold this responsibility alone; now everyone does. This is also where goal selection lives: agents optimize within a goal but do not originate one. Deciding which business lines to open, which problems to solve, which customers to serve is part of the Architect responsibility — frequently confused with "organizing the doing," which is now agentic.
Relationships, trust, and communication with other humans. The interpersonal layer agents do not provide: reading the room on a customer call, defending a budget in a resource conflict, mediating a disagreement, building the trust that makes a hard conversation possible. Interpersonal skill becomes more valuable in an AI-native company, not less, because the technical-execution layer that used to require headcount no longer does. A team that can move fast operationally but cannot hold a difficult conversation with a key customer — or with each other — will lose the advantage the substrate bought.
Validation. Reviewing what agents produce and what other humans produce against the defined spec. The quality gate. Validation is the bulk of many roles in the 2026 window: every finance team reviews AI-generated variance analyses, every content team reviews AI-drafted copy, every legal team reviews AI-extracted clause-risk flags. It is the most weight-bearing responsibility for the 2026–2027 window. Over time, as agents self-validate better and validation itself becomes automatable for narrow domains, it shrinks relative to the other three. The pattern across drafted chapters of this playbook: validation is where most of the human attention sits right now; the other three responsibilities are where the long-run leverage lives.
Accountability. Taking legal, fiduciary, and reputational responsibility for the work produced. The regulator, the customer, the board, the court all require a human in the decision chain. Accountability does not dissolve as agents improve. The agent-at-machine-speed asymmetry means accountability becomes harder, not easier, because the human on the hook must explain work done at machine speed without having personally reviewed each step. Accountability is the responsibility that cannot be automated, because there is no one else for a court or a regulator to hold responsible.
Someone who only executes becomes redundant, because execution is now an agent's job. Someone who cannot design specs or validate becomes a liability: either the agents run against weak specs and produce plausible wrong outputs, or the validation gate fails silently. Someone unable to own accountability creates governance gaps. The hiring signal that follows: look for people who already hold these four responsibilities in their current role, not people who execute narrow specialty work well.
Management becomes a text file
The dominant pattern for agent-mediated coordination inside an AI-native team is management-as-text-file. A DRI (directly responsible individual) maintains a three-hundred-to-five-hundred-line instruction file that encodes the team's priorities, operating principles, approval thresholds, and escalation rules. Agents read the file continuously and generate the daily priorities, the dashboards, the first-pass deliverables. Human specialists act on what the agents produce — specifying, validating, building relationships. Agents then validate the outputs against the file, escalating anomalies. Management ceases to be a layer and becomes a file, updated by the human owner and consulted by agents and other humans alike. This reframes the word "manager": the manager is now the spec-author for the team, not the distributor of tasks.
Browserbase is the most public 2026 variant. The entire company interacts with one Slack-native agent across every function — engineering, ops, sales, support, exec — describing intent at a high level and course-correcting in-thread. The instruction file is made visible in Slack, the agent picks up which skills to load per request via a routing table, and the work distributes across a team of three to five humans plus an agent fleet that would previously have required eight to twelve people and a manager above them. The pod's "manager" is the shared instruction file and the shared intelligence layer, not a person doing work distribution.
The instruction file is the living contract for the team; updating the file is how organizational learning gets compiled. When the file drifts from what the team actually does, the failure mode is worse than having no file at all — agents will act on the file anyway and produce consistent, confidently-wrong outputs. A single named human owner for the file plus a weekly review against validation data are non-negotiable. The practical starting point for a team trying to build its first instruction file: have the current manager write down what they decide when they are "in the loop" for a week. That becomes version zero. Iterate weekly against what the team actually did. The file converges fast.
Management shifts from imperative to declarative. Instead of step-by-step instructions on how to do the work, the manager specifies the goal and the expected outcomes; the agent layer figures out the how. The question a manager used to ask — "has anyone started on the Johnson renewal?" — becomes a query against the intelligence layer, answered in seconds by anyone on the team.
A three-to-five-person pod covers what a fifty-person department used to
A pod of three to five humans typically covers the functional surface a traditional technology company used to staff across fifteen to fifty specialists — design, engineering, growth, operations, analytics, customer success. None of these functions need the specialist depth they once did, because the depth lives in skills and agents. What the pod needs is the four responsibilities distributed across its members: a named architect-primary for each domain the pod covers, a named owner of the key external relationships, clear validation ownership, and unambiguous accountability for each outcome. The composition question changes shape entirely.
Shopify's own articulation of this, from Tobi Lütke's 2026 podcast with David Senra, is the cleanest public statement of the same pattern in a business context. Lütke names the five-person pod as Shopify's core operating unit, notes the team can temporarily stretch to eight, and explicitly cites the military convergence: "It's sort of what the military ends up figuring out too — they test these things and come to the same conclusions"Shopify's own articulation of this, from Tobi Lütke's 2026 podcast with David Senra, is the cleanest public statement of the same pattern in a business context. Lütke names the five-person pod as Shopify's core operating unit, notes the team can temporarily stretch to eight, and explicitly cites the military convergence: "It's sort of what the military ends up figuring out too — they test these things and come to the same conclusions"↗. Shopify has roughly three and a half thousand people in its R&D organization, which is really lots and lots of small teams rather than a single enormous one, and each gradation upward in team size costs, in Lütke's estimate, roughly a factor-of-ten loss in productivity. Work moves through phase-gated transitions — prototype, build, operate — with explicit meetings at the boundaries rather than continuous status flow.
The empirical research now supports the pattern at the individual level. A pre-registered field experiment at Procter & Gamble with seven hundred seventy-six professionals found that individuals working with AI matched the performance of two-person teams working without AI on real product-innovation tasks. The study, by Dell'Acqua and collaborators at Harvard Business School, the Wharton School, and Warwick Business School, also found that AI produced another effect the small-team analogies predict: it broke down functional silos between R&D professionals and Commercial professionals, who would otherwise have proposed solutions biased toward their own specialization. The AI-assisted individuals produced balanced solutions regardless of their background, which is the "cross-trained operator" property the SEAL analogy names directly.
The same pattern shows up outside technology in 2026. A regional logistics dispatch desk that previously ran with a dispatcher, two route planners, a customer-service lead, and a shift manager can operate with a three-person pod when the routing and customer-communication agents are built against the firm's dispatch history and service-level agreements. The pod composition question — does someone own the Architect responsibility for the routing playbook, does someone own the relationships with the three biggest customers, who validates the agent's dispatch decisions on the edge cases, who holds accountability when a route fails — is the same question whether the work is code, claims, dispatch, or content.
Hierarchy still holds in four categories of work
Flat-org does not generalize everywhere. Four categories of work keep tiered structure because the binding constraint is not information-processing:
- Regulated command hierarchies — military operations, airline flight operations, nuclear plant operations. Failure cost is catastrophic; the chain of command is the control mechanism, and it works at scale specifically because no one in the chain is authorized to act outside their rung.
- High-coordination physical operations — manufacturing lines, large construction projects, complex surgical teams. The coordination problem is physical, not informational; sequencing and physical safety require roles and rank.
- Crisis response — incident command at fire, flood, or public-health scale. Decision speed under ambiguity exceeds what a flat protocol can absorb; someone must hold the call, and the organization must know who that is in advance.
- Compliance-heavy finance workflows — where deployer-liable doctrine and specific regulatory sign-off thresholds force senior approval. The hierarchy here is a legal requirement, not an information-processing solution.
The chapter's claim is not that all hierarchy dissolves. The claim is narrower: information-processing-bound work moves to flatter, substrate-coordinated small teams because the specific binding constraint — the human span of control — has moved. In the categories above, the binding constraint is different, and the structure stays.
Block makes the sequencing concrete
The small AI-native team is the structural consequence of moving information processing onto agents, not a rebadging of the startup team. Galbraith's 1974 argument stands, with one variable moved: the coping strategy that used to require hierarchy now requires an intelligence layer instead. The organizations that pull ahead in 2026 are the ones where the substrate came first and the team redesign followed.
Block is the public case that makes the sequencing concrete (the workforce restructure itself is developed in 1.3)Block is the public case that makes the sequencing concrete (the workforce restructure itself is developed in 1.3)↗. The chapter-specific signal is what the senior-team exercise concluded: starting from scratch with the current tools, the company would not have been built the way it was built. The Sequoia discussion from Dorsey and Roelof Botha laid out the replacement team shape — three roles (Individual Contributor, Directly Responsible Individual, Player-Coach), a target reporting depth of two to three layers from CEO to anyone in the company, and an intelligence layer doing the work the middle layers used to do. The substrate came first; the team redesign followed.
Six structural failure modes recur
The failure modes are as structural as the mechanism underneath the chapter. Each pairs a specific misapplication of the substrate-and-redesign move with the predictable organizational cost it produces.
- Flat structure without substrate. Remove layers, keep the same information-compression practices, and the organization collapses within a year. Legibility and the compression hierarchy above must exist before the layers come out.
- Legible but still hierarchical. The intelligence layer gets built; the org chart does not change. The firm pays the substrate cost without capturing the team-size benefit. The most common mid-2026 failure mode among traditional enterprises attempting AI transformation.
- Instruction-file drift. The file is not maintained; the team works around it; agents still read it; the outputs become consistently wrong in ways that pass shallow validation.
- Executor without responsibilities. A team member holds none of the four human responsibilities and is only useful for executing narrow tasks. The role is redundant, whether or not anyone has said so.
- Overclaiming flat. Treating every workload as information-processing-bound. Hierarchy still holds for command-chain, physical-coordination, crisis-response, and compliance-approval work; flattening those surfaces produces the incidents hierarchy was designed to prevent.
- Middle management eliminated without redistribution. The layers come out, but the Architect and Validation responsibilities the layers carried do not get reassigned to anyone. Spec quality degrades, validation coverage drops, and the firm actually slows down.