The Shift
What AI does to your cost structure
Unit economics per task compress by an order of magnitude, and three shifts follow: AI spend moves from the IT line to the payroll line, gross margin re-rates behind the unit-cost move, and role composition tracks demand elasticity at the function level. Software production cost collapses as a fourth, related shift — the bottleneck moves from building the product to finding users for it.
The cost base of every firm restructures when products are produced by systems instead of by people. Unit cost per task compresses by an order of magnitude, and three downstream shifts arrive on the same clock: gross margin re-rates behind the unit-cost move, role composition tracks demand elasticity at the function level, and software production cost collapses so the bottleneck moves from building a product to finding users for one. A practical consequence sits between the driver and the shifts — AI spend moves off the IT line and lands next to payroll, changing who owns the budget and how the decision logic runs. Firms that restructure through the sequence now set the cost baseline, and the rest will spend the next two years trying to match it.
For professionals, the same restructuring reaches the individual role. Job descriptions stay the same while expected output changes, and a role is evaluated against what an AI-fluent counterpart produces with the same tools. The work that still carries margin is the operation of agent systems rather than the execution of individual tasks by hand.
Unit economics per task compress by an order of magnitude
Intuit's Q2 2026 earnings call (February 26 2026) anchors the shift at enterprise scale. CEO Sasan Goodarzi reported that Intuit's accounting agents categorized over 237 million transactions in January 2026 alone — more than half of all transactions the platform categorized that month — while more than three million customers used Intuit AI agents with all-time repeat engagement above 85 percent↗. Business Insider reporting from March 2026 adds the P&L side of the same picture: Intuit's AI investments generated approximately $90M in efficiencies in the first half of 2025 through task automation and AI-to-human matching in the assisted-tax segment.
Two structural features produce the compression. Human labor enters accounting as fixed OpEx in salary-sized chunks — annual planning commits headcount against next year's forecast, and a customer-service team hired to handle ten million tickets still costs the firm the same if ticket volume drops to eight million. Inference enters as metered variable cost billed per action — the firm pays for exactly what gets used, and the unit is measured in fractions of a cent rather than fractions of a salary. The two cost bases do not average cleanly, and the arithmetic of how each scales — not operational inefficiency — is what makes the old firm lose on every comparable unit of work.
The compression shows up in vendor-side pricing as well. Salesforce's May 2025 Agentforce rollout moved from a flat $2 per-conversation conversational-AI fee to a Flex Credits consumption model at roughly $0.10 per agent action — a 20× compression in the lowest denomination of the vendor price surface. One conversation can span many actions, so the shift does not literally cut a customer's bill twenty-fold; it signals that the vendor-side unit of work is already priced at the lower order of magnitude the token substrate supports.
A worked-example parameterization makes the shift portable across industries. Let L be the labor share of a function's cost base, R the fraction of that labor an agent can replace today, and C the per-task cost ratio of inference to the human baseline. New unit cost becomes (1 − L) + L · (1 − R) + L · R · C. Three plug-in examples:
- Customer-service function at L = 0.6, R = 0.7, C = 0.1 moves unit cost from 1.00 to 0.62, which — given unchanged revenue — lifts margin on that unit from 40 percent to about 78 percent.
- Law firm first-pass contract review at L = 0.85, R = 0.4, C = 0.15 lands unit cost at 0.70, a 21-point margin lift.
- Logistics dispatcher at L = 0.25 of function OpEx, R = 0.6, C = 0.1 compresses unit cost by 14 points — smaller in absolute terms but large relative to the function's existing margin.
The three variables move independently by industry, and a firm that plugs its own L/R/C into the formula gets a defensible starting estimate before any deployment.
Warning: The compression is real but not free. Per-task inference cost trends up with frontier-model routing; the standard hedges are dynamic routing (small model first, frontier model only when needed), per-request spend caps enforced at the orchestration layer, and per-function cost-per-resolution dashboards before any pilot scales beyond its starting volume.
Four structural drivers each restructure margin on its own axis
The compression is not a single force. Four mechanisms run in parallel, each answering a different question about how AI changes the margin line, and a firm whose plan addresses one of them while ignoring the others captures only the part of the move that driver covers.
The first is cost compression per task. Salesforce's Agentforce repricing from $2 per conversation to roughly $0.10 per agent action and Intuit's accounting agents categorizing 237 million transactions in January 2026 alone both measure this driver. The L/R/C arithmetic in the prior subsection makes it portable across industries. The failure mode is firms that capture the per-task drop but leave the rest of the cost stack (audit storage, eval runs, governance, change management) underbudgeted, then discover loaded AI cost runs two to three times the API line.
The second is speed compression of the deliberation cycle. The cycle is the loop a firm runs from question raised to artifact produced to decision made. AI-native firms collapse the meeting-then-document-then-meeting pattern into a single working session in which a clickable prototype enters the room with the people debating its direction. The prototype-instead-of-meeting rule is the operational form: the cost of producing a working artifact has dropped below the cost of an hour of senior time, and the firms that absorb the shift run their decisions against artifacts rather than against slides. The failure mode is keeping the meeting cadence intact while bolting AI onto the artifact-production step, which captures none of the cycle compression.
The third is Dunbar throughput dissolution. The classic management ceiling sits at roughly 15 direct reports because human attention compresses past that limit. An AI-fluent operator does not manage 15 humans; they manage one human-agent fleet whose throughput tracks compute spend rather than headcount. The 3.1 thesis on small AI-native teams follows directly from this driver — the team architecture aligns naturally with the management constraint that just lifted. The failure mode is firms that lift the agent layer in name only while keeping middle-management routing intact, which keeps the Dunbar ceiling in place even as the technical capability for a wider span has arrived.
The fourth is zero marginal cost of cloning a working agent. Once an agent reliably performs a task, the marginal cost of running another instance is the per-call inference price. Replicating a human team requires hiring, onboarding, and monthslong calibration. Replicating an agent requires copying a skill file. The driver is what produces the Bessemer "AI Supernova" pattern named earlier in the chapter: $1.13M ARR per FTE in year one is what becomes possible when the cost curve for adding capacity is the API bill rather than recruiting and salary. The failure mode is firms that discover a working agent and then under-replicate it, leaving the second and third instance unbuilt and the productivity lift confined to a single pilot.
The four drivers do not substitute for each other. A firm that captures the per-task cost compression without compressing the deliberation cycle still routes work through the old meeting topology and pays for the old coordination overhead. A firm that compresses cycles without lifting the Dunbar ceiling still bottlenecks at middle-management span. The four together are why the Block, Shopify, and Bessemer-Supernova evidence below describe the same phenomenon at different scales.
AI compute belongs in the payroll line rather than the IT line
The FinOps Foundation's State of FinOps 2026 survey (n=1,192) measures how quickly AI spend moved into mainstream financial governance: 98 percent of FinOps teams now manage AI spend, up from 63 percent a year earlier and 31 percent two years earlier↗. The number answers a question most CFOs had not surfaced two years ago. The follow-up question — which budget owner gets it — has a structural answer: once AI spend scales with work output rather than tool count, the decision logic matches payroll rather than procurement.
Scale reference points are public. Jensen Huang framed the per-engineer token budget at Nvidia's GTC 2026 keynote and on the All-In Podcast filmed at GTC's close: a $500,000 engineer should consume roughly $250,000 of tokens annually — half of base salary. Huang said Nvidia itself is trying to spend $2 billion a year on tokens for its engineering team and compared an engineer still working without tokens to a chip designer insisting on paper and pencil. Ramp's public articulation of the same principle fills in the near-term gap: "token consumption per employee today isn't even close to double-digit percentages of their salary. But if someone is 2× more productive with AI, you should be willing to spend their entire salary again in tokens". Ramp's public articulation of the same principle fills in the near-term gap: "token consumption per employee today isn't even close to double-digit percentages of their salary. But if someone is 2× more productive with AI, you should be willing to spend their entire salary again in tokens"↗. The direction across both statements is the same — token spend rises to meet the work it produces, and the CFO gets to the destination faster by building the budget line for it today rather than after the function-level spend has already passed the compensation-committee threshold.
Why this owner, not the CIO. IT procurement budgets optimize for total cost reduction and per-seat licensing. AI token budgets behave the opposite way — spend grows with output expansion, and throttling the line when it scales is exactly the wrong move. A function VP with strategic-finance oversight evaluates the per-dollar output question that the CIO's ledger cannot surface; once per-employee token spend crosses a few percent of loaded compensation in any function, the decision belongs with the person who owns headcount for that function, not with the person who owns the SaaS vendor list.
The API bill is not the whole cost. Teams building the first token-budget line underestimate the loaded number by roughly 40 to 60 percent because they count only inference spend. The real cost stack adds audit storage, eval runs, governance tooling, periodic retraining on updated internal data, and the change-management overhead of moving workflow off legacy tools. Practitioner reports with the cleanest numbers cite token-only cost at roughly 4 percent of payroll at scale, with full loaded AI cost closer to two to three times that. The FinOps KPI that matters is cost per successful task, not cost per token — a cheaper per-token model that requires more retries or more human correction is not actually cheaper.
Gross margin flips when labor gets replaced at machine speed
Two 2026 public companies put hard numbers on the margin move:
- Shopify (April 2025). Tobi Lütke's on-X memo acts as a forcing function: every hiring request must first prove that AI cannot perform the work, and the CEO expects the bar to read as obvious within two years. The memo does not show a realized margin shift; it shows a firm rewiring the hiring loop so the shift can arrive. It is the leading-indicator case.
- Block (February 2026). The Q4 2025 shareholder letter documents a workforce reduction from over 10,000 to just under 6,000 — approximately 40 percent — described by Dorsey as driven by "intelligence tools [that] have changed what it means to build and run a company." Management guided 2026 Adjusted Operating Income at $3.20 billion, a 26 percent margin, up 54 percent year over year↗. The companion Sequoia article from Dorsey and Botha describes the organizational half of the move — the firm now runs on three roles (Individual Contributor, Directly Responsible Individual, Player-Coach) and a $2M revenue-per-employee target. Block is the realized case.
The two cases anchor different clocks of the same move. Block's letter makes the sequencing explicit: the organizational changes "will begin to more meaningfully impact Adjusted Operating Income in the second quarter, with the full impact of our new cost structure improving profitability in the second half of the year." At firm scale the margin move takes quarters, even when unit-cost effects on a single function appear within weeks.
Block was already profitable when the cut landed. Q4 2025 adjusted operating income reached $588M, up 46 percent year-over-year, with gross-profit growth accelerating to 24 percent and the firm clearing the rule of forty in the same quarter. The 40 percent workforce reduction was not survival math. It was a velocity move taken from a position of strength, with growth re-accelerating through the cut rather than dipping. Reading the restructure as downsizing misses the operative claim: the firm sized itself to the new capability surface rather than to its prior topology, and the operating-income guide reflects the resized capability rather than a cost cut against a deteriorating revenue line.
Warning: The margin flip only compounds if the underlying process actually gets redesigned. A firm that deploys AI onto its old labor topology without touching role composition pays for both the tool and the redundant labor at the same time. Section 2.1 develops the software-and-org co-design pattern that produces the move; 3.3 develops the political and environmental work that makes the restructuring land at firm scale.
Demand elasticity decides which roles shrink and which grow
The intuition that AI lands as a cheaper workforce breaks against the data. Cheaper software production expands the universe of outputs the firm builds, so software-engineering demand rose as productivity rose; the Jevons-Paradox-style dynamic Mollick has described in public is the shape, even if the size of the expansion shows up in employer-level rather than BLS-level data right now. The mechanism is elasticity of demand meeting task substitutability: the firm runs both directions at once on different functions.
- Where demand for output is bounded, productivity gains compress headcount. A BDR team covers a fixed number of qualified leads per quarter; a contact-center queue clears a fixed inquiry volume; a financial-analyst team covers a set coverage list; a restaurant chain runs a fixed number of store-level shifts. Productivity doubles, and the work gets done with fewer people.
- Where demand for output is expandable, productivity gains hold or grow headcount. Software engineering, creative content, legal research, strategic analysis all sit in categories where the firm could use more output if it were cheaper. Productivity doubles, output doubles (or more), and headcount stays flat or grows.
BCG's April 2026 labor model makes the two-axis structure rigorous: AI augments versus substitutes, and demand for output is expandable versus bounded, producing six segments with their own headcount trajectories. Section 1.4 develops the per-segment shape a founder uses to pick markets and an executive uses to plan function-level restructuring; section 3.3 develops the political and environmental work required to land the restructuring at firm scale.
The consequence for role design: headcount plans built on "AI replaces junior, leaves senior intact" get both directions wrong. In an expandable-demand function, the senior role holds or grows; in a bounded-demand function, the senior role also compresses, slower than the junior but on the same trajectory. The diagnostic a COO runs per function is two questions — would the firm buy more of this output at ten-times-lower cost, and does the task structure decompose into AI-addressable sub-tasks today? — with the function landing in one of the six BCG segments based on the answers.
Software production collapses and the bottleneck moves to finding users
Frontier models already do most of what any given business product requires. Felix Rieseberg at Anthropic states the point directly: "Execution is essentially free; taste is the constraint." "The overhang is in the product, not the model" — the unshipped value sits in packaging the capabilities into workflows people actually use. The specific components of the overhang are concrete: the model ships on a monthly cadence, and everything else ships on firm time — the harness, the context engineering, the distribution reach, the judgment about which workflow to redesign first. The build cost collapsed; the rest did not.
Two 2026 cases sit at different scales of the same phenomenon:
- Polsia. Ben Broca's one-person company had reached a $4.5 million annual revenue run rate by late March 2026, with Broca as sole employee. The founder operates not alone but through what he calls an outsourced "virtual team" — "a GC with a law firm... an infrastructure team with an infrastructure-for-agents company" — incentivized to scale with the business rather than employed full-time. The architecture of the firm has already changed in ways that the legal frame for what counts as an employee has not.
- Ramp Glass. A small internal team built the product and spread it to 700 daily active users at Ramp within three months through lateral adoption rather than top-down rollout. The case shows that the distribution phase compresses as well when the product is built on top of a firm's existing identity, credential, and data infrastructure.
The aggregate shape lines up with Bessemer's State of AI 2025 benchmark. The top ten AI Supernovas Bessemer surveyed averaged $1.13M ARR per FTE in year one — four to five times the typical SaaS benchmark — on an unusual 25 percent gross margin (distribution cost through the current token-price surface is the active headwind). Steadier "Shooting Stars" reached $164K ARR per FTE with SaaS-like 60 percent margins and a growth trajectory that reaches $100M ARR in four years rather than the traditional SaaS trajectory. The build-cost collapse is uniform; the distribution-cost drag varies by go-to-market shape.
Consequence for founders. Development budget collapses into marketing budget. The first hire for a new product is a growth, content, or enterprise-GTM role; build work compresses into the founder's own hands and into the agent fleet the founder runs on top of a $49-to-$500 monthly subscription. The founder-level trigger for the go-to-market hire is the product reaching working-prototype status with early users, not code-complete; the old sequencing of engineering-first, GTM-after no longer matches the cost curve.
Consequence for incumbents. The competitor set widens by an order of magnitude in any category where the underlying work is information-based. A fully-automated entrant like Polsia runs at a cost structure the existing workforce cannot match in its current shape, and the only way to reach that structure is by running the sequence 1.2's traps describe, through the transformation logic 3.3 develops.
Section 1.4 develops where the gains land across sectors — which markets a founder should attack on the demand side, and which workforce segments an executive should identify on the supply side. The interstitial after 1.4 translates the shifts in this chapter into the autonomy-map exercise that is the first concrete action for any reader still holding the old cost structure.