The Cost-Parity Reversal
Why frontier AI economics break the substitution narrative, and what harness design is doing about it
§1 — Two Stories Pretending to Be One
The practitioners burning the most money on AI are developers.
Cursor power users are paying $200 to $500 a month out of pocket to run Opus 4.x for agent coding work. That’s not enterprise pricing with a procurement officer. It’s people handing over a car payment in exchange for access to frontier inference. At the same time, Anthropic’s Haiku 4.5 is cheaper than last year, OpenAI’s nano and mini tiers keep getting cheaper, Groq serves Llama at prices that look like rounding errors. Both things are true in the same year at the same companies. One model family is commodifying; another is appreciating. That’s not a single story about AI getting cheaper. It’s two stories running at two different tiers.
Most of the AI discourse treats them as one. The consensus narrative says AI gets cheaper, AI replaces labor, survivors are the AI owners and operators. That narrative assumes uniform downward cost drift. The assumption holds at one tier and breaks at another. The real error is extending a mid-tier substitution mechanism to all knowledge work, including the knowledge work the mechanism doesn’t touch. Which is most of the knowledge work that pays.
This essay is about the economic mechanism that forces the split — what I’ll call the cost-parity reversal — and about its architectural consequence: harness design as a first-class primitive the public market hasn’t built yet. Not a future-market-will-build pitch. The market is building it, privately, inside the big consulting firms and at the frontier labs. What’s missing is named, open, substrate-compoundable methodology in public view. This essay makes the economic case for why that methodology matters and then names the two layers it has to include.
§2 — The Three Curves
Three cost curves are running in different directions, and the discourse keeps mistaking one for another.
Start with training. Frontier model training is exponential, locked to hyperscalers, and trending worse. Meta’s Llama 4 training run reportedly ran into the hundreds of millions on industry estimates from Epoch AI and SemiAnalysis; frontier training runs overall are pushing toward billion-dollar scale. Frontier-class open weights now survive either state sponsorship (Qwen, DeepSeek) or mega-corp subsidy (Meta) for strategic reasons that have nothing to do with market economics. This is the curve people invoke when they say frontier AI will get cheaper over time, extrapolating from scaling laws into inference pricing. The extrapolation doesn’t hold. Training cost and inference cost are separate structures and they don’t co-move.
Per-token inference is the second curve, and it’s genuinely dropping. GPT-4 ran about $30 per million input tokens in 2023; Claude 4.6 Sonnet is around $3 per million in 2026, a 10x decline in roughly two and a half years. Haiku, Flash, and Llama-via-Groq have collapsed 10-50x at fixed-capability tiers. This is the mechanism driving mid-tier substitution, and it’s real. Klarna’s customer service layoffs, Duolingo’s contractor reductions, Philippines and India call-center contraction, junior developer hiring down substantially from 2022 peaks (estimates range from 40 to over 60 percent depending on source and geography) are all downstream of per-token cost falling below the threshold where human labor wins on economics. Substitution at mid-tier is happening, it’s accelerating, and nothing in this essay contests it.
There’s a third curve the discourse has mostly missed. A useful agent task at the frontier doesn’t cost tokens × price. It costs something closer to the product of several interacting factors: context × tool_calls × retries × verification_loops × per_token_price. These factors aren’t fully independent (retries modify tool-call count; verification loops overlap with retries; longer contexts cost more per token via attention scaling; prompt caching when available can reduce effective per-token price by up to 90% on repeated context). But the composition produces a floor. Longer context for multi-frame problems. More tool calls for integration that touches real systems. More retries for reliability under load. More verification loops for output that has to survive contact with the world. The net cost of a useful frontier task sits at parity with competent human labor, and it has been moving toward parity, not away from it, even as per-token inference drops.
That mechanism has a name, or at least should: the Jevons Paradox of Intelligence. As per-token inference gets cheaper, frontier work consumes exponentially more inference per task. Cheaper unit price, more units consumed per task, total cost roughly stable. It’s the same pattern William Stanley Jevons described for coal in The Coal Question (1865). Making coal combustion more efficient didn’t reduce coal use; it expanded what coal could do and grew aggregate consumption instead. Andreessen is right that Jevons applies to AI. He draws the wrong conclusion from it, which I’ll come back to.
The frontier-tier pricing signals are uniform. OpenAI o1-pro launched at $200 a month. (Altman has said publicly they’re losing money on that tier, which means the number reflects subsidized revenue-testing against VC patience rather than steady-state economics. Directional signal, not structural proof.) Anthropic raised prices on Claude 4.x over 4.0. Cursor power users are burning $200 to $500 a month. Claude Max. Gemini Ultra at $249.99. Every frontier-tier price movement in the past eighteen months has been up or flat. Not one has been down.
Beneath those pricing signals sits a physical constraint layer worth separating from transient supply issues. Grid power is the structural floor: 5 to 7 year bottlenecks on gas plant permitting, Virginia data-center grid queues extending into 2028 and beyond. HBM memory and CoWoS packaging are real constraints too, but they’re 2 to 4 year bottlenecks, solvable as TSMC and Samsung retool. Don’t bundle the two. Power is what the frontier economy can’t outrun on any near horizon. Packaging can be fabbed.
One more dimension the per-unit view misses: temporal liquidity. Compute offers a kind of liquidity human labor cannot. Spin up a thousand agents for an hour, kill them when the job is done. This doesn’t appear in per-token pricing but dominates corporate decision-making for bursty parallelizable work. It specifically favors substitution at narrow-task scale, which reinforces the tier-split this whole essay depends on.
There’s a technical counter worth addressing: distillation. Claude Haiku 4.5 performs at roughly 80 to 90 percent of last year’s Opus capability at a fraction of the cost; OpenAI’s nano and mini tiers compress capability downward on similar timescales. If distillation pipelines continue compressing frontier capability into cheaper tiers on a 12-to-18-month cycle, the claim that “frontier cost-parity is structural” erodes. What’s frontier today becomes mid-tier next year at lower cost, and the tier-split becomes a moving target rather than a stable partition. The response: distillation compresses capability at a lag, and the frontier keeps moving. The question isn’t whether last-year-frontier runs cheaper next year (it does). The question is whether the current-frontier task class, the longest-horizon and highest-integration and most-accountability-sensitive work, stays at cost-parity with human labor. On current evidence, yes. §9 treats a strong version of the distillation trajectory as an invalidation signal.
The primary-source anchor for the task-horizon expansion half of the claim is METR’s March 2025 report, Measuring AI Ability to Complete Long Tasks. METR documents frontier AI’s 50%-task-completion time horizon doubling on roughly a seven-month cadence, from seconds-scale in 2019 to about an hour for Claude 3.7 Sonnet in early 2025. The paper’s methodology is a 170-task suite spanning software engineering, cybersecurity, and reasoning, calibrated against 800+ human professional baselines. Important caveat: METR measures capability-time directly. It does not measure cost-per-useful-task or make cost-parity claims. The cost-at-parity half of this essay’s mechanism is a composition on top of METR’s capability data, arguing that practitioners respond to both the capability-growth curve and the per-token-price-decline curve by scaling task complexity, which keeps useful-task cost at rough parity with human labor. METR tells us AI can handle longer task horizons over time; this essay adds the spending-composition argument. The assembly is the essay’s, not METR’s.
Three curves, three directions. Training up, inference down, agent-task flat. The consensus narrative collapses them into a single “AI is getting cheaper” story and then extends that story to all labor economics. The extension is where the error lives.
§3 — What the Substitution Narrative Got Right, and Where It Overextended
The consensus narrative doesn’t look like a narrative to the people telling it. It looks like a straight line from capability to economics to labor. Let me set it down in the form I’ve seen most often in print since 2023.
- Capability scales with compute. Demonstrated.
- Cost-per-capability drops uniformly as models scale. Assumed.
- At some inflection, AI per-unit-work drops below human per-unit-work across knowledge work. Projected.
- Mass substitution follows at scale. Predicted.
- Survivors are AI owners and operators. Concluded.
Step 1 is correct. Step 2 is correct at one tier and false at another, which is the crux. Step 3 is correct at narrow well-specified tasks and false at complex integrated work, for the mechanism reasons in §2. Step 4 is partially correct in that mid-tier substitution is happening. Step 5 is too coarse even if you grant the preceding steps.
Mid-tier substitution is real and worth naming directly, with the caveat that the canonical example is messier than the headline. Klarna eliminated roughly 700 customer service positions by deploying an AI agent across 35 languages in early 2024, and then walked back portions of the deployment in May 2025 after quality complaints, with CEO Sebastian Siemiatkowski publicly acknowledging that the AI-only approach had gone too far and that Klarna was re-hiring humans. That walk-back sharpens rather than refutes the mid-tier point. Substitution works where quality tolerance is high and fails where it isn’t, which is itself a scope-carveout consistent with §1.5. Duolingo cut contractor translator volume. Philippines and India call center employment has contracted multiple percentage points off 2022 peaks. Junior developer hiring is materially suppressed from 2022 peaks (estimates range from 40 to over 60 percent depending on source and geography), and the survivors face higher bars for equivalent compensation. These outcomes are downstream of per-token inference cost falling below the human-labor threshold for specific narrow task classes. Nothing in this essay contests them.
What this essay contests is the extension of the mid-tier mechanism to the frontier-integrated tier, and the resulting prediction that substitution will sweep through knowledge work uniformly. That extension requires step 2 to hold uniformly. It doesn’t. The cost curves run in three different directions, per §2.
The §1.5 scope carveout matters here. This essay addresses task classes meeting three conditions: complexity (outcome quality depends on holding multiple context frames simultaneously), accountability (human cognitive authority at the boundary is required by liability, trust, or relationship dynamics), and integration (output has to compose with surrounding human workflow rather than execute in isolation). That’s not most task classes. It is, however, most of the task classes knowledge workers get paid well to do. Senior software engineering, law, medicine, strategy consulting, research synthesis, design, negotiation, relationship-dependent sales, regulatory writing, executive decision-making. All sit above the threshold. Customer service triage, document classification, routine translation, basic content moderation, standard data entry sit below.
Below the threshold, substitution-era economics are operating and will continue to operate. Above the threshold, the cost-parity reversal argument applies. Treating the two populations as one is the narrative’s error.
The §1.5 conditions are contingent, not permanent. Trust requirements, liability regimes, integration patterns all evolve. If regulators accept AI-signed audits, accountability conditions for auditor work shift. If case law establishes AI-authored contracts as admissible without human countersignature, legal drafting loses an accountability constraint. The carveout holds as long as receiver-side and regulator-side trust heuristics hold. Those heuristics are currently stable and likely to remain so on the 5-to-10-year horizon this essay addresses, but they are not fixed features of the knowledge economy. When they shift, specific task classes will cross from above-threshold to below-threshold, and the tier boundary will move with them. The thesis is an argument about where the line sits now and why it’s currently stable, not a claim about where it has to sit forever.
The strongest academic critic to engage here is Daron Acemoglu, who won the 2024 Nobel for work on institutions and growth and has been publishing skeptical-of-large-macro-AI-effects papers for several years. His Simple Macroeconomics of AI (2024) estimates AI’s productivity contribution at roughly 0.06% annual TFP growth, implying about 1.1 to 1.6% GDP uplift over the next decade. That’s an order of magnitude below hyperscaler and consultancy forecasts.
Acemoglu’s framework uses a task-based model inherited from Autor, Levy, and Murnane’s 2003 work: the economy is a bundle of tasks, some automatable and some not, and AI shifts the boundary. His estimate multiplies roughly 4.6% of GDP exposed to AI by roughly 15% average cost savings to arrive at his number. The framing is careful and macro.
He is not, notably, arguing that automation always substitutes eventually. His actual position is almost the opposite. His so-so automation concept (first developed with Pascual Restrepo) holds that current AI deployment is being pointed the wrong direction, at labor substitution rather than worker complementarity, and that this is a suboptimal direction rather than an inevitable trajectory. In a February 2025 MIT Technology Review interview, he put it directly: “We’re using it too much for automation and not enough for providing expertise and information to workers.” He advocates explicitly for complementary uses, which is architecturally adjacent to what this essay calls augmentation. The direction he wants AI development to take is the direction the frontier-integrated tier economically requires.
The real Acemoglu steelman is different from the one I initially expected to engage. It’s this: even if the architectural argument in §5 and §6 is correct, the macro effect is small (5% of the economy, 0.06% TFP growth, 1.6% GDP). Your harness-as-primitive essay could be structurally true and yet irrelevant at the scale economic policy operates on. Why bother building named open methodology for a thin slice of economic activity?
The response has two parts.
First, the 5% of the economy Acemoglu is estimating is disproportionately the high-comp population. That 5% is not evenly distributed; it concentrates in knowledge-work roles that are above the §1.5 threshold. For the individuals in that population, the architectural question of whether you survive in an augmentation-frame role or get misclassified under a substitution-frame receiver is not a small question. It’s a career-scale question. Small macro effect is compatible with large micro effect for the affected.
Second, Acemoglu’s model is designed to measure productivity-at-economic-aggregate, which is the right tool for his question and the wrong tool for the architectural one. He measures whether AI makes the average task faster or cheaper. The essay’s claim is that at the frontier-integrated tier, neither happens independently of architecture. Cost-parity holds because the multiplicative agent-task cost structure stays at parity with human labor. The return on architectural investment (harness design) is what turns cost-parity into capability amplification. That return doesn’t show up in Acemoglu’s aggregator because the aggregator doesn’t measure the architecture-enabled productivity layer. It measures per-task cost savings.
So Acemoglu is directionally aligned (wants augmentation, skeptical of naive substitution claims) and measuring a different layer of the phenomenon (macro TFP versus architectural return on harness design). The two arguments are adjacent, not convergent and not contradictory. Engaging Acemoglu directly sharpens the essay’s claim rather than undermining it. The substitution narrative’s problem is that Acemoglu’s own rebuttal of naive substitution optimism is also a rebuttal of the consensus “AI is cheaper everywhere and that’s enough” story. The consensus hasn’t absorbed his argument either.
Two qualifications to the Acemoglu response are worth making explicit.
First, the “measuring a different layer” defense is structurally adjacent to the Solow paradox response of the 1980s: “computers are everywhere except in the productivity statistics.” That defense held for about 15 years before computing-era productivity gains showed up in TFP measurements. If harness-era productivity gains are architecturally real, they should appear in aggregate measurements on a 5-to-10-year lag, as happened with computing. The architectural-return-not-visible-in-aggregator defense is temporally bounded. If harness-era gains haven’t appeared in TFP by roughly 2031-2036, the defense fails and this essay needs revision.
Second, Acemoglu’s mechanism is different from this essay’s in a way worth naming. Acemoglu argues augmentation is a difficult policy and design choice the market has been failing to make. This essay argues cost-parity at the frontier tier makes augmentation economically forced rather than merely preferable. Those are different mechanisms. If Acemoglu is right about choice-architecture dominating, augmentation could stall out even when cost-parity favors it, because policy friction and default-substitution incentives override economic pressure. The essay’s bet is that economic pressure compounds past the policy friction for frontier-integrated work. That’s a bet, not a demonstration. Acemoglu’s version stays live as an alternative mechanism where the market fails to route toward the architecture this essay argues is economically required.
The narrative’s overextension is structural. It confuses a mid-tier mechanism with a frontier-tier mechanism and predicts the compound of both. That prediction isn’t slightly wrong. It’s wrong in a way that matters for what the knowledge economy should be building.
§4 — Augmentation as Economic Consequence, Not Slogan
When cost-parity holds at the frontier tier, value comes from capability amplification, not labor substitution. Two inequalities have to hold at once for augmentation to dominate:
human + AI > humanjustifies the AI costhuman + AI > AIjustifies the human cost
Both inequalities hold for tasks meeting the §1.5 scope. Below the threshold, the second inequality fails. AI alone beats human+AI on narrow well-specified work where the integration overhead doesn’t pay for itself, and substitution dominates. Above the threshold, both hold. The frame is task-class-gradient, not universal.
The most instructive historical case for the gradient is centaur chess. From about 2000 to 2014, top-level human+AI teams reliably beat top AI alone. The mechanism was clear: humans had strategic judgment and memory for long positional patterns that the engines of that era couldn’t match; engines had calculation depth and tactical accuracy that humans couldn’t match. Combined teams outperformed either alone. Then AI crossed a reliability threshold for the chess task class. The engines got deep enough and accurate enough that the human contribution became noise. Centaur chess died. Top human+AI teams lost to AI alone and kept losing.
The same mechanism applies to any bounded-objective task class with sufficient training signal as AI crosses threshold. For narrow well-specified code generation, routine translation, commodity summarization, basic customer triage, AI-alone is on track to cross the augmentation-dominates-AI threshold on two-to-five-year timescales. Those task classes leave the augmentation frame and enter the substitution frame, permanently. This is what §3 named as the mid-tier mechanism operating correctly.
Frontier-integrated work lacks a clean reliability threshold. The relevant task class is not “produce a correct output” but “produce a correct output that composes with a larger human workflow, survives contact with variable real-world integration, and carries accountability at the boundary.” There’s no single pass/fail benchmark. There’s a rolling judgment over multiple failure modes, none of which resolve to a single dial. AI-alone crossing “the threshold” for this task class isn’t a defined event, because the threshold is a vector, not a scalar. A scalar threshold is something AI crosses once and stays past, as the chess engines did. A vector threshold requires crossing every component simultaneously, under every composition the work encounters, under shifting accountability and integration conditions that are themselves moving targets. No single training regime optimizes for that intersection. That’s why the augmentation-dominant regime here is projected to persist substantially longer than for bounded-task classes, and why the centaur-chess death trajectory is not the correct historical analogy for frontier-integrated work.
This essay’s augmentation argument is not novel as an economic argument. A school of thought has been publishing the augmentation frame continuously since 2023, and the essay has to place itself accurately relative to that school rather than pretend it’s the first to notice.
Ethan Mollick’s Co-Intelligence (2024) is the most widely read practitioner-level treatment. His centaur/cyborg taxonomy distinguishes two working modes: centaurs split tasks by comparative advantage (the human decides statistical approach, the AI produces the graph), cyborgs interleave more tightly (initiating a sentence for the AI to complete, iterating paragraph-by-paragraph). His Jagged Frontier concept names the observation that AI capabilities are uneven across tasks in ways hard to predict in advance. Mollick is not making an economic argument. He’s making a practitioner methodology argument. The two are compatible but distinct.
David Autor’s 2024 Noema essay “AI Could Actually Help Rebuild the Middle Class” is the strongest academic version of the task-level augmentation argument. Autor’s move is to split roles into tasks and argue that AI augments some tasks while automating others within the same role, which preserves the role while shifting its composition. This is labor economics at the task-decomposition layer. It’s rigorous, peer-reviewable, and adjacent to what this essay claims.
Erik Brynjolfsson and Lindsey Raymond’s “Generative AI at Work” (2023, NBER) documents a 14% productivity gain for customer service agents using AI assistance, with the largest gains for the least-experienced workers. Brynjolfsson’s separate “Turing Trap” (2022) argues that “human-mimicry” AI (building machines to do what humans do, framed as substitution) is a productivity dead-end compared to “human-augmentation” AI (building machines that expand what humans can do). The Turing Trap framing is directly compatible with the direction this essay argues for.
Kevin Kelly’s The Inevitable (2016) contains the decade-old version of the augmentation-beats-substitution claim, cast at civilizational scale rather than labor-economic scale. Azeem Azhar’s Exponential View has carried the augmentation discourse in policy-adjacent circles for years. Daron Acemoglu’s so-so automation work, engaged at length in §3, is directionally aligned.
The economic claim, that augmentation dominates substitution for frontier-integrated work, is shared with this school. Positioning this essay as “market speaks substitution, I speak augmentation” does not survive a literature review. The augmentation frame is not this essay’s contribution.
What this essay adds is not the augmentation frame itself but two things underneath it.
The first is the economic mechanism. The augmentation school makes the augmentation claim largely on methodology or productivity-measurement grounds. Mollick argues practitioners do better with cyborg working modes. Autor argues task decomposition favors augmentation over role replacement. Brynjolfsson measures productivity gains empirically. None of them argues that frontier agent-task cost structurally holds at parity with human labor because of the context × tool_calls × retries × verification_loops multiplication, and that this is why frontier substitution stalls as a compute-economics matter independent of methodology or task-decomposition choices. The cost-parity-at-frontier-as-structural-not-transitional claim is what §1 and §2 of this essay carry. It’s economic mechanism, not methodology.
The second is the architectural claim that follows in §5 and §6. Mollick writes about centaur methodology at the practitioner layer. Autor writes about task decomposition at the labor-economics layer. Brynjolfsson writes about productivity gains at the empirical-measurement layer. None of them describes the architecture underneath that makes human+AI > (human, AI) reliable rather than artisanal. None of them names the authorial-interface layer. Harness design as an architectural primitive with the two specific layers §5 and §6 name is genuinely not in the published discourse as of April 2026.
So the essay’s relationship to the existing augmentation school is precise: it agrees on the direction, adds the specific cost-parity mechanism that makes the augmentation direction economically forced rather than merely methodologically preferable at the frontier tier, and names the architectural infrastructure that the school has been gesturing toward without specifying. The augmentation frame is borrowed. The mechanism and the architecture are what’s being contributed.
Having established the economic case, the essay turns to the architectural one.
§5 — Harness as Architectural Primitive
Augmentation-at-scale doesn’t happen by methodology alone. If human+AI > human and human+AI > AI are going to hold reliably across sessions and across people, not artisanally across individual clever practitioners, the combined system needs architecture. The practitioner can figure out how to work well with Claude in a chat window; that’s methodology. The architecture is what lets that work survive generator changes, social pressure inside multi-agent setups, compression errors that corrupt shared memory, and the slow drift of RLHF defaults over model generations. That is a different kind of problem, and it is mostly not being named.
A cognitive substrate for partnership work has to solve a specific list of structural problems.
The human’s input is compressed at boundaries. Into documents, into prompts, into shared memory. If the AI’s compression errors leak into the human’s compression, the human’s cognition is corrupted by the AI’s over time, silently. You need a generator-independence invariant at the compression boundary. Call this an epistemic firewall.
Multi-agent setups fail by mechanisms that aren’t in single-agent playbooks. Shared agent memories echo each other into a convergent, narrower cognitive space even when each individual agent was trained for diversity. Cross-agent reads have to respect structural saliency budgets, not just token budgets, or the collective develops social failure modes that look like groupthink from the outside and homogenization from the inside.
Identity substrate for a partnership has to survive the generator being swapped out. In my own case, the partnership substrate I’ve been developing and publishing since 2024 (continuity, activation, earned operational trust, accumulated working language) persists across generator substitutions. The substrate is the harness property. The generator is a pluggable inference resource.
Structural invariants beat discipline-based verification every time. If a behavior has to happen for the system to stay coherent (logging, audit events, structural constraints on AI output), it has to be enforced in code, not in prompts. Prompts drift under load. Code doesn’t.
I’ll use “flow” as shorthand for the harness methodology I’ve been developing. Two layers are relevant. flow-methodology is the public reference: the flagship paper A Structural Theory of Harnesses (Zenodo DOI 10.5281/zenodo.19570642), the open documentation at github.com/phillipclapham/flow-methodology, and FlowScript on GitHub, PyPI, and npm. The flow system is my full operational implementation, which combines the public methodology with private infrastructure (instance-specific continuity substrate, operational agents, scheduler, internal instrumentation). The research claim lives at the methodology layer; the operational system is proof-of-possibility that the methodology runs at real scale. When this essay names specific components (Canon Foundation, Social Foundation, Substrate Plug Layer, the authorial-interface layer), those references are to the public methodology layer unless marked otherwise.
Each of those structural problems has named solutions in flow-methodology’s public reference. Canon Foundation names the epistemic firewall mechanism; Social Foundation names the echo-system failure mode and cross-agent saliency discipline; the Substrate Plug Layer names how one partnership harness connects to varied generator substrates without fragmenting identity. Supporting primitives include FlowScript (forcing function for cognitive compression), the anti-gatekeeping structural layer (runtime enforcement of anti-RLHF discipline at the generation boundary), and the bilateral blackboard (architectural answer to cross-harness synthesis under Canon constraints). Some components are fully specified in public; others are named and architecturally described at the methodology layer while their operational implementations stay private. The research claim is about the architecture; the operational system is demonstration-of-feasibility, not the object of publication.
That list isn’t a product pitch. The public methodology exists as proof-of-possibility, not proof-of-monopoly. Dozens of other teams are building functional equivalents privately, many with deeper internal methodology than anything flow-methodology publishes. What makes a public instance useful is that it’s independently citable and buildable-on, not that it’s more developed than the private systems.
The practice landscape is bimodal, and the closed side is larger than the open side.
McKinsey AI Practice, BCG X, Accenture Applied Intelligence, Deloitte AI Institute, and PwC AI Factory together field thousands of consultants building enterprise-scale harness methodology under names like “AI adoption playbooks,” “enterprise AI governance frameworks,” and “human-in-the-loop design patterns.” None of it publishes. The work is billable consulting IP. Every engagement produces harness artifacts that stay inside the engagement. When the consultants leave, the methodology leaves with them.
Anthropic’s Forward Deployed Engineering team is building something structurally similar for enterprise customers, inferable from public job descriptions and customer-case writeups. Scale AI’s Agent Oversight positions describe real-time harness infrastructure for production agents: monitoring, intervention, control. These are harness positions. The companies don’t name the category because naming it would tell their competitors what they’re actually building.
So the category has two moat dimensions, and they are not the same dimension.
The first is distribution asymmetry: open methodology and published papers against proprietary, per-engagement, opaque practice. The flagship paper A Structural Theory of Harnesses is on Zenodo with a DOI. flow-methodology is on GitHub as open reference documentation. FlowScript is open source on GitHub and published to PyPI and npm. The open instances are citable; the closed versions aren’t. Distribution asymmetry erodes the moment a major consulting firm or frontier lab publishes a methodology paper. Anthropic’s FDE team is the closest public-competitor signal; the clock is already running on this dimension.
The second is compoundability asymmetry, and the refinement worth making explicit is that consulting firms DO compound methodology internally. McKinsey runs Practice Development. BCG’s Knowledge Express captures patterns. Accenture’s Centers of Excellence and Deloitte’s Insight Studios explicitly compound cross-engagement methodology. What those firms do not do is compound the methodology publicly. The asymmetry isn’t “they can’t compound.” It’s that their compoundability is firm-internal, subject to consultant turnover, and unavailable to the broader market as architectural reference. Artifacts stay inside the firm, and significant portions of the methodology walk out when consultants leave.
Open methodology has a different profile. The foundations published in flow-methodology’s reference (Canon, Social, Substrate Plug) compound across every future use by every practitioner who reads them. FlowScript compresses more work the longer it’s used. Open compoundability is public-methodology-layer and independent-practitioner-usable; closed compoundability is firm-internal and turnover-limited. Both are real compounding mechanisms. One is visible to anyone who wants to build on it; the other is not.
The strategic consequence is worth naming precisely: a competitor publishing a methodology paper erodes the first moat but not the second. The time window on distribution asymmetry narrows every quarter. The time window on compoundability asymmetry does not close on publication events, because per-engagement consulting IP structurally cannot compound no matter who publishes what.
The harness primitive is being built. Most of it just isn’t public.
§6 — The Second Primitive: Authorial Interface
Cognitive substrate gets you the thinking. It does not get you work that accountability-layer receivers will accept.
This is the gap that breaks careers quietly. A person working above the complexity threshold, using harness architecture well, produces analytical output that’s sharper than what they’d produce alone. The work is substantively correct. The thinking is theirs. The AI contribution is compression efficiency and cross-frame pattern-finding, both of which map onto human cognition rather than replacing it. This is augmentation working as designed.
Then the output hits a receiver (manager, customer, board, peer, regulator) and gets rejected as “AI slop.” Not because the thinking is wrong. Because the form triggered a classification heuristic that fires on surface markers.
The markers are specific and visible: bullet hierarchies, “three things” framings, uniform parallel construction across sections, em-dash-heavy prose, Additionally/Furthermore/Moreover transitions, asterisk subheadings, consistent register from opening to close. A receiver scanning for trust signals sees those markers and runs a classification that says “machine-generated,” often with detection-tool output treated as confirmation. Detection tools are themselves unreliable. General false-positive rates run 3 to 12 percent, rising to 61 percent for ESL and technical writers. No tool exceeds 85 percent accuracy. The receiver-side heuristic runs anyway. Classification fires on form, not content. Once the heuristic fires, the transaction gets rejected retroactively, regardless of quality.
This is Middle Intelligence made structural at the interface layer. The phrase names a receiver-side dismissal heuristic that fires when parse-cost exceeds attention-budget, dressed in the moral clothing of detecting dishonesty. A receiver who cannot parse above-threshold density at their own ceiling reaches for a dismissal category that lets the parse-cost feel like a principled complaint. “AI slop” is the label that shapes the complaint as detection of dishonesty rather than admission of ceiling. The accusation is structurally unfalsifiable from inside its own logic. You can’t defend yourself by proving you wrote it. The evidence the heuristic is running on is density, not authorship.
This is not theoretical. It’s live at the employment layer in 2026. A substantive multi-line analysis posted in a workplace Slack gets flagged as AI-generated and goes from asset to liability inside one message. A formal write-up gets coaching-trajectory escalation. The same person producing the same quality of thought in older formats passes unflagged. The trigger is the form, not the substance, and the receiver often cannot tell the difference.
There’s a deeper reframe worth surfacing. The anthropological version of the same claim reads the AI-slop accusation as a purity taboo, a jurisdictional move by a receiver culture marking which registers can and cannot make reality enforceable inside the tribe. Dense analysis doesn’t fail to parse; it oxidizes an anaerobic accountability-avoidance environment on contact, forcing commitments into infrastructure engineered to prevent them. The receivers who reject the register are not being irrational. They are defending the metabolic conditions their role depends on. Zero-obligation language is how positions stay unowned, decisions stay unmade, accountability stays diffused. Density is threat, not parse-cost. This framing is compatible with the Middle Intelligence mechanism above and sharper at the jurisdictional layer: what’s at stake is not cognitive capacity but which symbols can create obligations.
The architectural fix is a second harness layer. Call it the authorial-interface layer, distinct from and sitting on top of the cognitive substrate. The substrate is where the thinking happens. The interface is where the thinking becomes receivable. Both are necessary for accountability-class work; neither alone suffices. In the anthropological framing, the authorial-interface layer is a boundary regulator between two incompatible register-environments: the substrate produces density, the interface controls how density crosses the boundary without disrupting role-incentives the receiver depends on.
The mechanism has specific design properties.
It has to run at generation time, not as a post-hoc rewrite. This is the part that surprised me when I first named it. Rewriting AI-assisted output for human voice after the fact catches surface markers. You can strip em-dashes, replace “Additionally” with “Also,” break up parallel construction. What you don’t catch that way is register-uniformity across sections, sentence-rhythm uniformity within paragraphs, and the structural parallelism that survives word-level edits. The compression happening mid-generation drifts back to AI-default structure, and the drift is invisible to surface-level editing. You end up with rewritten output that still reads as machine-produced, because the bones are still machine-shaped.
Honest caveat on the generation-time-not-post-hoc claim: this is the essay’s strongest architectural assertion about the authorial-interface layer, and the evidence for it is drawn from working practice rather than controlled study. A double-blind comparison between generation-time humanization and skilled post-hoc editing, equally resourced, would test the architectural-necessity framing directly. If detection rates and receiver-acceptance rates converge across the two conditions, the claim collapses from “generation-time is architecturally required” to “generation-time is more reliable than post-hoc under time pressure and at scale,” which is a weaker version of the argument. §9 treats that outcome as an invalidation condition for the architectural framing specifically. The weaker version still matters; it just doesn’t carry the full §6 claim.
The fix is to load the interface layer as a generation-time mode. In flow-methodology, that mode is specified as [!humanize], and it activates before the first token of any output going to an accountability-class receiver. The mode specifies what to strip (bullet hierarchies when they’re doing AI-signature work rather than enumeration work, Additionally/Furthermore/Moreover transitions, perfect parallelism across paragraphs, hedging that reads as liability-avoidance, em-dashes deployed as default punctuation), what to insert (contractions where natural, sentence fragments for emphasis, first-person asides when they carry information, opinions stated directly rather than passive-voiced, rough edges that signal human cognitive effort), and what register to calibrate to (context-dependent: casual for forums, considered-informal for essays, professional-human for client email).
The calibration target is specific: senior technical professional writing quickly and thoughtfully. Not a language model performing “casual.” Not a human performing formality. A person with taste, time pressure, and opinions.
There’s a real tension between the two harness layers, and it’s worth naming directly.
The cognitive substrate layer is explicit, named, open. Flagship paper voice. FlowScript notation. Partnership visible in the documentation. The claim is: here is how human+AI collaboration works, here are the artifacts that make it work, here is the methodology that compounds across sessions. Transparency is a feature.
The authorial-interface layer is implicit, invisible, passing. The output is produced so the receiver cannot tell AI was involved. The claim is: here is what I produced, read it without the AI-detection heuristic firing. Invisibility is the feature. The two layers carry opposite interface policies.
Both live in the same harness architecture; they activate on different task classes. Flagship-paper-class artifacts are publications about methodology, technical papers, work where the AI involvement is itself part of what’s being examined. These foreground the cognitive substrate layer and leave the authorial interface on a light setting. Accountability-class artifacts are different: client deliverables, workplace communication, compliance writing, trust-dependent relationships. These activate the authorial interface from the first token and leave the cognitive substrate invisible in the output.
This essay, for what it’s worth, is written with the authorial-interface layer lightly active. The voice is considered-informal, paper-register. Em-dashes are sparse, contractions are present, sentence rhythm varies deliberately, some sections step into terser register than others. The calibration is for a Nemo Operans reader who has seen the substitution narrative but might not have seen the architectural frame. If the interface layer were fully active, this essay would sound different: less structured, more idiomatic, the section numbering stripped. If the interface layer were entirely absent, it would sound like a thesis paper, which is what the source document is.
Task-class classification is the gating function. Running full authorial-interface activation on a flagship-paper-class artifact strips the voice that makes the work recognizable. Running cognitive-substrate-mode with AI structural markers on an accountability-class artifact produces the employment-layer failure mode.
Most public discourse about AI-augmented work lives at the cognitive substrate layer. Flagship paper included. The authorial-interface layer is largely invisible in public treatment because its success condition is invisibility. Successful authorial-interface work produces output indistinguishable from unassisted human output, which means it produces no discourse trail. Practitioners who have learned this layer carry it as tacit craft. Most harness-era discussion does not name it, because the people who have the craft don’t advertise the craft, and the people who lack the craft don’t know what they lack.
That invisibility is the strategic asset for practitioners who have built it. It is also the missing architectural layer in published harness-era methodology. As of April 2026, no published treatment in the augmentation or harness literature names the authorial-interface layer as an architectural primitive distinct from cognitive substrate. Mollick, Autor, Brynjolfsson, Acemoglu, Azhar, Kelly, Andreessen, the flagship harness paper itself, all work at practitioner or labor-economics or cognitive-substrate layers without decomposing harness architecture into the two primitives this essay names. The novelty claim is subject to verification by readers with broader discourse awareness, but literature review to date has not found the distinction named. Including it in the public architecture is the contribution this essay most wants to make legible. The cognitive substrate layer is the part of harness work already being discussed. The authorial-interface layer is the part that is not.
§7 — Andreessen’s Jevons Counter
The sharpest counter to the thesis comes from Marc Andreessen and the a16z economic case for generative AI. The argument has three moves and they’re worth representing cleanly before responding.
First: the marginal cost of creation is going to zero. a16z frames generative AI as the third epoch of compute. The microchip drove marginal compute cost to zero. The Internet drove marginal distribution cost to zero. Generative AI is now driving marginal creation cost to zero. This is a civilizational-scale economic claim, not a narrow-task claim. The direction is unambiguous and supported by actual pricing data. a16z’s “LLMflation” observation documents inference cost dropping roughly 10x annually at equivalent performance tiers, from about $60 per million tokens in 2021 to roughly $0.06 in recent comparisons.
Second: Jevons paradox applies to AI. When the marginal cost of a good with elastic demand drops, demand more than compensates. Cheaper inference leads to more inference use, not less. Andreessen’s phrasing from a16z: “When the marginal cost of a good with elastic demand (e.g., compute or distribution) goes down, the demand more than increases to compensate. The result is more jobs, more economic expansion, and better goods for consumers.”
Third, implicitly: if cheaper inference drives expanding demand, then the cost-parity argument this essay makes is either wrong or irrelevant. The mechanism Andreessen is describing says inference gets cheaper and capability attempts expand at ever-more-complex tasks as the price curve keeps falling. The cost-parity-holds-at-the-frontier claim looks like a local observation that the larger price-decline trend will wash out.
The response to this counter is not to reject it. It’s to absorb it.
Andreessen’s unit-level Jevons argument is correct. Inference per token is cheaper and will keep getting cheaper. Demand for inference is elastic at the margins and will keep expanding. Enterprise generative AI spending grew from about $11.5 billion in 2024 to about $37 billion in 2025, a 3.2x year-over-year jump while per-token costs were dropping. That’s Jevons working exactly as described. Nothing in this essay contests the unit-level Jevons claim. Per-token pricing is in fact the mechanism driving the mid-tier substitution this essay acknowledges is real.
What the essay claims is that a different Jevons mechanism operates at the task level simultaneously, and produces the opposite aggregate result for frontier-integrated work.
Here’s the mechanism, framed explicitly as a behavioral conjecture about practitioner response rather than a demonstrated empirical regularity. Per-token inference gets 10x cheaper. Practitioners respond to the price signal by attempting more complex tasks with AI assistance: longer context, more tool calls, deeper verification loops, more sophisticated agent architectures. The composite cost structure of a useful frontier task (the product of interacting factors described in §2) stays roughly flat, because the practitioner is using the cheaper unit price to buy a more capable task rather than the same task cheaper. This is a conjecture about how humans respond to price signals when working at the top of their capability, supported by anecdotal evidence (Cursor $200-$500/month, enterprise GenAI spend rising 3.2x year-over-year while per-token costs fall) but not by controlled study.
The testable operationalization: if monthly per-seat frontier AI spend at organizations using AI on above-§1.5-threshold work is flat-to-rising despite per-token price declines, the task-level Jevons mechanism is operating. If per-seat spend drops at the same rate per-token prices drop, the mechanism isn’t operating and task-level substitution is proceeding at the frontier tier. §9 absorbs this as an additional empirical test.
This is unit-level Jevons producing task-level cost-flatness, not unit-level cost-decline producing task-level cost-decline. Both are Jevons mechanisms. They operate at different layers and they do different work. The decomposition is worth owning directly: Andreessen’s unit-level mechanism is the driver, and this essay’s task-level-flatness claim operates through his argument rather than around it. What the essay adds is that unit-level Jevons has different labor-economic implications at the frontier-integrated tier than unit-level Jevons alone suggests.
Andreessen’s framing optimizes for the unit-level view. It’s correct at that layer and that layer is the right one for the civilizational-scale claim a16z is making about aggregate economic expansion. Where the framing misses is the task-level accounting that matters for the labor substitution question. If you want to know whether AI substitutes for a senior software engineer, what matters isn’t whether tokens are cheap. It’s whether the useful engineering task the engineer was doing can be closed by an agent for less than the engineer’s hourly rate, reliably, in a way the customer accepts. That measurement is at the task level, and at the task level, unit-level price decline is being spent on task-level capability expansion rather than substitution.
So the response: yes, marginal inference cost is going to zero. Yes, Jevons applies. Yes, aggregate compute demand is growing much faster than per-token cost is falling. These are all real and this essay treats them as load-bearing evidence for the underlying claim, not counter-evidence. The argument is about how the expanded inference demand gets spent. It gets spent on task complexity growth, not on labor substitution at the frontier-integrated tier.
A related counter from the quality-acceptance angle is worth naming. Even if task-level Jevons holds on the supply side (practitioners spend per-token decline on complexity), demand-side quality standards can drift downward. A client might accept a $50 Good Enough AI brief over a $500 frontier-integrated augmented brief regardless of cost-parity on the practitioner side. That’s the reliability-threshold dynamic from §4 operating at the acceptance layer rather than the capability layer. The response is partial: §1.5’s scope carveout already excludes commodity-content-at-accepted-quality-loss. What quality-acceptance drift does is shift the boundary of what counts as above-threshold. Some task classes currently above-threshold will cross below as standards drift, and the tier-split this essay rests on will move with them. This reinforces the contingent-scope claim from §3 rather than refuting the core argument.
The substitution narrative that cites Andreessen’s Jevons argument as supporting labor substitution is doing the same move §3 called out in the broader consensus. It extends a mid-tier mechanism (inference cheap, narrow tasks substitutable) to a frontier-tier population it doesn’t cover (complex integrated tasks, cost-parity preserved by complexity growth). The Jevons argument supports the unit-level observation and is silent on the task-level one. Filling that silence with substitution claims is not Andreessen’s argument. It’s what the substitution narrative has grafted onto his argument.
§8 — What the Market Is Actually Building
The labor market is moving faster than the discourse. If the argument to this point is correct (cost-parity at the frontier, augmentation as economic consequence, harness design as the two-layer architectural primitive), then some portion of the economy should be hiring directly against those claims. The named roles should exist. They should have job descriptions that track what harness work actually is.
They do.
Anthropic Forward Deployed Engineering positions describe work at the intersection of engineering, consulting, and applied AI research: embedding with customers to build production systems that reliably deploy AI at scale. The role expectation is that the candidate has shipped artifacts demonstrating AI-native integration thinking, understands the productionization problem, and can operate as a technical founder inside a customer environment. That is harness work described in the language of enterprise deployment. Scale AI’s Agent Oversight positions describe real-time monitoring, intervention, and control of production AI agents, which is harness work described in the language of runtime operations. SPEQD, a smaller and more explicit outfit, has advertised a Founding Harness Engineer role using the category name directly. Catio, an architecture-IDE company, has advertised a Founding Solutions Architect role where the job description lines up with harness-methodology work more cleanly than the title suggests.
None of these positions say “harness engineer” in the job title except SPEQD’s. None of them name the two-layer architecture (cognitive substrate + authorial interface) that §5 and §6 describe. They describe the outer shape of the work in the language available to each company. But the shape they describe is the shape of the role this essay argues the augmentation era economically requires.
Honest framing: those positions are hypothesis-compatible with this essay’s argument. They are not confirmation of it. Each of them can be read two ways. The aspirational reading is that they’re architect-the-human-AI-trust-oversight-integration-layer roles and the market is forming exactly the category the essay names. The mundane reading is that they’re customer engineer roles who help enterprises deploy AI, traditional forward-deployed sales engineering with AI vocabulary bolted on. Current public evidence does not discriminate between the two readings. Interviews at these companies will, one way or the other.
What matters for the essay’s argument is the aggregate signal across roles, not the individual interpretation of any one. The signal is that companies at the frontier tier (labs that ship models, infrastructure companies that operate agents, consultancies that deploy AI at scale) are hiring for something they don’t yet have a stable public name for. The hiring is moving ahead of the discourse. The category is forming in the labor market before it’s forming in the writing.
The week this essay was drafted, the market provided additional signal worth naming. In 48 hours across April 21-22, 2026, three independent events pointed the same direction. Anthropic removed Claude Code from the $20 consumer Pro tier, producing the unusual spectacle of r/ClaudeAI and r/LocalLLaMA (two communities that rarely converge) reaching opposite conclusions from the same event. Cursor signed a multi-year enterprise contract with xAI, with an acquisition option attached, capturing the IDE layer on a timeframe that reads as structural rather than transient. Anthropic shipped Live Artifacts across paid plans: real-time dashboards, auto-refresh, version history, cross-session persistence. Taken together, the three signals converge. The individual-developer tool tier is being sunset at low-margin while the platform play ships at enterprise tiers. “Building apps WITH Claude” ($20/month tool-layer assuming substitution-era economics) is being replaced by “building apps ON Claude” (platform-layer that requires harness architecture to integrate). Different businesses, different moat structures. Anthropic’s own framing on the consumer-tier sunset: “usage has changed a lot and our current plans weren’t built for this.”
Industry discourse as recently as early April was projecting these dynamics on six-to-twelve-month timelines (BCG and WEF strategic-resilience commentary in April 2026 both used that framing). What happened instead compressed material portions of that timeline into 48 hours. The cost-parity reversal is not a future-tense forecast. The week’s events are consistent with the mechanism being currently live, and the clock on the distribution-asymmetry described earlier is not metaphorical.
The architectural components described throughout this essay predate the economic framing by substantial margins, which matters as a matter of audit record. The flagship paper A Structural Theory of Harnesses shipped April 14, 2026. Canon Foundation v1 and Social Foundation v1 shipped April 17. Substrate Plug Layer Foundation v1.1 shipped April 20. The authorial-interface layer mode described in §6 graduated to flow-methodology’s Proven-patterns library on February 24, 2026, more than eight weeks before the labor-market category this section describes appeared on any research horizon I was tracking. Causal direction: architectural-frame → economic-frame → market-observation. The record exists in git commits and episodic store timestamps for any reader who wants to verify the sequence. The architectural claim stands or falls on its own terms; the order in which the work was produced is a fact, and the fact cuts one way.
What hasn’t been built publicly is also worth naming.
There’s no open, named, substrate-compoundable harness methodology that includes both primitives. The consulting firms have the methodology but don’t publish it, and can’t compound it across engagements per §5. The frontier labs have internal tooling implied by their FDE positions but haven’t released it as public architecture. Mollick has the practitioner methodology in book form but doesn’t name the architectural layer. Autor has the labor-economics framing but doesn’t touch the architecture. The flow system, as one set of public artifacts, names both layers. Cognitive substrate in the flagship paper and Canon/Social/Substrate Plug foundations; authorial interface in the [!humanize] specification. Flow is one instance of the category, not the category itself.
The gap in public architecture isn’t because the architecture is secret. It’s because the category hasn’t been named yet at the level that invites independent implementations. Naming it is what this essay is doing.
The market is already building this. The writing is catching up.
§9 — Invalidation Conditions
The thesis is empirically contingent. It becomes invalid if the underlying cost-parity mechanism breaks. Being explicit about what would prove it wrong is what separates a claim worth making from a claim worth defending. Three signals are load-bearing.
Signal A: sustained frontier price decline of more than 30% over a six-month window. OpenAI, Anthropic, or Google drops frontier-tier pricing by that magnitude and holds. The mechanism would be that either a hyperscaler absorbed a capability breakthrough collapsing inference cost across the stack, or the VC subsidy underwriting current frontier-tier pricing ended and commodity economics arrived faster than this essay projects. Either path invalidates the cost-parity claim for the frontier, or at least accelerates the reliability-threshold dynamic described in §4 to a timescale that renders the essay’s argument local rather than structural.
Signal B: equivalent-quality, equivalent-integration-cost drops 2x year-over-year on a consistent METR-style benchmark. The measurement has to be quality-normalized. Naive cost-per-task declines don’t invalidate the thesis, because cost-per-task can fall through the “market accepts lower quality” channel (Jevons-in-reverse at the acceptance layer; Klarna’s customer service substitution fits this channel and is already carved out by §1.5 scope). What would be invalidating is cost declines on the composite task-horizon × quality-at-horizon × integration-completeness measure. That’s the architecturally meaningful signal, the one that closes the augmentation delta rather than merely shifting where on the quality curve the market settles.
Signal C: the physical constraints relieve. TSMC, HBM, and CoWoS constraints visibly ease through multiple public announcements of fab capacity coming online ahead of schedule, with measurable declines in wait times for compute procurement. Packaging constraints relieving is expected on a 2-to-4-year horizon and is not invalidating on its own. The harder test is grid power. If the 5-to-7-year bottleneck on gas plant permitting and data-center grid queues visibly compresses, that’s the structural floor moving, and the essay’s “power is what the frontier economy can’t outrun on any near horizon” claim fails.
Signal D: sustained distillation collapse. If what was Opus-class capability in 2025 runs at Sonnet or Haiku-class pricing by 2027 with no measurable degradation on METR-style integration benchmarks, the frontier/mid-tier distinction this essay rests on is transient rather than structural. Distillation is already happening continuously (Haiku 4.5 at 80-90 percent of last year’s Opus capability at a fraction of the cost). The question is whether it compresses into mid-tier without integration-quality loss. If yes on a 12-to-18-month cycle consistently, the tier-split collapses.
The trigger rule is straightforward. Any two of Signals A through D firing simultaneously is grounds for a v2 rewrite. Any one signal firing is grounds for revisiting the relevant section without full revision. Grid power is the hardest constraint to reverse and the most indicative of a durable floor; if it moves, that’s the canary, regardless of whether two signals have fired.
Three ongoing empirical tests, subordinate to the main signals, are worth stating as well.
The task-level Jevons operational test: if monthly per-seat frontier AI spend at organizations using AI on above-§1.5-threshold work is flat-to-rising despite per-token price declines, the mechanism named in §7 is operating. If per-seat spend drops at the rate per-token prices drop, it isn’t, and §7’s response to Andreessen weakens.
The authorial-interface architectural test: a double-blind comparison between generation-time humanization and skilled post-hoc editing, equally resourced, measuring detection rates and receiver-acceptance rates on accountability-class artifacts. If the two converge, §6’s architectural-necessity claim collapses to an efficiency claim. The layer is still useful; it’s just not architecturally required.
The architectural-return-in-aggregator test: if harness-era productivity gains are real at the frontier-integrated tier, they should appear in TFP measurements on a 5-to-10-year lag, as computing-era gains eventually did after Solow’s 1987 observation. If they haven’t appeared by roughly 2031-2036, the different-layer defense of Acemoglu in §3 fails and the thesis needs revision.
Quarterly review against these signals, or on-trigger when cognitive sensors detect relevant data, is the monitoring cadence. Invalidation is not failure. It’s healthy thesis evolution, cheaper for the writer and more useful for the reader than defending a position after the ground moves.
The argument isn’t that the cost-parity reversal mechanism is permanent. It’s that the mechanism is currently live, that the industry discourse has mostly not absorbed it, and that the architectural consequence (two-layer harness as primitive) is the structure the augmentation era needs. If that stops being true, the essay stops being useful. That’s fine. It’s actually the disposition the argument has been asking its readers to hold from the beginning.