The $200 Bet: Anthropic Is Subsidizing You Today to Own You in 2028
How a loss-making subscription plan, a coding revolution and the unstoppable economics of inference cost decline are quietly building one of the most defensible moats in tech history.
The math that does not add up … until it does
On the surface, it looks like a bad deal for Anthropic. A Claude Max 20x subscriber pays $200 a month. In return, they receive roughly 900 messages per five-hour window, long agentic sessions, full Claude Code access and priority access to Anthropic’s most powerful models. Measured against raw API rates (Claude Opus 4.6 priced at $5 per million input tokens and $25 per million output tokens ) a genuinely heavy user of Claude Code could plausibly consume thousands of dollars worth of compute in a single month.
On January 5, 2025, OpenAI chief executive Sam Altman posted on X that his company was “currently losing money on OpenAI pro subscriptions” because users were consuming far more than anticipated. Altman acknowledged to TechCrunch “I personally chose the price and thought we would make some money”. Anthropic, by every available indicator, operates in the same territory. Internal financial documents obtained by the Wall Street Journal in November 2025 and subsequently reported by TechCrunch and Fortune show Anthropic expected to burn approximately $3 billion in cash in 2025, against $4.2 billion in sales with break-even targeted for 2028, contingent on gross margins reaching 77% by that year. Those margin targets are already under pressure: Anthropic’s inference costs ran 23% above internal projections in 2025, cutting gross margin to roughly 40% and prompting downward revisions to forward estimates, according to MarketWise’s reporting in March 2026.
This is not a pricing strategy. It is a territorial claim on the future.
Part I: The Inference Cost Curve Is the Whole Story
To understand what Anthropic is betting on, the starting point is one of the most remarkable and underreported economic phenomena in recent technology history: the collapse of AI inference costs.
Epoch AI’s research, published in 2025, found that the price to achieve GPT-4’s performance on PhD-level science questions fell by roughly 40 times per year between 2022 and 2025. Across different benchmarks, the range of annual price reductions runs from 9 times to 900 times depending on the task. The Stanford HAI AI Index 2025 report, released in April 2025, found that inference costs for GPT-3.5-level performance dropped over 280-fold in just 18 months between November 2022 and October 2024. At the hardware level, Stanford’s researchers found costs declining 30% annually while energy efficiency improved 40% each year.
The same capability that cost $20 per million tokens in late 2022 now costs less than $0.40.
The forces driving this compression are accelerating on multiple fronts simultaneously. AWS reduced H100 instance pricing by 44% in a single June 2025 announcement, bringing the cost from approximately $7 per hour to $3.90. Specialized providers now offer H100 access at $1.49 per hour. NVIDIA’s Blackwell architecture, widely deployed through 2025 and 2026, delivers up to 10 times cost reductions per token compared to the previous Hopper generation for the workloads that matter most.
The infrastructure cost of running a frontier model query will continue to fall. This is not speculative. It is the logical continuation of a three-year curve that has never reversed.
The Architecture Revolution
More significant than hardware gains is the fundamental shift in how AI models are constructed. Through 2025, virtually every leading frontier model migrated to Mixture-of-Experts (MoE) architecture. DeepSeek-V3, LLaMA 4 and Mistral Large 3 all adopted this approach, as did almost certainly Google’s Gemini family. The principle is structurally elegant: instead of activating all 600 billion parameters of a model for every single token, MoE models route tokens to only the most relevant specialized sub-networks, typically activating around 40 billion parameters out of 600 billion total. The result is the capacity of a massive model delivered at a fraction of the computational cost.
As NVIDIA’s technical documentation on its Blackwell platform confirms, MoE models on the GB200 NVL72 rack-scale system achieve 10 times better generational performance compared to the same models on previous generation H200 hardware. DeepSeek-V3 achieved frontier-level performance for under $6 million in total training compute, a figure that would have seemed impossible two years prior. Beyond MoE, speculative decoding, quantization (reducing numerical precision from FP16 to INT4) and prompt caching each contribute compounding cost reductions.
A November 2025 arXiv preprint from researchers applying a benchmark-level rather than token-level methodology found that inference costs for reaching any given capability level are declining at roughly 5 to 10 times per year. Separately, researcher JS Denain’s analysis published via Epoch AI’s Substack found that a complex task requiring 43 million output tokens in April 2025 required only 5 million tokens to complete at equivalent quality in December 2025 (roughly 3 times cost reduction in eight months) driven purely by model efficiency improvements.
Anthropic’s own model trajectory validates the underlying bet. Claude Opus 4.5 delivered Opus-class capability at $5 per million tokens, a 67% price reduction versus the prior Opus generation priced at $15 per million. Opus 4.6, released February 5, 2026, launched at that same $5 price while offering significantly improved long-context retrieval, agent team capabilities and 128K token output windows.
More capability. Same price.
The Bet Stated Plainly
A full-throttle Max 20x user likely costs Anthropic $1.000 to $5.000 in compute to serve for $200 today. That is a painful subsidy. In 18 months, given the documented cost trajectory, the same usage pattern might cost $200 to $500 to serve. In 36 months, perhaps $50 to $100. The economics reverse completely and Anthropic will have spent those 36 months building one of the deepest habitual user bases in the history of software tooling.
Compute is not the product being sold. Dependency, familiarity and professional identity are. Those are not commodities.
Part II: Why Coding Is the Perfect Vertical
Of all the categories Anthropic could have chosen to anchor its land grab, it chose software development and the results have been extraordinary.
Claude Code launched as a research preview in February 2025 and reached general availability in May. By November 2025, Anthropic confirmed that the product alone had surpassed $1 billion in annualized run-rate revenue $1 billion in annualized run-rate revenue (Anthropic, Dec 2 2025). By February 2026, per Anthropic’s Series G announcement, that figure had more than doubled to $2.5 billion (Anthropic, Feb 2026). The broader company trajectory was equally striking: Reuters reported in May 2025 that Anthropic’s overall annualized revenue crossed $3 billion that month, up from roughly $1 billion in December 2024, a growth that Meritech General Partner Alex Clayton, who has no financial stake in the company, described as a rate he had never seen across more than 200 public software company IPOs. Bloomberg reported in March 2026 that the run rate had reached $9 billion at year-end 2025 and had surpassed $19 billion by early March 2026, driven in each instance primarily by Claude Code adoption.
Three compounding reasons explain coding’s strategic centrality.
Code either runs or it does not. Unlike most AI use cases, software development produces outputs that are immediately verifiable through tests and execution making productivity gains from AI coding assistants measurable and undeniable, giving enterprises a clear return-on-investment calculation. As Fortune reported in February 2026, Spotify co-CEO Gustav Söderström stated that the company’s best developers had not written a single line of code manually since December 2025, with the streaming giant shipping over 50 new features through Claude Code-powered workflows. At an Anthropic enterprise event covered by VentureBeat in March 2026, the New York Stock Exchange’s chief technology officer Sridhar Masam described his organization as “rewiring our engineering process” with Claude Code and building internal AI agents that could take instructions from a Jira ticket all the way to a committed piece of code.
Coding also creates genuine switching costs not through hostage-taking but through workflow integration so deep it becomes invisible. Once a development team has built its muscle memory, its Slack integrations, its CI/CD pipelines and its CLAUDE.md configuration files around a specific AI coding assistant, switching becomes an act of self-disruption rather than a vendor decision. At a Seattle developer meetup in January 2026, a Google principal engineer publicly noted that Claude had reproduced a year of architectural work in one hour. That kind of experience does not produce curiosity about alternatives.
Third, and strategically most important, coding provides a developer-shaped wedge into enterprise. By fidelizing developers (the people who build, approve and architect corporate AI infrastructure) before expanding to the broader organization, Anthropic converted them into internal advocates. The Register reported in February 2026 on Anthropic’s partnership with CodePath, placing Claude Code in the hands of more than 20,000 students at community colleges and Historically Black Colleges and Universities. The Register noted the parallel directly: at the start of the personal computer revolution, Apple and Microsoft worked to get their products into schools precisely because early familiarity drives long-term retention. Anthropic is running the same play, against a larger addressable market, at considerably higher velocity.
Part III: The Historical Rhyme
History contains several technologies that appeared unprofitably expensive during their formative period and became structurally cheap, with the companies that subsidized early adoption reaping advantages that compounded for decades.
AOL’s Free Hours
In the early 1990s, AOL mailed floppy disks offering free internet access to millions of American households. The economics looked absurd: dial-up access was metered and expensive, and AOL was distributing a perishable commodity at a loss. The Smithsonian’s National Museum of American History documents the strategy’s architect, chief marketing officer Jan Brandt, as recognizing something her competitors had not: the internet was a network-effect business, and no one would pay for it until they had experienced it. The campaign generated conversion rates of 10 to 15 percent when the industry standard was 2 percent. Britannica’s account of AOL’s corporate history records that by the end of the 1990s, AOL had surpassed 20 million paying subscribers and briefly became the most valuable media company in the United States.
The parallel to Anthropic’s pricing structure is precise. The Max plan at $200 per month is Anthropic’s floppy disk: a subsidized introduction to a transformative technology, offered at a loss, designed to build habitual dependency before the underlying economics are sustainable. The difference is that Anthropic’s cost structure is improving at a rate AOL could not have imagined.
Moore’s Law and the Democratization of Computing
NBER research by economist Kenneth Flamm documents that quality-adjusted IT equipment prices declined 16% annually on average across the five decades from 1959 to 2009, accelerating to 23% per year in the late 1990s. CSIS’s analysis of Moore’s Law concludes that this relentless cost decline transformed the computer from a military research instrument to a household appliance and that the corollary effect was the enabling of semiconductor adoption across every major industry.
AI inference is Moore’s Law operating on a steeper curve. Hardware improvements account for part of it (30% annual decline at the chip level, per Stanford’s AI Index). AI additionally benefits from algorithmic improvements and software optimizations with no real analogue in the transistor-density world. Companies that subsidized computing access during the expensive early years built relationships that became structural. The hardware got cheap. The relationships persisted.
The Broadband Build-Out
In the late 1990s, internet service providers invested enormous sums to lay fiber optic cable, operating at substantial losses for years. Bandwidth costs fell as Butters’ Law predicted: the cost of transmitting a bit over optical fiber halved roughly every nine months. Companies that built user bases during the expensive era served those users at dramatically lower costs once the infrastructure matured. Those that sat out the land-grab period did not recover the ground.
AI infrastructure today occupies an analogous phase. The companies building habitual users now will serve those users at a fraction of the current cost within 24 to 36 months.
Part IV: The Big Picture
Anthropic is not competing for AI users. It is competing for professional identities.
When a developer describes themselves as a “Claude Code developer,” when an enterprise architect has Claude integrated into every step of their workflow, when a cohort of computer science students learns to code with Claude as their pair programmer … that relationship has moved beyond a customer relationship. Changing tools becomes an act of professional self-disruption. This is precisely what Microsoft accomplished with Office in the 1990s, and notably it had nothing to do with subsidization. Excel, Word and PowerPoint were not technically superior to all competitors in every dimension. Lotus 1–2–3 was a more capable spreadsheet at the time; WordPerfect was a widely beloved word processor. Microsoft prevailed not by being cheaper but by being everywhere: bundled into OEM deals, standardized across organizations, embedded in the muscle memory of a generation of knowledge workers. At some point, the tool stopped being a choice. It became the assumed medium of professional thought. Anthropic is pursuing the same destination through a different route: where Microsoft used distribution leverage, Anthropic is using the developer as the entry point, building irreplaceable habits at the individual level first and watching those habits calcify into organizational standards. The goal is identical: make switching not a vendor decision but an act of professional self-disruption.
The scale of what is being subsidized matters. At full utilization, the Max 20x plan likely delivers $2,000 to $5,000 in compute value for $200. That is a subsidy ratio of 10 to 25 times. No rational company accepts that ratio without a specific theory of the future. Anthropic’s theory: inference costs will fall 5 to 10 times annually; users who become dependent on Claude Max today will still be subscribers when the marginal cost of serving them drops to $20 to $50 per month; and the switching costs built through years of deep workflow integration will ensure they remain.
A second bet runs alongside the first. MoE architecture is already transforming cost structures. State Space Models address the quadratic bottleneck of traditional attention, enabling million-token context windows without exponential memory costs. Test-time training (models that adapt to specific tasks in real time) is emerging as the next structural frontier. Each architectural shift brings another step-change in cost reduction. Anthropic, operating at the frontier of model research, is positioned to ride each wave.
The Risks Are Real
Rate limit interventions signal the friction point in the subsidy model. In August 2025, Anthropic introduced weekly usage caps for Claude Code heavy users after what it described as abuse by a small subset of subscribers. The distribution of usage matters critically: if the average Max subscriber consumes 20 to 40 percent of their available capacity, moderate users subsidize heavy ones and the arithmetic holds. If heavy users cluster, the per-subscriber losses become unsustainable.
The open-source threat carries genuine weight. DeepSeek disrupted the pricing narrative in early 2025 by delivering GPT-4-level performance at a fraction of incumbent costs. Stanford’s AI Index 2025 found that the performance gap between leading proprietary and open-weight models had narrowed to 1.7% on some benchmarks. If capable open models become trivially self-hostable at scale, Anthropic’s pricing power weakens considerably and the value proposition shifts entirely to the ecosystem (Claude Code integrations, Cowork workflows, institutional knowledge embedded in configuration files) rather than raw model capability.
Google’s structural advantage warrants acknowledgment. Google AI Ultra at $249.99 per month bundles YouTube Premium, 30 terabytes of storage and 25,000 AI credits, all subsidized by over $300 billion in annual advertising revenue. Google owns its TPU infrastructure, eliminating cloud provider margins entirely. Anthropic does not have that cost structure. The bet works only if the inference cost curve falls fast enough to outpace Google’s subsidy capacity.
The Long Arc
The history of transformative technology is not a history of the best technology winning. It is a history of the technology that achieved sufficient habitual adoption before the economics became obvious to everyone else.
AOL did not win the early internet because dial-up was the optimal connection method. It won because it made the internet feel like something ordinary people could actually use, at the precise moment when no one else was doing that at scale.
Microsoft did not win the productivity suite wars because Word was the superior word processor. It won because it subsidized early adoption through bundling and OEM deals until switching costs rendered competition irrelevant.
AWS did not win cloud computing because Amazon possessed the most sophisticated infrastructure. It won because it offered compute at prices that made experimentation possible, built developer habits before competitors recognized the stakes and then watched the cost curve complete the rest of the work.
Anthropic is running the same play. The $200 Max plan that costs Anthropic $1,000 to $5,000 to serve today will cost $50 to $200 to serve in 2027. The users fidelized now like the developers who reach for Claude Code first, the knowledge workers routing every complex problem through Claude Cowork, the CS students who have never submitted a pull request without an AI collaborator, those users will still be subscribers when the economics reverse.
The quarterly losses are visible. The strategy pays out in 2027 and beyond, in a currency that does not appear on balance sheets.
The $200 plan is not a product. It is a land grab.
The $200 Bet: Anthropic Is Subsidizing You Today to Own You in 2028 was originally published in Mind In The Loop on Medium, where people are continuing the conversation by highlighting and responding to this story.


