Table of Contents
A recent debate on X brought skepticism back to the surface:
How can Chutes (Subnet 64), earning roughly $54k/day in TAO emissions (halving to $27k), ever compete with centralized inference providers backed by billions in venture capital?
It's a fair question. Inference is a scale game. Whoever delivers the cheapest, fastest compute wins the volume. And when your competitors are flush with VC cash and enterprise contracts, it can seem like an unwinnable fight.
But this framing misses what makes Chutes structurally different from every other inference provider in the market. Chutes is not a company that buys and operates its own GPUs. It does not sign datacenter leases or maintain an enterprise sales team. It has no cap table full of investors demanding a return on capital.
Chutes is a decentralized compute marketplace. And that distinction changes the economics entirely.
What Is Chutes?
Chutes is an open-source, decentralized compute provider for deploying, scaling, and running open-source AI models in production. It operates as Subnet 64 on Bittensor, currently the #1 subnet by emissions.
The mechanics work like this:
GPU operators (miners) register their hardware and deploy models as endpoints called "chutes." Developers send inference requests, and miners compete to serve them. Validators audit fairness and quality. Revenue flows in as fiat; miners are scored on compute capacity (55%), speed (20%), availability (20%), and bounties (5%) over rolling 7-day windows.
The throughput numbers reflect production-scale usage:
- 34+ trillion tokens processed (lifetime)
- ~120 billion tokens/day post-monetization, with peaks at 160B
- 696,000+ users (excluding OpenRouter)
- $5.5M annualized revenue (75% organic, 25% sponsored)
- 50+ models supported across LLMs, diffusion, speech, and embeddings
For context, OpenRouter, the major LLM API aggregator serving 5M+ developers across 60+ providers, processes roughly 1 trillion tokens per day. About 20-25% of Chutes' daily volume flows through OpenRouter, where Chutes ranks as the top provider.
The Structural Cost Advantage
This is the core of the bull case, and where critics get the analysis wrong.
Traditional inference providers operate on a familiar model: raise capital, sign long-term GPU contracts (or buy hardware outright), build out infrastructure, hire sales and ops teams, and then try to undercut competitors on price while keeping margins alive. Every dollar of compute they provision carries contract risk, counterparty risk, and capital allocation pressure. Investors expect returns. Margins need to exist.
Chutes flips this model. Its supply side is entirely permissionless. Any datacenter or individual GPU operator can plug into the network and start serving inference, as long as the hardware meets requirements. There are no contracts to negotiate, no vendors to vet, no procurement teams to staff.
The miners competing for emissions and inference fees create natural downward pressure on compute pricing. Chutes' GPU inventory runs at roughly 50-60% of market pricing for equivalent hardware. This is not because Chutes is subsidizing anything out of pocket. It's because miners are competing against each other for a share of TAO emissions, and many are running on sunk capital expenditure at marginal cost.
Think about a datacenter with $100M worth of idle B200s. The hardware is already bought and racking up costs whether it runs or not. Deploying on Chutes for whatever payout is available beats letting it collect dust. This dynamic creates a cost floor that venture-backed competitors cannot replicate without continuous fundraising or margin compression.
As Bittensor co-founder Const put it:
"Inference is highly subsidized cross industry. 40% cost coverage is a great place to be compared to labs at 1-2%. As the market for tokens reaches maturity everything will rise to >100%. At that point we will be at X multiples."
The GPU Supply Thesis
The structural advantage gets stronger over time, not weaker.
NVIDIA's Blackwell architecture (and Rubin after it) will deliver massive performance gains. Roughly 2,000 Blackwells can do the work of 8,000 Hoppers. When datacenters upgrade, where do the displaced Hoppers and H200s go?
They need somewhere to earn a return. Networks like Chutes, Targon, and Lium become natural destinations for stranded compute. The supply of available GPUs flowing into decentralized networks will grow as each hardware generation rolls over, and incentive mechanisms will route that compute to wherever demand exists.
This is not a one-time event. It's a recurring cycle tied to NVIDIA's product roadmap. Each generation displaces the last, expanding the pool of hardware that can profitably serve inference on decentralized networks at prices centralized providers cannot match.
Why "Cheapest Inference Wins" Is the Wrong Frame
The most common bearish argument goes like this: inference is a commodity, and the moment a well-funded competitor undercuts Chutes on price, miners leave overnight.
We’ve spoken to the Chutes team about this. A high priority for them is to become more efficient with their GPU capacity by raising their overall utilization rates. They can do this by incentivizing more miners to host highly-demanded models, and spinning down models that have… https://t.co/hny9pkvxH4
— seth bloomberg (@bloomberg_seth) February 19, 2026
This misunderstands what miners are optimizing for. GPU operators on Chutes are earning TAO emissions on top of inference revenue. The combined yield (emissions + fees) on hardware they already own makes Chutes competitive even if a centralized provider offers lower per-token inference pricing. A pure price comparison ignores half the equation.
Beyond that, Chutes is building capabilities that centralized providers are structurally slow or unwilling to offer:
Permissionless access
No API keys, no Terms of Service risk, no accounts to get banned. For autonomous agents, trading bots, DeFi automation, and onchain programs, this is not a nice-to-have. It's a requirement.
Censorship resistance
Hyperscalers will always face regulatory pressure to restrict certain models, prompts, or use cases. Chutes doesn't have that constraint.
TEE-enabled privacy
Chutes runs GPU-level Trusted Execution Environments, meaning miners cannot access user prompts. Compare this to Venice (a project with similar valuation whose entire premise is privacy inference) where, by its own models' admission, the team could see prompts from their API and the GPU fleet lacks TEE. Chutes does not build its brand around privacy, yet it is further down the road toward true confidential inference than projects that do.
Crypto-native settlement and composability
Every model is a native API endpoint. For the growing ecosystem of AI agents that need to call inference programmatically, pay in crypto, and compose with other onchain services, Chutes is natural backend infrastructure.
The real question is not whether Chutes can beat AWS on commodity chat inference. It's whether a neutral, permissionless inference layer becomes necessary infrastructure as agent economies scale. If it does, the market could be very large.
The Revenue Flywheel
Chutes' economic model creates a self-reinforcing loop:
- TAO emissions subsidize compute costs, allowing Chutes to offer inference at below-market rates
- Low pricing attracts developers and users (696k+ and growing)
- Usage generates organic revenue ($5.5M annualized, 75% organic)
- Revenue flows to token buybacks (no cap table to service)
- Growing usage and subnet performance attract more miners
- More miners increase competition, which drives compute pricing down further
- Cheaper, faster inference attracts more users
This loop has a TAO price multiplier built in. Higher TAO prices increase the dollar value of emissions, which increases miner incentives, which brings more hardware online, which improves the service.
The key metric to watch is the organic-to-sponsored revenue ratio. At 75/25, it's already strong. A shift toward 80%+ organic would validate that real developer demand, not subsidized usage, is driving the throughput numbers.
Distribution and First-Mover Advantage
Chutes has consistently been among the first providers to offer newly released open-source models (DeepSeek V3, Kimi K2, and others). In inference, being first to serve a new model drives developer trial and creates sticky usage patterns.
This matters because Chutes' position as the top provider on OpenRouter gives it compounding distribution across millions of developers. OpenRouter is the default API aggregator for a large portion of the AI developer ecosystem. Ranking #1 there is not a vanity metric. It's a growth engine.
Recent integrations with Vercel SDK, n8n, and OpenClaw are expanding Chutes' reach beyond crypto-native users into the broader developer tooling ecosystem. Each integration creates a new funnel of users who may not know or care that their inference is running on Bittensor.
Risk Factors
No honest bull case ignores the risks.
Validator concentration
The main validator is operated by Rayon Labs (~16 H200 GPUs). Single-operator dependency is a real concern for network resilience.
Revenue quality
25% of revenue is sponsored inference, meaning user fees are partially subsidized. Real organic revenue may be lower than the headline figure suggests.
Inference commoditization
Together AI, RunPod, Fireworks, and DeepInfra are all racing to zero on pricing. Margins will compress across the board. The question is whether Chutes' structural cost advantage holds up as centralized competitors also get more efficient.
Model pricing pressure
Loss-leader pricing from models like DeepSeek may compress margins further across the entire inference market.
Emission dependency
If TAO emissions decline significantly (through halving or price drops) before organic revenue scales to compensate, the miner incentive structure weakens.
The Bull Case, Summarized
Chutes is not trying to be decentralized OpenAI. It's building a permissionless inference marketplace with structural cost advantages that centralized competitors cannot replicate.
The thesis rests on three pillars:
- Cost structure - Permissionless supply side + miner competition + sunk CapEx economics = a pricing floor that VC-funded competitors need continuous capital to match.
- Growing GPU supply - Hardware upgrade cycles create a recurring wave of displaced compute that flows toward networks offering the best yield. Chutes is positioned to absorb that supply.
- Agent-native infrastructure - Permissionless access, censorship resistance, TEE privacy, and crypto-native settlement make Chutes the natural inference backend for autonomous agents and onchain programs. If agent economies grow (and there are strong reasons to believe they will), this niche becomes very large.
$5.5M in annualized revenue is modest against centralized providers. It is the strongest revenue signal in Bittensor. The scale is not there yet. The trajectory is.
Disclaimer: This article is for informational purposes only and does not constitute financial, investment, or trading advice. The information provided should not be interpreted as an endorsement of any digital asset, security, or investment strategy. Readers should conduct their own research and consult with a licensed financial professional before making any investment decisions. The publisher and its contributors are not responsible for any losses that may arise from reliance on the information presented.