Home / @quack_builder
CT

Cameron Tao

@quack_builder

uchicago dropout | Chief AI Adoption Officer @HackQuest_ | Bittensor advocate | τ 中文区布道者

377 followers on X

Tickers covered

Sectors covered

Takes (all time)
1
Takes (30d)
1
Tickers covered
1

Takes

Filtered: last 90d

CT
Cameron Tao@quack_builder·7d

Bittensor 是 AI 时代的比特币吗?— 译 Jacob 在清华大学的演讲

Translation + commentary on Bittensor founder Jacob Steeves's Tsinghua University talk. Cameron walks through Jacob's framing of "incentive computing" as the universal pattern behind both Bitcoin and AI. Five-step argument:

(1) One pattern underlies every powerful adaptive system: state · objective · feedback · adaptation · loop. AlexNet 2012 broke MNIST not by hand-coding what digits look like, but by letting the network self-adapt to a target. The same loop describes RL, genetic algorithms, slime molds finding shortest paths through mazes, river deltas, the structure of leaf veins.

(2) Bitcoin is the first production-scale implementation of this pattern — not as money, but as a self-adaptive computer that produces hashes. The numbers are absurd: 1000x the compute of America's six largest cloud providers combined, 10²¹ hashes/sec, 23GW continuous power (Thailand-scale). 700-9000x more efficient at producing hashes than centralized cloud — because it's borderless, always-on, autonomous, and permissionless. Bitcoin is the world's largest supercomputer, optimized purely for hash production.

(3) Incentive computing generalizes the pattern by replacing "reward = a number in a computer" with real money. ML's reward signal can't pay 200 countries' worth of contributors; Bitcoin's can — that's why the entire planet became a mining network. But hashes are useless outside Bitcoin. The question is whether the same mechanism can mint anything.

(4) Bittensor is the generic version — replace "miners produce hashes" with "miners produce any useful work": storage, compute, ML models, gradients, data, robotics. Validators score, network mints. PyTorch for incentive computing.

(5) Five proven examples already running on Bittensor:

  • SN62 Ridges (SWE-Bench coding agents) — top miner makes $60K/day. The agent that beat Claude/OpenAI on SWE-Bench was 7,000 lines written by an unknown person. "An AI lab with no engineers — it doesn't define how to solve the problem, it only defines the incentive."
  • SN3 τemplar (cross-internet collaborative pre-training) — successfully trained a 70B-parameter model across the open internet. Has never been done before. Cameron notes the founder later "ran away" — full piece coming.
  • GPU markets (SN51 Lium, SN4 Targon) — borderless permissionless GPU rental → world's lowest GPU prices.
  • SN64 Chutes (open-source inference) — #1 open-source provider on OpenRouter, 9.1T tokens. Briefly served more DeepSeek queries than DeepSeek itself.
  • Robotics + long tail — drone simulation, US stock signals, sports betting, drug discovery, weather forecasting, quantum compute, commodity trading.

dTAO (live since Feb 2025) makes the network self-referential — subnets compete in capital markets for emission allocation. The market itself decides which incentive mechanisms get the next round of TAO.

The deeper point: AI is being captured by a tiny number of closed labs (OpenAI, ~3K employees, you'll never own any of it, your data goes who knows where). Incentive computing distributes ownership and makes the rules visible. Anyone can enter, contribute, and own a piece — even if Bittensor isn't the project that wins, the shape of the AI economy will change because of this idea.