Balaji’s Provocative Take: AI Supercharges Intelligence—But Only for the Already Smart
AI isn’t the great equalizer—it’s the ultimate amplifier. If you’re sharp, it makes you razor-edged. If you’re not? Well, good luck keeping up.
### The Smart Get Smarter
Balaji’s argument cuts through the hype: AI tools don’t magically confer genius. They compound existing intellect. The gap between the cognitively privileged and everyone else? It’s widening—fast.
### A Cynical Finance Twist
Wall Street’s already pricing this in. Hedge funds with PhD quants will lap retail traders using ChatGPT for stock picks. The rich get richer—even in brainpower.
### No Free Lunch
Forget ‘democratizing intelligence.’ The AI era rewards those who can ask the right questions, spot flawed logic, and—crucially—afford the best models. Everyone else gets auto-complete for their grocery lists.
AI makes you smarter, but only if you are already smart
Balaji doesn’t call it artificial intelligence. He calls it amplified intelligence. Because the AI isn’t acting on its own, it’s not fully agentic, it doesn’t set long-term goals, and it can’t verify its own output. “You have to spend a lot of effort on prompting, verifying, and system integrating,” he said. So, how useful AI is depends on how smart you are. If you give it bad instructions, it gives you bad results.
He also says AI doesn’t replace you, it just helps you do more jobs. With it, you can fake your way into being a passable UI designer or game animator. But don’t expect expert quality. AI makes you good enough to be average, not excellent. For real quality, you still need specialists.
There’s another job it does take, and that’s the job of the last version of itself. Midjourney pushed Stable Diffusion out of the workflow. GPT-4 took GPT-3’s spot. As Balaji puts it, “AI doesn’t take your job, it takes the job of the previous AI.” Once companies create a space for AI in a workflow, like image creation or code generation, that space stays filled. It just gets handed off to the newer, better model.
He also says AI is better at visuals than text. Easier for humans to judge a picture than to verify a wall of code or paragraphs of text. “User interfaces and images can easily be verified by the human eye,” Balaji says. With text, it’s slower and more costly for people to check the accuracy.
Crypto limits what AI can and can’t do
Balaji draws a line between how AI works and how crypto works. AI is probabilistic; it guesses based on patterns. But crypto is deterministic; it runs on hard, provable math. So crypto becomes a boundary that AI can’t easily cross.
AI might break captchas, but it can’t fake a blockchain balance. “AI makes everything fake, but crypto makes it real again,” he says. AI might solve simple equations, but cryptographic equations still block it.
There’s also already a version of killer AI out there. It’s drones. “Every country is pursuing it,” Balaji says. It’s not image generators or chatbots that pose the threat, it’s autonomous weapons. That’s the area where AI’s real-world impact is already lethal.
He argues that AI is decentralizing, not centralizing. Right now, there are tons of AI companies, not just one or two giants. Small teams with good tools can do a lot. And open-source models are improving fast. So even without massive budgets, small groups can build strong AI systems. That breaks up power instead of concentrating it.
Balaji also rejects the idea that more AI is always better. He says the ideal amount is not zero, not 100%. “0% AI is slow, but 100% AI is slop.” Real value lives in between. Too little AI means you’re behind. Too much, and quality falls apart. He compares it to the Laffer Curve, a concept in economics that says there’s a sweet spot between extremes.
In his final argument, he lays out why today’s systems are constrained AIs, not godlike machines. He breaks that into four kinds of limits:
- Economic: Every API call costs money. Using AI at scale isn’t free.
- Mathematical: AI can’t solve chaotic or cryptographic problems.
- Practical: You still need humans to prompt and verify results. AI can’t complete the full task alone.
- Physical: AI doesn’t gather real-world data on its own. It can’t sense its environment or interpret it like people do.
He ends by saying these limits might be removed later. It’s possible that future researchers can merge System 1 thinking (fast and intuitive, like AI) with System 2 thinking, which is more logical and careful, like traditional computing. But right now, that’s just theory. It’s still an open problem. There is no all-knowing AI. There are just tools (expensive, limited, competitive tools) that do what they’re told and need constant checking.
Cryptopolitan Academy: Want to grow your money in 2025? Learn how to do it with DeFi in our upcoming webclass. Save Your Spot