Discussion about this post

User's avatar
Andrew's avatar
4hEdited

Nice summary, and I think your point at the end is also what a lot of people think when they hear Alexandr Wang or Aschenbrenner say "I trust the US government to handle this correctly! Elections = ethical and accountable officials". In fact I think some authoritarian states are structurally more accountable to their population in some ways.

I would point out that people like to ask "who gets control of superintelligence?" but the premise is the user has to be able to control and trust it in the first place, which is exactly the agent adoption problem. Giving unlimited access to your email and credit card is frankly insanity at this point in time. If superintelligence helps you build weapons for example, whatever it outputs has to be something the user understands and trusts to kill only who they want it to.

I did olympiad math and studied lots of very abstract math when I was young, worked on a lot of industry deep learning applications since circa 2016, and I would say that in my experience there is a sort of qualitative gap between being able to do math and being able to model things. It takes more brain power for a human to do advanced math but not everyone who is good at math comes up with good ideas for modelling (I've seen this with friends from top math PhD programs, though I would say most mathematicians are good at modelling because math already requires asking lots of questions). A lot of this is perhaps just real world context that is very messy.

It doesn't feel like AI has crossed that gap yet, but I don't rule out that it imminently could (I think at least a few more years though). When I saw what Deepmind was working on 10 years ago it hit me that humans really aren't as smart as we think. I thought we would have till ~2040 back in 2020, but I think everyone with a nice job should be prepared to be humbled and lose their job pretty soon. If your identity and pride revolves around your work, you need soul searching on top of financial preparation.

The one thing to note however is that this exponential increase in compute is physically constrained. The non-linear speedup has mainly come from bandwidth increases (because the GPUs of the past had nowhere near enough memory to store trillions of parameters, so you had to shuttle data over relatively big distances). "Thinking" is inherently a non-parallel kind of scaling (can have multiple models "talking to each other" but they still block each other). It is hard to find 10x more natural gas to pipe to one location or 10x more copper wire. Even if superintelligence can improve itself, scaling compute in non-parallelizable areas is an uphill climb.

The run in SNDK for example makes a lot of sense given they want to build stacked NAND flash. You can store a core model that is well trained for logical reasoning and language on HBF. It's both cheaper and more dense with only a small read speed penalty. However these kinds of improvements sort of become harder to find going forward. At some point, it will be obvious 10x the compute just doesn't deliver 10x the economic value anymore, and finding that point will be the key to trading this run successfully.

Andrew's avatar

On the question of "who is allowed to have superintelligence", I'd like to make another comment: A lot of top researchers are not American and would not want the US to have this monopoly on power. This includes employees at Anthropic and OpenAI. This includes decelerationists (counterexample being Sutskever who IIRC thinks the US should have the lead to provide regulation). Out of the American researchers, a decent number have globalist leanings as well. Their "loyalty" to the US as defined by Aschenbrenner et al can't be relied on either.

Right now the state of research is that there aren't many "trade secrets", people are still working through backlogs of ideas that many are broadly aware of. If we go to a world where the state of the art becomes a guarded secret, it is not clear the US government would have the state of the art if it restricts access to TS/SCI researchers. This is not 1945.

If those private lab employees can tell their talents aren't needed once the AI can self-improve, they have nothing to lose from proliferating the state of the art, and might even feel that is more ethical than allowing a OpenAI or NSA monopoly of power.

We can see examples of the globalist leanings if you go on Twitter and observe for example that Lucas Beyer, Karpathy, Hinton, etc. seem to be supportive of Qwen and Deepseek and Moonshot developments.

No posts

Ready for more?