Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think gwern's point (whose opinion on deep learning I respect a lot more than his on genetics) is that GPT-3 is nearly not the end of the story. OpenAI released a GPT pretty much every year and each one is a spectacular improvement on the previous one with no sign of that trend ever stopping. If size is really what matters at all, there's no telling what GPT-4 or GPT-5 might be capable of, let alone a giant GPT run by state-sized actors.


The machine that trained GPT-3 is one of the largest GPU clusters in the world at the moment. (10k v100 GPUs hooked up with fast interconnect).

For comparison, Nvidia's "Selene" supercomputer only has 2k a100 GPUs and is #7 fastest supercomputer in the world: https://www.hpcwire.com/2020/06/22/nvidia-nabs-7-spot-on-top... a100 GPUs are around 3x faster than v100 GPUs.

It would take some time for governments to catch up :)


I think the point Gwern makes is that if a government wanted to they could easily allocate enough resources to do this. More bluntly I think he's saying if the US decided tomorrow to begin a Manhattan Project for AGI there is a non-zero chance that they might succeed in 7 years.


If they started a Manhattan Project for AGI, would we know by now? Or only after it explodes?


Only after, secrecy and surprise is the whole point.


1. This is a list of publicly known supercomputers (it’s likely that the intelligence community have larger secret assets).

2. Most of the largest super computers are owned by states.


Lots of large companies don't include their clusters on those supercomputer lists. Including Google. IDK about DoD, but I wouldn't be surprised if they have top 10 clusters we don't know about. The list of largest supercomputers is definitely only a list of the largest disclosed supercomputers.

That said, only so many actors could've secretly sourced 10K+ v100s with fast interconnect...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: