Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

These transformer models are so huge, they require extremely expensive and specialist hardware beyond what enthusiasts and even many academica access to.

There is no chance in the near future consumers or Edge devices will be able to run these models locally, data is going to have to be fed back into the cloud.



Thanks for replying! I had no idea there were models this large. Feels a bit like going back to the mainframe age.


Smaller models with better performance are beginning to arrive. Things like RETRO, better training data, longer training time, and scale optimization will have these models on phones and desktops doing crazy things in the near future.


They are but performance is decreased. In many cases transformers are encoding vast amounts of training data within the insane number of parameters.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: