Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> which presumably hasn't done a fresh pre-training over the web

What makes you think that?

> Did they figure out how to do more incremental knowledge updates somehow?

It's simple. You take the existing model and continue pretraining with newly collected data.





A leak reported on by semi-analyses stated that they haven't pre-trained a new model since 4o due to compute constraints.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: