Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The link only mentions requirements to run inference at "decent speeds", without going into details about what they consider to be decent speeds.

In principle you can of course run any model on any hardware that has enough RAM. Whether the inference performance is acceptable depends your particular application.

I'd argue that for most non-interactive use cases, inference speed doesn't really matter and the cost benefit from running on CPUs vs GPUs might be worth it.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: