Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> We use the same data filtering techniques as Gemma 1. Specifically, we filter the pre- training dataset to reduce the risk of unwanted or unsafe utterances.

Hmmm. I'd love to know what qualifies as "unsafe".



It will refuse to describe the process of making napalm using only double entendres.


I don't understand the point of this sort of censorship when I can go to google, ask how to make napalm, and get a million results telling me to dissolve styrofoam in gasoline.

I've seen documentaries and science shows on cable TV that demonstrate basic facts like this, or how the IRA produced IEDs, or how molotov cocktails were made in the spanish civil war.

The information is beyond easy to access, and has been for decades.


True. But an LLM product is closely associated with a single company and unlike a search engine which can claim it only shows you what is already available, the LLM will seem like it personally tells you something harmful. When they want to sell it as a helpful assistant that kind of behavior will undermine that goal.

We saw all the bad press companies have got in recent years for all kinds of unintended AI outputs.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: