Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

A lot of negativity towards this and OpenAI in general. While skepticism is always good I wonder if this has crossed the line from reasoned into socially reinforced dogpiling.

My own experience with GPT 5 thinking and its predecessor o3, both of which I used a lot, is that they were super difficult to work with on technical tasks outside of software. They often wrote extremely dense, jargon filled responses that often contained fairly serious mistakes. As always the problem was/is that the mistakes were peppered in with some pretty good assistance and knowledge and its difficult to tell what’s what until you actually try implementing or simulating what is being discussed, and find it doesn’t work, sometimes for fundamental reasons that you would think the model would have told you about. And of course once you pointed these flaws out to the model, it would then explain the issues to you as if it had just discovered these things itself and was educating you about them. Infuriating.

One major problem I see is the RLHF seems to have shaped the responses so they only give the appearance of being correct to a reasonable reader. They use a lot of social signalling that we associate with competence and knowledgeability, and usually the replies are quite self consistent. That is they pass the test of looking to a regular person like a correct response. They just happen not to be. The model has become expert at fooling humans into believing what it’s saying rather than saying things that are functionally correct, because the RLHF didn’t rely on testing anything those replies suggested, it only evaluated what they looked like.

However, even with these negative experiences, these models are amazing. They enable things that you would simply not be able to get done otherwise, they just come with their own set of problems. And humans being humans, we overlook the good and go straight to the bad. I welcome any improvements to these models made today and I hope OpenAI are able to improve these shortcomings in the future.



I feel the same - a lot of negativity in these comments . At the same time, openai is following in the footsteps of previous American tech companies of making themselves indispensable to the extent that life becomes difficult without them, at which point they are too big to control.

These comments seem to be almost a involuntary reaction where people are trying to resist its influence.


precisely: o3 and gpt5t are great models, super smart and helpful for many things; but they love to talk in this ridiculously overcomplex, insanely terse, handwavy way. when it gets things right, it's awesome. when it confidently gets things wrong, it's infuriating.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: