Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It’s a curious wording. It mentions a process of improvement being attempted but not necessarily a result.




because all the safety stuff is bullshit. it's like asking a mirror company to make mirrors that modify the image to prevent the viewer from seeing anything they don't like

good fucking luck. these things are mirrors and they are not controllable. "safety" is bullshit, ESPECIALLY if real superintelligence was invented. Yeah, we're going to have guardrails that outsmart something 100x smarter than us? how's that supposed to work?

if you put in ugliness you'll get ugliness out of them and there's no escaping that.

people who want "safety" for these things are asking for a motor vehicle that isn't dangerous to operate. get real, physical reality is going to get in the way.


I think you are severely underestimating the amount of really bad stuff these things would say if the labs put no effort in here. Plus they have to optimize for some definition of good output regardless.

The term "safety" in the llm context is a little overloaded

Personally, I'm not a fan either - but it's not always obvious to the user when they're effectively poisoning their own context, and that's where these features are useful, still.


but... we do all drive motor vehicles, right.



Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: