Hacker Newsnew | past | comments | ask | show | jobs | submit | am17an's commentslogin

Isn't this factually wrong? Grok-4 used as much compute on RL as they did on pre-training. I'm sure GPT-5 was the same (or even more)

It was true for models up to o3, but there isn't enough public info to say much about GPT-5. Grok 4 seems to be the first major model that scaled RL compute 10x to near pre-training effort.

The point is those kernels exist already, you can just use them off the shelf. In the case where you're trying to write a production grade kernel without operating at that part of the stack... well good luck with that.

There’s a GitHub link which is open from last year, about the missing license in ollama. They have not bothered to reply, which goes to show how much they care. Also it’s a YC company, I see more and more morally bankrupt companies making the cut recently, why is that?


I think most of them were morally bankrupt, you might just be realizing now.


This is nice and useful because the new GPT-OSS model uses this technique. Kudos to the original authors!


And, as always, the FOSS ecosystem moves quickly, llama.cpp already fully support them! https://github.com/ggml-org/llama.cpp/pull/15157


Seriously, people astroturfing this thread by saying ollama has a new engine. It literally is the same engine that llama.cpp uses and georgi and slaren maintain! VC funding will make people so dishonest and just plain grifters


No one is astroturfing. You cannot run any model with just GGML. It's a tensor library. Yes, it adds value, but I don't think that saying that ollama also does is unfair.


I’ve never seen a PR on ggml from Ollama folks though. Could you mention one contribution you did?


Asymptotically, prompting is a programming language.


> Perhaps you meant to say “UNhelpful”?

Yes, thanks for pointing it out!


In the future there will be safe havens where LLM generated code has not been merged. It will marketed as “hand-crafted” by Romanian programmers or something like that, akin to Swiss watches. It will be extremely high quality, but too expensive to mass produce.


There's going to a wrapper around github PRs to summarize these issues, the mess they created in the first place. BTW this is the same guy which has famously stopped hiring engineers


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: