Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
hansworst
35 days ago
|
parent
|
context
|
favorite
| on:
30% drop in O1-preview accuracy when Putnam proble...
Isn't that just the LLM equivalent of hardcoding though?
Trasmatta
35 days ago
|
next
[–]
I wouldn't call that hardcoding, otherwise you'd have to call everything it does "hardcoded".
freehorse
35 days ago
|
prev
[–]
"Overfitting" would be a bit more accurate term if the problem lies in the specific examples existing in its training set in various forms, places, languages etc but with the same values.
Consider applying for YC's Spring batch! Applications are open till Feb 11.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: