Hacker News new | past | comments | ask | show | jobs | submit login

I stopped at: "causal sequence of “thoughts” "



Interpretability research is basically a projection of the original function implemented by the neural network onto a sub-space of "explanatory" functions that people consider to be more understandable. You're right that the words they use to sell the research is completely nonsensical because the abstract process has nothing to do with anything causal.


All code is causal.


Which makes it entirely irrelevant as a descriptive term.


"Servers shall be strict in formulation and flexible in interpretation."




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: