Hacker News new | past | comments | ask | show | jobs | submit login

I don't think there is a language processor before or after it, just based upon the training data it's most likely tokens to return are spanish if question is largely in Spanish



It works decently well as a translator, correct? I wonder how it's been doing that - is it "native" to being an LLM or is it somehow processing it before?


It's native. There is nothing in the LLM except more LLM.

(And a safety model afterwards. And a tokenizer. But those things make it behave worse, not better.)




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: