Hacker News new | past | comments | ask | show | jobs | submit login

Could you explain more on how to do this? e.g if I am using the Claude API in my service, how would you suggest I go about setting up and controlling my own inference endpoint?



You can't. He means by using the open source models.


Runa local LLM tuned for coding on LM Studio. It has a server and provides endpoints.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: