Hacker News new | past | comments | ask | show | jobs | submit login

I have a fine tuned version of Mistral doing a really simple task and spitting out some JSON. I'm getting equivalent performance to GPT-4 on that specialized task. It's lower latency, it's outputting more tokens/sec., more reliable, private, and completely free.

I don't think we will have an Open Source GPT4 for a long time so this is sorta clickbait, but for the small, specialized tasks, tuned on high quality data, we are already in the "Linux" era of OSS models. They can do real, practical work.




Been my thought for awhile now.

Can you recommend where I can learn more about hardware requirements for running Mistral/Mixtral?


> completely free

Not according to my calculation. For low request rate it is likely more expensive than GPT4.


How are you guys fine tuning?




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: