Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

With Ollama i got the 20B model running on 8 TitanX cards (2015). Ollama distributed the model so that the 15GB of vram required was split evenly accross the 8 cards. The tok/s were faster than reading speed.


For the price of 8 decade old Titan X cards, someone could pick up a single modern GPU with 16GB or more of RAM.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: