Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Designing it is easy and always has been. Programming it is the bottleneck. Otherwise Nvidia wouldn't be in the lead.


but programming it is "import pytorch" - nothing nvidia-specific there.

the mass press is very impressed by Cuda, but at least if we're talking AI (and this article is, exclusively), it's not the right interface.

and in fact, Nv's lead, if it exists, is because they pushed tensor hardware earlier.


Someone does, in fact, have to implement everything underneath that `import` call, and that work is _very_ hard to do for things that don't closely match Nvidia's SIMT architecture. There's a reason people don't like using dataflow architectures, even though from a pure hardware PoV they're very powerful -- you can't map CUDA's, or Pytorch's, or Tensorflow's model of the world onto them.


I'm talking about adding Pytorch support for your special hardware.

Nv's lead is due to them having Pytorch support.


Eh if you're running in production you'll want something lower level and faster than pytorch.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: