Hacker News new | past | comments | ask | show | jobs | submit | more Palmik's comments login

Would love to hear how you went about doing things at Framer!


Quite shady indeed: https://www.saashub.com/theres-an-ai-alternatives -- see the icon, it's almost the same as https://theresanaiforthat.com/


The name is a prefix of the most popular website in the niche: thereisanaiforthat.com. Maybe some shady SEO tactic?

There are tons of these. Many of them are quite fishy -- they often reach out to you offering you to list your tool for a fee -- there are only a few directories where it's worth to do so.

Btw, was this supposed to be a Show HN?


For a reference, that's at least 40% more than what H100 sxm would cost if you are willing to reserve for a month (so not apples to apples).

H100 will also be much faster, especially if you are willing to use fp8. Maybe 3-4x


I don't think these analogies work.

Meta provides open source code to modify the the weights (fine tune the model). In this context, fine-tuning the model is better converted to being able to modify the code of the game.


So do video game developers (provide source code to modify their games) the analogy absolutely works. I can list a huge amount of actually open source software that I can see the source code and data for which is very different from Llama etc.


In the LLM world there are many open source solutions to find tuning, maybe the best one being from Meta: https://github.com/pytorch/torchtune

In terms of inference and interface (since you mentioned comfy) there are many truly open source options such as vLLM (though there isn't a single really performant open source solution for inference yet).


Thanks! Good to know.


It's much better than sharing a binary artifact of regular software, since the weights can be and are easily and frequently modified by fine tuning the model. This means you can modify the "binary artifact" to your needs, similar to how you might change the code of open source software to add features etc.


They specifically call out fp8 aware training and TensoRT LLM is really good (efficient) with fp8 inference on H100 and other hopper cards. It's possible that they run the 7b natively in fp16 as smaller models suffer more from even "modest" quantization like this.


People arguing that this is "just extension" are ignoring the fact that extensions have special priviledges compared to websites, and you would not want all websites to have the full power of arbitrary extension.

If it's "just extension", make it available to all domains.


The primary "special privilege" is that the extension is shipped with the browser and hidden. The API itself is available to any extension developer.

https://developer.chrome.com/docs/extensions/reference/api/s...


It has fp8 support. Not sure whether fp8 on MI300x is supported by vLLM yet.

Also, many of these comparisons use vLLM for both setups, but for Nvidia you can and should use TensorRT-LLM which tends to do quite a bit better than vLLM at high loads.


Elio, the person who did the testing confirmed with me that he has fp8 working.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: