Hacker News new | past | comments | ask | show | jobs | submit | jchonphoenix's comments login

Meta tools are best in class when the requirement is scale. Or that the external tools haven't matured yet


OpenAI initially raised 50m in their institutional round.

1b was a non profit donation, so there wasn't an expectation of returns on that one.


This is another inflection style "acquisition." Highly unethical of the founders and screws over all your employees and investors who are left holding the bag.

For those asking, c.ai has very high cost and looks like a typical consumer company that burns money for use, so they were decent on revenue but not near profitability.


> Character’s leaders told staff on Friday that investors would be bought out at a valuation of about $88 per share. That’s about 2.5 times the value of shares in Character’s 2023 Series A, which valued the company at $1 billion, they said.

https://www.theinformation.com/articles/google-hires-charact...


> investors would be bought out at a valuation of about $88 per share. That’s about 2.5 times the value of shares in Character’s 2023 Series A, which valued the company at $1 billion

Those investors almost certainly have a liquidation preference. How much did employee shareholders get? I'd guess zero.

"I am confident that the funds from the non-exclusive Google licensing agreement, together with the incredible Character.AI team, positions Character.AI for continued success in the future,” Shazeer said in a statement given to TechCrunch."

That's a pretty hilarious statement from a Founder/CEO, given the circumstances.


>liquidation preference. How much did employee shareholders get? I'd guess zero.

But they aren't filing for chapter 11? I assume all shareholders will be bought out, including the employees, and this will be paid for by Google who will license their models, presumably as a scheme to pay off of the investors as I doubt they actually need those models at all.

(assuming the linked source is correct.)


got it. so instead of employees and investors screwed, only employees are screwed..


To do better by the employees, the CEO really should have fought harder to have the whole company get acqui-hired, even if Google would have shut down the service. Maybe there were some other considerations that I’m not seeing (ie. There are good reasons the company should be kept going, and there’s a good path to success even without Noam. The article doesn’t specify how many employees are going over, so it’s hard to tell.) Landing the employees a relatively cushy Google SWE gig after helping build your company is the least you could do for them.


When you don't understand reasoning, always look at who the beneficiaries are in these situations.


I'm also wondering how much money they spend on legal fees, given that they are copying then likenesses of many celebrities without their permission (that's the only way I've heard abot them before).


Everybody has a price


don't forget Adept


"Low level python" is an amusing statement


> low-level Python implementation

Python is written in C, with raw pointers everywhere, manual reference counting, etc. Python C extensions are certainly low-level too. Knowing these details is important to write good code in some domains, and if you need to get as much performance out of Python as you can.


Technically the Field's medal is the top honor in mathematics.

But who am I to talk.


There's no formal hierarchy so there's nothing technical about it. Field's medal is certainly prestigious but it's not widely applicable, it has a bunch of restrictions that don't have anything to do with math itself, including an age limit. For example no one who has ever been awarded an Abel Prize would qualify for a Field's medal strictly due to age.


Six Abel laureates (out of 22) had previously won the Fields Medal: Serre, Aliyah, Thompson, Milner, Deligne, and Margulis.


True, it is considered one of the two top honors in Math since last decade. Previously it was the only distinguished prize.

There was a growing need for another award which bridged few gaps. The 40 year cutoff age, awarded every 4 years to living mathematical prodigies failed to honor several prominent mathematical breakthroughs which came after decades of painstaking research.

As the field has progressed, monumental breakthroughs are harder to come by early into career. Many of the ingenuity comes from cross-study of disciplines for e.g. Riemannian hypothesis being approached by Algebraic geometry and Topology rather than number theory. These require years of mastery - not just prodigy. Also the prize money offered by Abel Foundation is a good incentive for research into pure math


Yes the Fields is much more similar to the MacArthur Genius Grants.

The Abel is much more similar to the Nobel, though both the Abel and Fields are Nobel-caliber in prestige.


As an alum, if the students don't love it, they can withdraw. Someone needs to foot the bill for research and education and CMU does not have a large endowment fund


"If you don't like this country, just leave!"


Arguably one of the earliest consumer use cases that found footing was AI girlfriend/boyfriend. Large amounts of revenue spread across many small players are generated here but it's glossed over due to the category.


Weird how close your user name is to Joaquin Phoenix, star of the film "Her" centered around an AI girlfriend.


Given that how widespread romance scam schemes already are (the "market" is at least $0.5 billion/year), I would expect any reasonably functioning AI girlfriend/boyfriend model to be massively (ab)used also against unwilling/unwitting "partners".


I think one related area we'll start seeing more of in the future is "resurrected" companions. You have a terminally ill family member, so you train a model on a bunch of video recordings of them, then you can talk to "them" after they've shuffled off this mortal coil.


Licensing my soul as closed source now ..


DNR: Do Not Retrain


Be right back.


I think this is why we're seeing that the type of founders YC usually funds in these industries aren't going through YC and choosing alternative methods of getting started.


Hawaii has a lot of sun. Not sure they can be a model for any of the mainland that doesn't look like their weather.


None of this matters if they can't get the hardware stack to work correctly.

The media keeps missing the real lock in Nvidia has: CUDA. It's not the hardware. It's the ability for someone to use it painlessly.


TPUs have the second best software stack after CUDA though. JAX and Tensorflow support it before CUDA in some cases and it's the only Pytorch environment that comes close to CUDA for support.


TPUs are single case use, contrary to CUDA.


Google has historically been weak at breaking into markets that someone else has already established and I think the TPUs are suffering from the same fate. There is not enough investment in making the chips compatible with anything other googles preferred stack (which happens to not be the established industry stack). Committing to getting torch to switch from device = “cuda” to device = “tpu” (or whatever) without breaking the models would go a long way imo.


I always thought Google was actually pretty good at taking over established, or rising markets, depending on the opportunity or threat they see from a competitor. Either by timely acquisition and/or ability to scale faster due to their own infrastructure capabilities.

- Google search (vs previous entrenched search engines in the early '00s)

- Adsense/doubleclick (vs early ad networks at the time)

- Gmail (vs aol, hotmail, etc)

- Android (vs iOS, palm, etc)

- Chrome (vs all other browsers)

Sure, i'm picking the obvious winners, but these are all market leaders now (Android by global share) where earlier incumbents were big, but not Google-big.

Even if Google's use of TPUs are purely self-serving, it will have a noticeable effect on their ability to scale their consumer AI usage at diminishing costs. Their ability to scale AI inference to meet "Google scale" demand, and do it cheaply (at least by industry standards), will make them formidable in the "ai race". This is why altman/microsoft and others are investing heavily in AI chips.

But I don't think their TPU will be only self-serving, rather, they'll scale it's use through GCP for enterprise customers to run AI. Microsoft is already tapping their enterprise customers for this new "product". But those kinds of customers will care more about cost than anything else.

The long-term game here is a cost game, and Google is very, very good at that and has a headstart on the chip side.


TPUs were originally intended to just be for internal use (to keep google from being dependent on Intel and nvidia). Making them an external product through cloud was a mistake (in my opinion). It was a huge drain on internal resources in many ways and few customers were truly using them in the optimal way. They also competed with google's own nvidia GPU offering in cloud.

The TPU hardware is great in a lot of ways and it allowed google to move quickly in ML research and product deployments, but I don't think it was ever a money-maker for cloud.


> The media keeps missing the real lock in Nvidia has: CUDA. It's not the hardware. It's the ability for someone to use it painlessly.

Really? What if someone writes a new back-end to PyTorch, TensorFlow and perhaps a few other popular libraries? Then will CUDA still matter that much?


if someone writes a new back-end to PyTorch

If that was easy to do surely AMD would have done it by now? After many years of trying?


I am starting to wonder if AMD were even trying all this time.


PyTorch has had an XLA backend for years. I don't know how performant it is though. https://pytorch.org/xla


It's pretty fast, just not as nice to use. You need statically defined tensors, and some functions are just not supported (last time I used it).


Can you do Unreal engine's Nanite, or Otoy Ray tracing in Pytorch?


TensorFlow and PyTorch support TPUs. It's pretty painless.


Having used it heavily it is nowhere near painless. Where can you get a TPU? To train models you basically need to use GCP services. There are multiple services that offer TPU support, Cloud AI Platform, GKE, and Vertex AI. For GPU you can have a machine and run any tf version you like. For tpu you need different nodes depending on tf version. Which tf versions are supported per GCP service is inconsistent. Some versions are supported on Cloud AI Platform but not Vertex AI and vice versa. I have had a lot of difficulty trying to upgrade to recent tf versions and discovering the inconsistent service support.

Additionally many operations that run on GPU but are just unsupported for TPU. Sparse tensors have pretty limited support and there's bunch of models that will crash on TPU and require refactoring. Sometimes pretty heavy thousands of lines refactoring.

edit: Pytorch is even worse. Pytorch does not implement efficient tpu device data loading and generally has poor performance no where comparable to tensorflow/jax numbers. I'm unaware of any pytorch benchmarks where tpu actually wins. For tensorflow/jax if you can get it running and your model suits tpu assumptions (so basic CNN) then yes it can be cost effective. For pytorch even simple cases tend to lose.


> TensorFlow and PyTorch support TPUs. It's pretty painless.

Unless you physically work next to the TPU hardware team, the torch support for TPUs is pretty brittle.


mojo language joins the chat: https://www.modular.com/max/mojo


Mojo is a closed source language that will never reach mainstream adoption among ML engineers and scientists.


> Mojo is a closed source language that will never reach mainstream adoption among ML engineers and scientists.

[Citation needed]

The creator, Chris Lattner, previously created LLVM, clang, and Swift. In each case he said these projects would be open sourced, and in each case they were. In each case they reached mainstream adoption in their respective target markets.

He's stated that Mojo will be open source.

If you're going to claim with great confidence that this language will have a different outcome to his previous ones, then you probably should have some strong evidence for that.


hmm the creator says (from his podcast with Lex Friedman when I listened to him) that they are open sourcing it, but that it is a project borne out of their private effort at their company and that it is still being used privately - so the aim is open sourcing it while taking community input and updating their private code to reflect the evolving design so that when they release it their internal lang and the open sourced lang will not diverge.

of course not ideal, but better than "open sourcing" it and refusing every request because it does not work for their codebase. worse than having it open source from the get go, of course.

assuming that day comes, does it have a competitor in the works? a python superset, compatible with python libs, but enables you to go bare metal to the point that it enables you to directly program GPUs and TPUs without CUDA or anything?

"never" means you believe it will never be open sourced, or a competitor will surpass it by the time it is open sourced. or that you believe the premise of the lang is flawed and we don't need such a thing. Which one is it?

Here is their github btw: https://github.com/modularml/mojo

From what I see, they have a pretty active community and there is demand for such a system.

The github says something similar:

>This repo is the beginning of our Mojo open source effort. We've started with Mojo code examples and documentation, and we'll add the Mojo standard library as soon as we get the necessary infrastructure in place. The challenge is that we use Mojo pervasively inside Modular and we need to make sure that community contributions can proceed smoothly with good build and testing tools that will allow this repo to become the source of truth (right now it is not). We'll progressively add the necessary components, such as continuous integration, build tools, and more source code over time.


yes there is a much more performant competitor that actually supports Nvidia GPUs [1] https://centml.ai/make-your-ml-models-run-faster-with-hidet/


...this has very little to do with mojo. mojo is not an nvidia accelerator for a couple ML frameworks.


And Nvidia does actually sell their hardware. Nobody will ever get their hands on one of these outside Google Cloud. It might as well not exist.


Well, sometimes they fall of the back of trucks I guess: https://www.ebay.com/itm/134540730431 Archive link: https://archive.ph/7dPFo


Doesn't really matter. Google's infra is all the client you need to continue pouring tens of billions into a project like this, bonus if others start using it more in the cloud, but they have so much use for accelerators across their own projects they aren't going to stop


What's painful about using TPUs?


in a sentence: google ai stuff is the vendor lockin of 2024 apple with the ecosystem value of 1994 apple.

c.f. https://news.ycombinator.com/item?id=39149854


So you trade one vendor lockin for another. Nothing lost.


Consider applying for YC's W25 batch! Applications are open till Nov 12.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: