There are two interesting "safety nets" at play here: the classic "Nobody ever got fired for choosing IBM" principle (where ubiquity creates an implicit guarantee of continuity), and a less visible but equally powerful one where certain open source tools become such fundamental infrastructure that they're essentially "too critical to fail." Think curl, gpg, or apt - most users never directly interact with these, but they're so deeply embedded in the internet's fabric that the ecosystem ensures their maintenance. One heuristic I've found helpful is looking at major corporate adoption patterns - it can be a decent signal for identifying which tools fall into these categories.
I highly recommend js13k (https://js13kgames.com), an annual game jam where your entire game must fit in 13 kilobytes. The tight size limit forces you to get creative with optimization - think procedural generation, custom minimal engines, and code golf techniques that actually matter. While the games might be "useless" in a practical sense, you'll learn more about low-level JavaScript optimization than from most serious projects, all while being part of an incredibly supportive community that loves sharing tricks and tips.
js13k is a game jam competition. Make a game that fits in a 13 kilobyte zip.
"The thirteenth, anniversary edition of the online js13kGames competition starts… NOW! Build a Web game on a given theme within the next month and fit it into a 13 kilobyte zip package to win lots of cool prizes, eternal fame, and respect!"
Medplum (YC S22) is an open source, API first, healthcare developer platform. "Headless EHR", we take care of the security, compliance, and regulatory burdens of healthcare software development. Well funded and growing fast.
We're hiring an amazing Dev-Ex / Dev-Rel engineer to delight customers, build sample apps, and promote the Medplum platform.
“Training costs scale with the number of researchers, inference costs scale with the number of users”.
This is interesting, but I think I disagree? I'm most excited about a future where personalized models are continuously training on my own private data.
How can you disagree with that statement? Training takes significantly more processing power than inference, and typically only the researchers will be doing the training, so it makes sense that training costs scale with the number of researchers, as each researcher needs access to their own system powerful enough to perform training.
Inference costs scaling with the number of users is a no-brainer.
I'm pretty dumbfounded how you can just dismiss both statements without giving any reasoning as to why.
EDIT:
> I'm most excited about a future where personalized models are continuously training on my own private data.
Non-technical people will not be fine-tuning models. A service targeted at the masses is unlikely to fine-tune a per-user model. It wouldn't scale without being astronomically expensive.
We will need at least one- if not several- research and data capture breakthroughs to get to that point. One person just doesn't create enough data to effectively train models with our current techniques, no matter what kind of silicon you have. It might be possible, but research and data breakthroughs are much harder to predict than chip and software developer ergonomics improvements. Sometimes the research breakthroughs just never happen.
Nothing out of the ordinary for individual government departments to turn to private contractors when the GSA doesn't offer them a service they need when they need it.
GSA has since developed login.gov, but there hasn't been a mandate that other agencies have to use it over third-parties.
There was another one (census, maybe? can't recall which agency it was) using a .gd for a while, too... don't see it on the list anymore. Not sure who signed off on putting government services behind the "control" of a country we've invaded before.
I think their comment was to the contrary, that the copyright/legal implications of 'stolen' code could seriously hobble the wider development, proliferation, adoption, and commercialization of AI software.
Maybe I misunderstood, but the comment seemed to dismiss copyright issues as a cheap way to kill AI ("soft underbelly"). I think stealing code is a pretty serious deal and the onus is on AI software companies to make sure they aren't doing it; it's not "slowing the development of AI" to keep them accountable.
Are you seriously arguing that using short snippets open source code to inspire similar, yet not exactly the same, original code is "stealing code"? Human developers do that all day long. And just because a piece of code exists in a GPL project doesn't mean it originated there. Every algorithm or sort function likely originated in a more permissively licensed project before it got included in a GPL project.
What happens if I (a human) read GPL code and then reuse the knowledge gained from it in my own commercial projects? It's not as clear cut as you make it sound.
It could be as clear-cut as you've just made it: "a human". An LLM is not a human.
You could get into the semantics of "learning" - does JPEG encoding count as the computer "learning" how to reproduce the original image? But trying to create some metric for why LLMs "learn" and JPEG doesn't "learn" on the basis of the algorithms is a philosophical endeavor. Copyright is more about practicality - about realized externalities - than it is about philosophy. That's why selling cars and selling guns are regulated differently, despite the fact that you could reduce both to "metal mechanical machines that kill" by rhetorical argument.
Even from a strictly legal perspective, it actually is fairly clear-cut. The answer to "what if I (a human) read GPL code and then reuse the knowledge gained from it..." comes down to a few straightforward properties of the license. GPL doesn't cover "reduced to practice" as many corporate contracts do, so terms covering "the knowledge gained" are lenient. GPL covers "verbatim" copies which is what LLMs are doing, that's as clear cut as it gets. Inb4: "So what if I add a few spaces here and there?" - well, GPL also covers "a work based on"; this is where I (who am not a lawyer) can't speak confidently, but surely there are legal differences between "based on" and "reduced to practice", considering that both are very common occurrences in contracts, so there actually would be a lot of precedent.
I agree with you that verbatim copies are obviously covered by copyright. What if LLMS reproduce code with changed variable and function names (which would be a great improvement to `cs_gaxpy` in the original article)? What if just the general structure of an algorithm is used? What if the LLM translates the C algorithm from the original article into Rust? This discussion is only scratching the surface.
It's going to be an uphill battle just to get people to even understand what the problems are. And this is even a technical forum. Now imagine trying to explain these nuances to a judge or jury.
It's not so much an ability to understand as it is a desire to not understand in order to be able to ignore the rightsholders' licensing terms.
Plenty of tech companies exist by putting a thin layer on top of the hard work of others and if those others can be ignored then that's what they'll do.
4. Defense of Third Party Claims.
If your Agreement provides for the defense of third party claims, that provision will apply to your use of GitHub
Copilot. Notwithstanding any other language in your Agreement, any GitHub defense obligations related to your
use of GitHub Copilot do not apply if (i) the claim is based on Code that differs from a Suggestion provided by
GitHub Copilot, or (ii) you have not enabled all filtering features available in GitHub Copilot.
> If your Agreement provides for the defense of third party claims
do any of them?
it also states:
> You retain all responsibility for Your Code, including Suggestions you include in Your Code or reference to develop
Your Code. It is entirely your decision whether to use Suggestions generated by GitHub Copilot. If you use
Suggestions, GitHub strongly recommends that you have reasonable policies and practices in place designed to
prevent the use of a Suggestion in a way that may violate the rights of others. This includes, but is not limited to,
using all filtering features available in GitHub Copilot.
I think it's pretty clear. If you're not filtering, you're liable. If you are and something transpires, they'll fight your legal battle for you which is probably better than any monetary indemnity clause. I assume this is for enterprise users where it actually matters.
Looks like this code in https://github.com/ChRis6/circuit-simulation/blob/2e45c7db01... is older then the GPL code in question or provided by the example. Uh oh, did we discover something? Who actually owns this code because this code predates the code in question by a calendar year using git blame, by a different author, and with no license attached to the oldest code. Is it possible the code in the codeium.com example is relicensed and not GPL code at all?