Hacker Newsnew | past | comments | ask | show | jobs | submit | singiamtel's commentslogin

I would hope not. The copyrighted content seems to be the link rather than anything in the app.

Your example sounds like stopping notepad from rendering copyrighted content


From a technical perspective, you're absolutely correct.

From a regulatory perspective, it seems unlikely that most courts would appreciate the difference. In their mind - you run a website, and that website contains copyrighted content. Take it down.

You'd probably have to just blacklist the link in question to avoid a legal headache.


I've been finding it fascinating to see the different approaches to this by the big players. I would've definitely expected Wikipedia to have a blanket ban (even if hard to enforce), but it seems like they are allowed as long as you're not just generating the full article


Is it a grant? My understanding is that they're raising money as a startup

https://www.reuters.com/technology/artificial-intelligence/o...


I found this principle particularly interesting:

    Human oversight: The use of AI must always remain under human control. Its functioning and outputs must be consistently and critically assessed and validated by a human.


Interesting in what sense? Isn't it just stating something plainly obvious?


It is, but unfortunately the fact that to you - and me - it is obvious does not mean it is obvious to everybody.


Quite. One would hope, though, that it would be clear to prestigious scientific research organizations in particular, just like everything else related to source criticism and proper academic conduct.


Did you forget the entire DOGE episode where every government worker in the US had to send an weekly email to an LLM to justify their existence?


I'd hold CERN to a slightly higher standard than DOGE when it comes to what's considered plainly obvious.


Sure, but the way you maintain this standard is by codifying rules that are distinct from the "lower" practices you find elsewhere.

In other words, because of the huge DOGE clusterfuck demonstrated how horrible practices people will actually enact, you need to put this into the principles.


Oddly enough nowadays CERN is very much like a big corpo, yes they do science, but there is a huge overhead of corpo-like people who running CERN as an enterprise that should bring "income".


Can you elaborate on this, hopefully with details and sources including the revenue stream that CERN is getting as a cooperation?


I want to see how obvious this becomes when you start to add agents left and right that make decisions automagically...


Right. It should be obvious that at an organization like CERN you're not supposed to start adding autonomous agents left and right.


Where is “human oversight” in an automated workflow? I noticed the quote didn’t say “inputs”.

And with testing and other services, I guess human oversight can be reduced to _looking at the dials_ for the green and red lights?


Someone's inputs is someone else's outputs, I don't think you have spotted an interesting gap. Certainly just looking at the dials will do for monitoring functioning, but falls well short of validating the system performance.


The real interesting thing is how does that principle interplay with their pillars and goals i.e. if the goal is to "optimize workflow and resource usage" then having a human in the loop at all points might limit or fully erode this ambition. Obviously it not that black and white, certain tasks could be fully autonomous where others require human validation and you could be net positive - but - this challenge is not exclusive to CERN that's for sure.


Do they hold the CERN Roomba to the same standard? If it cleans the same section of carpet twice is someone going to have to do a review?


It's still just a platitude. Being somewhat critical is still giving some implicit trust. If you didn't give it any trust at all, you wouldn't use it at all! So they endorse trusting it is my read, exactly the opposite of what they appear to say!

It's funny how many official policies leave me thinking that it's a corporate cover-your-ass policy and if they really meant it they would have found a much stronger and plainer way to say it


"You can use AI but you are responsible for and must validate its output" is a completely reasonable and coherent policy. I'm sure they stated exactly what they intended to.


If you have a program that looks at CCTV footage and IDs animals that go by.. is a human supposed to validate every single output? How about if it's thousands of hours of footage?

I think parent comment is right. It's just a platitude for administrators to cover their backs and it doesn't hold to actual usecases


I don't see it so bleakly. Using your analogy, it would simply mean that if the program underperforms compared to humans and starts making a large amount of errors, the human who set up the pipeline will be held accountable. If the program is responsible for a critical task (ie the animal will be shot depending on the classification) then yes, a human should validate every output or be held accountable in case of a mistake.


I take an interest in plane crashes and human factors in digital systems. We understand that there's a very human aspect of complacency that is often read about in reports of true disasters, well after that complacency has crept deep into an organization.

When you put something on autopilot, you also massively accelerate your process of becoming complacent about it -- which is normal, it is the process of building trust.

When that trust is earned but not deserved, problems develop. Often the system affected by complacency drifts. Nobody is looking closely enough to notice the problems until they become proto-disasters. When the human finally is put back in control, it may be to discover that the equilibrium of the system is approaching catastrophe too rapidly for humans to catch up on the situation and intercede appropriately. It is for this reason that many aircraft accidents occur in the seconds and minutes following an autopilot cutoff. Similarly, every Tesla that ever slammed into the back of an ambulance on the back of the road was a) driven by an AI, b) that the driver had learned to trust, and c) the driver - though theoretically responsible - had become complacent.


Sure, but not every application has dramatic consequences such as plane or car crashes. I mean, we are talking about theoretical physics here.


Theoretical? I don't see any reason that complacency is fine in science. If it's a high school science project and you don't actually care at all about the results, sure.


Half-Life showed a plausible story of how high energy physics could have unforeseen consequences.


The problem is that the original statement is too black and white. We make tradeoffs based on costs and feasibility

"if the program underperforms compared to humans and starts making a large amount of errors, the human who set up the pipeline will be held accountable"

Like.. compared to one human? Or an army of a thousand humans tracking animals? There is no nuance at all. It's just unreasonable to make a blanket statement that humans always have to be accountable.

"If the program is responsible for a critical task .."

See how your statement has some nuance? and recognizes that some situations require more accountability and validation that others?


Exactly.

If some dogs chew up an important component, the CERN dog-catcher won't avoid responsibility just by saying "Well, the computer said there weren't any dogs inside the fence, so I believed it."

Instead, they should be taking proactive steps: testing and evaluating the AI, adding manual patrols, etc.


That doesn't follow. Say you write a proof for a something I request, I can then check that proof. That doesn't mean I don't derive any value from being given the proof. A lack of trust does not imply no use.


> So they endorse trusting it is my read, exactly the opposite of what they appear to say!

They endorse limited trust, not exactly a foreign concept to anyone who's taken a closer look at an older loaf of bread before cutting a slice to eat.


I think you're more reading what you want to read out of that - but that's the problem, it's too ambiguous to be useful


I'm also in Switzerland, currently my approach is to invest in Vanguard VOO (tracks the S&P500) via Interactive Brokers. There is a way to setup auto transfer and invest every month

As a caveat your money will be in dollars and in American companies, which might not be what you want, but it's worked for me well so far


Does this mean Copilot tab complete is banned too? What about asking an LLM for advice and then writing all the code yourself?


The language is actually:

> Any contribution of any LLM-generated content

I read this as "LLM-generated contributions" are not welcome, not "any contribution that used LLMs in any way".

More generally, this is clearly a rule to point to in order to end discussions with low effort net-negative contributors. I doubt it's going to be a problem for actually valuable contributions.


> Does this mean Copilot tab complete is banned too? What about asking an LLM for advice and then writing all the code yourself?

You're brushing up against some of the reasons why I am pretty sure policies like this will be futile. They may not diminish in popularity but they will be largely unenforceable. They may serve as an excuse for rejecting poor quality code or code that doesn't fit the existing conventions/patterns but did maintainers need a new reason to reject those PRs?

How does one show that no assistive technologies below some threshold were used?


> How does one show that no assistive technologies below some threshold were used?

In this case, you don't:

> immediately be banned without recourse

In other words, if the maintainer(s) think it's LLM-generated, right or wrong, you're banned.


Idk why anyone would contribute to a project with an attitude like this


Make useful, carefully reviewed contributions and you'll be fine.


that's fine, they probably don't want you then


I don't want them either. I'll find someone else who likes me the way I am. Plenty of fish in the pond.


That seems unlikely. Probably, what is going to happen, is if during a code review, you can't actually explain what your code is doing or why you wrote it, then you will be banned.

I don't know much about this project, but looking at the diff with their previous policy, it's pretty clear that people were abusing it and not declaring that they use llms, and they don't actually know what they're doing


Then a friendlier and clearer wording for the policy would work better. The current one says you will be "immediately" banned "without recourse" which implies that nothing like you are describing will happen.


Or arguably that's the point. If you Copilot generate a few lines of code or use it for inspiration you're still paying attention to it and are aware of what it's doing. The actual outcome will be indistinguishable from the code you hand wrote so it's fine. What policies like this do is stop someone generating whole pages at once, run it with minimal testing then chuck it into the code base forever.


I'm pretty sure the point is that anything clearly generated will result in an instant ban. That seems rather fair, you want contributors who only submit code they can fully understand and reason about.


The language clearly says "If you use an LLM [...] to make any kind of contribution".


[I was wrong and posted a link to an earlier policy/discussion overridden by the OP]


The part you are quoting is being removed. The policy used to state "If you contribute un-reviewed LLM generated...", now simply states "If you use an LLM to make any kind of contribution then you will immediately be banned without recourse."


you seem to be reading that backwards, that's the content that was removed. it now just says "if LLM, banned": https://github.com/twpayne/chezmoi/blob/master/.github/CODE_...


[I was wrong and posted a link to an earlier policy/discussion overridden by the OP]


That’s not what it says. It’s pretty clear…

> Any contribution of any LLM-generated content will be rejected and result in an immediate ban for the contributor, without recourse.

You can argue it’s unenforceable, unproductive, or a bad idea. But it says nothing about unreviewed code. Any LLM generated code.

I’m not sure how great of an idea it is, but then again, it’s not my project.

Personally, I’d rather read a story about how this came to be. Either the owner of the project really hates LLMs or someone submitted something stupid. Either would be a good read.


The linked thread includes the story


Said policy:

    * Any contribution of any LLM-generated content will be rejected and result in
      an immediate ban for the contributor, without recourse.


It's you who isn't reading the policy. What you are reading instead is deleted sections in a Git commit which are not part of the policy.


Did you read the policy? "If you use an LLM to make any kind of contribution..."


Tab completions by LLM are code generated by an LLM.


Not sure about this project in particular, but many more popular projects (curl comes to mind) have adopted similar policies not out of spite but because they'd get submerged by slop.

Sure, a smart guy with a tool can do so much more, but an idiot with a tool can ruin it for everyone.


Isn't it then more reasonable to have a policy that "people who submit low quality PRs will be banned"? Target the actual problem rather than an unreliable proxy of the problem.

LLM-generated code can be high quality just as human-generated code can be low quality.

Also, having a "no recourse" policy is a bit hostile to your community. There will no doubt be people who get flagged as using LLMs when they didn't and denying them even a chance to defend themselves is harsh.


Banning LLMs can result in shorter arguments. "Low quality" is overly subjective and will probably take a lot of time to argue about. And then the possible outrage if it is taken to social media.


> Banning LLMs can result in shorter argument

Can it really? "You submitted LLM-generated contributions" is also highly subjective. Arguably more so since you can't ever really be sure if somethingi s AI generated while with quality issues there are concrete things you can point to (e.g. it the code simply doesn't work, doesn't meet the contributor guidelines, uses obvious anti-patterns etc.).


If you rtfa, you will find it's actually the other way around. The linked PR from the AI has "concrete things you can point to" like "the code simply doesn't work".


Here's a post from Daniel Stenberg (curl maintainer) announcing that he just landed 22 LLM-generated commits.

https://mastodon.social/@bagder/115241241075258997

So obviously curl doesn't have a blanket ban.


I am wondering why you are posting this link, then asking this question to the HN community, instead of asking the project directly for more details. I does look like your intent is to stir some turmoil over the project position, and not to contribute constructively to the project.


That kind of point could be made for a large fraction of HN comments, but that aside: if a project’s policy is to ban for any LLM usage, without recourse, just asking a question about it could put you on a list of future suspects…


Good point, I'll include my question on the original discussion


It's the em dash that does it for me


AIs learned that from humans because it's a normal, common bit of punctuation they see frequently.

AIs also use the word "the" frequently.


Friendly reminder that em and en dashes were part of English well before ChatGPT was launched. Anecdotally, I’ve been using them forever and English isn’t even my native language.


Also, a lot of programs autocorrect dashes to em-dashes.


I use em dashes, but always as two hyphens.

I think this notion that em dash always means chatgpt is an overview correction.


I have loved the em dash forever and i’m being punished for it now.


The project seems nice, but how good is that domain name


I'm so glad it's getting some recognition!


Small bug report, you can shoot several times each email


Did the web server go down?


Yeah, sorry! Hug of death, redeploying w/ a big boy hosting provider now


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: