this post was submitted on 23 Mar 2026
699 points (99.0% liked)

Technology

83102 readers
4982 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] infeeeee@lemmy.zip 413 points 3 days ago (8 children)

Saved you a click:

After much debate, the new policy is in effect: Wikipedia authors are not allowed to use LLMs for generating or rewriting article content. There are two primary exceptions, though.

First, editors can use LLMs to suggest refinements to their own writing, as long as the edits are checked for accuracy. In other words, it’s being treated like any other grammar checker or writing assistance tool. The policy says, “ LLMs can go beyond what you ask of them and change the meaning of the text such that it is not supported by the sources cited.”

The second exemption for LLMs is with translation assistance. Editors can use AI tools for the first pass at translating text, but they still need to be fluent enough in both languages to catch errors. As with regular writing refinements, anyone using LLMs also has to check that incorrect information hasn’t been injected.

[–] arcine@jlai.lu 3 points 1 day ago

Treating it like a tool instead of treating it like a God. What a novel idea !

[–] RIotingPacifist@lemmy.world 260 points 3 days ago (3 children)

AIbros: we're creating God!!!

AI users: it can do translation & reformating pretty well but you got to check it's not chatting shit

[–] halcyoncmdr@piefed.social 101 points 3 days ago (3 children)

The takeaway from all LLM-based AI is the user needs to be smart enough to do whatever they're asking anyway. All output needs to be verified before being used or relied upon.

The "AI" is just streamlining the process to save time.

Relying on it otherwise is stupid and just proves instantly that you are incompetent.

[–] Zagorath@quokk.au 12 points 3 days ago (2 children)

the user needs to be smart enough to do whatever they're asking anyway

I'm gonna say that's ideal but not quite necessary. What's needed is that the user is capable of properly verifying the output. Which anyone who could do it themselves definitely can, but it can be done more broadly. It's an easier skill to verify a result than it is to obtain that result. Think: how film critics don't necessarily need to be filmmakers, or the P=NP question in computer science.

[–] Pyro@programming.dev 16 points 3 days ago (4 children)

But if the output has issues, what're you going to do, prompt it again? If you are only able to verify but not do the task, you cannot correct the AI's mistakes yourself.

[–] Zagorath@quokk.au 9 points 3 days ago (2 children)

At the risk of sounding like an overly obsequious AI… You know what, you're completely right. I'm honestly not sure what use case I was imagining when I wrote that last comment.

You were thinking logically about a normal production chain. In that case, QA or whoever says "This is wrong, rework it and correct the issue" and that's that. With AI, it does the whole thing over again and may or may not come back with the same issue or an entirely new one.

[–] Redjard@reddthat.com 6 points 3 days ago

Making text flow naturally, grouping and ordeeing information, good writing.

You can verify two textst have the same facts and information, yet one reads way better than the other. But writing a text that reads well is quite hard.

I can't draw, but I could probably photoshop out some minor issues in an AI-generated image.

[–] fartographer@lemmy.world 1 points 2 days ago (1 children)

If you're unable to brute-force verification (research, testing, consulting the ancient texts), there's where you stop what you're doing, and take a breath. Then, consult an expert. Just like the film critic analogy, it's easier to verify than to create, so you're saving the expert time and effort while learning about something that you were obviously already passionate enough about to have started this endeavor.

[–] alsimoneau@lemmy.ca 2 points 1 day ago (1 children)

As someone who codes, it's not always easier to verify than to create.

[–] fartographer@lemmy.world 1 points 1 day ago

As someone who codes, I specifically didn't say "always" because of course it's not always true. Especially in the cases of "garbage in, garbage out."

But there's still an argument to be made for mental load and context, for which I'd argue that planning solutions and then writing the code generally is more taxing than someone handing you suggested solutions with semi-complete code or pseudo-code, and then identifying road blocks.

On the other hand, if someone you trust unexpectedly hands you hallucinated garbage, then you're likely to spin your wheels trying to identify what they did.

[–] Redjard@reddthat.com 1 points 3 days ago

If you don't habe the ability then you would do what you would have 5 years ago: not do it
Either submit without, or not submit at all.

[–] Aralakh@lemmy.ca 2 points 2 days ago

This is where domain expertise would come in, no? It's speeding up the work but it usually outputs generic content, and whatever else it injects while hallucinating. Therefore the validation part holds up I'd say.

[–] rumba@lemmy.zip 3 points 2 days ago

This is absolutely the case, and honestly, at least for now how it needs to be across the board.

Noone should be using AI to do things you're incapable of doing (or undoing).

[–] 7101334@lemmy.world 1 points 2 days ago

Relying on it otherwise is stupid and just proves instantly that you are incompetent.

Relying on it in any circumstances (though medical stuff is understandable if you're simply too poor or don't have access) while it is exhausting water supplies and polluting the planet is stupid and instantly proves that you are stupid and inconsiderate.

[–] youcantreadthis@quokk.au 10 points 3 days ago (1 children)

Fucking hate those anti human filth pushing slop into everything. I want to take one apart with power tools.

[–] Paranoidfactoid@lemmy.world 16 points 3 days ago (1 children)
[–] Scrollone@feddit.it 5 points 3 days ago (2 children)

Damn that movie was funny. I need to rewatch it.

Yaaah, but I'll need you to come in this weekend though. Yaaaahhhh....

[–] onlyhalfminotaur@lemmy.world 6 points 3 days ago

It holds up better than any movie from the late 90s that I can think of.

[–] XLE@piefed.social 5 points 3 days ago

I don't think AI users would say it does reformatting either (if they're honest): If you tell a chatbot to reformat text without changing it, it will change the text, because it does not understand the concept of not changing text. It should only take one time for someone to get burned for them to learn that lesson.

Seems pretty reasonable to use it as a grammar checker. As long as it's not changing content, just form or readability, that seems like a pretty decent use for it, at least with a purely educational resource like Wikipedia.

[–] Goodlucksil@lemmy.dbzer0.com 12 points 3 days ago

To save you another few clicks: this is the discussion (RfC) that implemented the changes, and the policy is linked at the top.

[–] ji59@hilariouschaos.com 22 points 3 days ago

So, it should be used reasonably, as it should have always been.

[–] daychilde@lemmy.world 18 points 3 days ago

Liar. I already read the article before opening the comments. YOU SAVED ME NOTHING.

;-)

[–] FauxPseudo@lemmy.world 0 points 3 days ago (1 children)

Seems like there should be a third exception. For those occasions where the article is about LLM generated text. They should be able to quote it when it's appropriate for an article.

[–] Zagorath@quokk.au 6 points 3 days ago (1 children)

That is a reasonable exception to no-AI policies in research papers and newspaper articles, but not for Wikipedia. As a tertiary source, Wikipedia has a strict "no original research" policy. Using AI to provide examples of AI output would be original research, and should not be done.

Quoting AI output shared in primary and secondary sources should be allowed for that reason, though.

[–] ricecake@sh.itjust.works 3 points 2 days ago

Eh, that's not quite original research. There are plenty of other examples of images and sound files created for Wikipedia. A representative example isn't research, it's just indicating what something is.

The Wikipedia article on AI slop and generative AI has a few instances of content that's representative to illustrate a sourced statement, as opposed to being evidence or something.

It's similar to the various charts and animations.

[–] errer@lemmy.world 0 points 3 days ago (3 children)

Wikipedia probably wants to sell access to LLMs to train. It’s only valuable if Wikipedia remains a high-quality, slop-free source.

I think even AI zealots think there should be silos of content to train from that are fully human generated. Training slop on slop makes the slop even worse.

[–] Grimy@lemmy.world 18 points 3 days ago (1 children)

Sell licenses of what? It's already all in the creative commons iirc.

[–] Zagorath@quokk.au 2 points 3 days ago (1 children)

The content is CC licensed, but they are trying to block AI scraping because it overloads their servers. They have a paid API that uses a lot less compute for both Wikipedia and the AI, as well as being a revenue source for Wikipedia.

[–] ricecake@sh.itjust.works 1 points 2 days ago

Yes, but...

https://en.wikipedia.org/wiki/Wikipedia%3ADatabase_download

That's because viewing the page uses server resources, as done API access. If you want the data you can download the database directly.

[–] SuspciousCarrot78@lemmy.world 13 points 3 days ago

AI already trains on Wikipedia.

https://commoncrawl.org/

This was only done because the editors pushed to minimize AI involvement. There's a comment here already mentioning that: https://lemmy.world/comment/22826863