this post was submitted on 07 Oct 2025
255 points (92.9% liked)

Technology

75792 readers
2941 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] MiDaBa@lemmy.ml 6 points 1 day ago (4 children)

Ai has been trained on current and past writing which could be considered plagiarism depending on if you're asking an Ai CEO or not. My question is, what happens when most writing is done by Ai? Do they continue to train it but now on itself? Will the language models experience deterioration at that point?

[–] nightlily@leminal.space 16 points 1 day ago

That’s part of the reason these models haven’t improved much in the last year or so. They‘ve absorbed all the public facing internet and whatever copyrighted works they could get away with pirating (pretty much all printed work), and now they are faced with a brick wall. They haven’t come up with a way to create new content, to reinforce a „correct“ statistical model without causing model collapse, and I don’t think they ever will. The well (the public internet) is already thoroughly poisoned so they have to use a snapshot of the pre-LLM internet, not even an up to date one.

If it isn’t good enough after consuming almost the entirety of humanity’s written output since the invention of the printing press, it’s never going to be.

This is actually a problem a lot of people are working on, they used to call the resulting failure 'model collapse'. Training AI on existing slop does tend to deteriorate and is overall a bad time for AI.

[–] luxyr42@lemmy.dormedas.com 4 points 1 day ago (1 children)

Even discounting the writing quality, we already have AI responses that reference AI hallucinations posted online as fact.

[–] Ilovethebomb@sh.itjust.works 1 points 1 day ago

Also Reddit shitposts, there's some notable examples of that.

[–] Jhex@lemmy.world 2 points 1 day ago

it becomes even dumber... we are already there