this post was submitted on 08 Jan 2026
601 points (99.7% liked)

Technology

78511 readers
3110 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] ramble81@lemmy.zip 12 points 3 days ago (5 children)

Serious question here. LLMs trained their data off SO. Developers now ask LLMs for solutions instead of SO. New technology comes out that LLMs don’t have indexed. Where will LLMs get their data to train on for new technologies? You can’t exactly feed it a manual and expect it to extrapolate or understand (for that matter “what manual).

[–] dantheclamman@lemmy.world 1 points 2 days ago

I am worried, because there are increasing cases where open source docs are going offline because they can't take the bandwidth costs of the big LLM bots recrawling hundreds of times per day. Wikipedia is also getting hammered. There is so much waste and diminishing returns

[–] Prox@lemmy.world 8 points 3 days ago* (last edited 3 days ago) (1 children)

Yes, that is the major problem with LLMs in general. There is no solution aside from "train on another different source (like Reddit)", but then we rinse & repeat.

[–] village604@adultswim.fan 1 points 3 days ago (1 children)
[–] Prox@lemmy.world 2 points 3 days ago (1 children)

I guess, though I'm pretty ignorant as to how RLVR would fix the issue that arises from new coding languages or even new major versions. I'm not sure how LLMs would ever get to a correct answer if they don't have good reference material to start from or reference.

[–] General_Effort@lemmy.world 1 points 3 days ago

The assumption seems to be that an LLM can't figure out a manual or source code. If it can't, then you have to pay people. But that's not a universally valid assumption.

[–] xthexder@l.sw0.com 5 points 3 days ago* (last edited 3 days ago)
[–] General_Effort@lemmy.world 2 points 3 days ago

You can’t exactly feed it a manual and expect it to extrapolate or understand (for that matter “what manual).

You can do that to a degree (RLVR). They are also paying human experts. But that's the situation now. Who knows how it will be in a couple more years. Maybe training AIs will be like writing a library, framework, ...

[–] ThomasWilliams@lemmy.world 0 points 2 days ago

From the questions people ask and from online accounts.