this post was submitted on 03 Dec 2025
387 points (98.3% liked)

Technology

77096 readers
3824 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] NotMyOldRedditName@lemmy.world 1 points 3 hours ago* (last edited 3 hours ago) (1 children)

rely on input data, which is running out.

Thats part of the equation, but there is still a lot of work that can be done to optimize the usage of the llms themselves, and the more optimized and refined they are, the cheaper it becomes to run, and you can also use even bigger datasets that weren't feasible before.

I think there's also a lot of room to still optimize the data in the data set. Ingesting the entire worlds information doesn't lead to the best output, especially if you're going into something more factual vs creative like a LLM trained to assist with programming in a specific language.

And people ARE paying for it today, OpenAI has billions in revenue, the problem is the hardware is so expensive, the data centeres needed to run it are also expensive. They need to continue optimizing things to narrow that gap. Open AI charges $20 USD/month for their base paid plan. They have millions of paying customers, but millions isn't enough to offset their costs.

So they can

  1. reduce costs so millions is enough
  2. make it more useful so they can gain more users.

This is so early that they have room to both improve 1 and 2.

But like I said, they (and others like them) need to figure that out before they run out of money and everything falls apart and needs to be built back up in a more sustainable way.

We won't know if they can or can't until they do it, or it pops.

[–] pahlimur@lemmy.world 3 points 2 hours ago* (last edited 2 hours ago) (1 children)

None of this is true.

I've worked on data centers monitoring power consumption, we need to stop calling LLM power sinks the same thing as data centers. Its basically whitewashing the power sucking environmental disasters that they are.

Machine learning is what you are describing. LLMs being puppeted as AI is destructive marketing and nothing more.

LLMs are somewhat useful at dumb tasks and they do a pretty dumb job at it. They feel like when I was new at my job and for decades could produce mediocre bullshit, but I was too naive to know it sucked. You can't see how much they suck yet because you lack experience in the areas you use them in.

Your two cost saving points are pulled from nowhere just like how LLM inference works.