this post was submitted on 28 Mar 2026
162 points (90.1% liked)

Technology

83990 readers
3261 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] XLE@piefed.social -1 points 3 weeks ago (13 children)

Where do you think the "new ones" are coming from?

[–] riskable@programming.dev 5 points 3 weeks ago (12 children)

Same places as usual: Academia and open source foundations.

That's where 99% of all advancements in AI come from. You don't actually think Big AI is paying as many people to do computer science and mathematics research as all the universities in the world (with computer science programs)?

It's the same shit as always: Big companies commercialize advancements and discoveries made by scientist and researchers from academia (mostly) and give almost nothing back.

Big AI has partnerships with tons of schools and if it weren't for that, they wouldn't be advancing the technology as fast as they are. In fact, the only reason why many of these discoveries are made public at all is because of the agreements with the schools that require the discoveries/papers be published (so their school, professors, researchers, and students can get credit).

Like I was saying before: You don't need a trillion dollars in data centers to do this stuff. Almost all the GPUs and special chips being used (and preordered, sigh) by Big AI are being used to serve their customers (at great expense). Not for training.

Training used to be expensive but so many advancements have been made this is no longer the case. Instead, most of the resources being used in "AI data centers" (and research) is all about making inference more efficient. That's the step that comes after you give an AI a prompt.

Training a super modern AI model can be done with a university's data center or a few hundred thousand to a few million dollars of rented GPUs/compute. It doesn't even take that long!

Generative AI improves at a ridiculously fast rate. In nearly all the ways you could think of: Training, inference (e.g. figuring out user intent), knowledge, understanding, and weirder, fluffier stuff like "creativity" (the benchmarks of which are dubious, BTW).

[–] XLE@piefed.social -2 points 3 weeks ago* (last edited 3 weeks ago) (11 children)

Before we spin into a tangent about theory and "what ifs" etc, care to link me to all these great models from academics and open-source institutions?

Because right now, the only companies I see making advancements in "AI" are burning through obscene amounts of cash, with no end in sight.

And there is no evidence the cost of inference is going down, and even Anthropic admits training will continue burning resources.

[–] XLE@piefed.social 1 points 3 weeks ago* (last edited 3 weeks ago) (2 children)

@ikidd@lemmy.world @ ingeanus@ttrpg.network do you two have a source for these supposed great models?

[–] riskable@programming.dev 4 points 3 weeks ago

I personally love glm-5 and qwen3.5, specifically: https://ollama.com/library/qwen3.5:122b

I've used them both for coding and they work really well (way better than you'd think). They're also perfectly capable of the usual LLM chat stuff (e.g. check my grammar) but all the models (even older, smaller ones) are capable of that stuff these days.

For a treat: Have someone show you using some of these models to search the web! It's amazing. You don't see ads, you don't have to comb through 12 pages of search results, and they read the pages that moment (not cached) to give you summaries of the content. So when you click the link to go to the content you know it's the thing you were looking for. They're not using a local index of the Internet, they're searching on your behalf using whatever search engines you configured. It's waaaaay better than ChatGPT (which uses Bing behind the scenes whether you like it or not) or Gemini (which uses Google, obviously). The (self-hosted) LLM will literally be running curl for you on Google, DuckDuckGo, Bing, or whatever TF else you want (simultaneously) then reading each of the search results and using your prompt to figure out what the most relevant results are. It's sooooo nice!

FYI: Ollama.com's library page is actually a great resource for finding info on all the models that can be self-hosted: https://ollama.com/library

[–] ikidd@lemmy.world 2 points 3 weeks ago (1 children)

What inclined you to @ me into this? As far as I can see, I haven't even replied in this thread, and you just seem like you're on the warpath with anyone that wants to defend using LLMs. If Greg KH thinks it's coming into it's own, you might want to heed him.

[–] XLE@piefed.social -1 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

Greg who? Does he have a basis for his beliefs?

[–] ikidd@lemmy.world 2 points 3 weeks ago (1 children)

Greg Kroah-Hartman (KH) or are you just being obtuse? About whom this article is written and not exactly some rando on the internet like you and me.

[–] XLE@piefed.social 0 points 3 weeks ago

I thought you were talking about the thread you were looking at, not the article. So these models don't exist? Pity.

load more comments (8 replies)
load more comments (8 replies)
load more comments (8 replies)