this post was submitted on 03 Apr 2026
23 points (62.9% liked)

Technology

83990 readers
3399 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 15 comments
sorted by: hot top controversial new old
[–] brucethemoose@lemmy.world 14 points 2 weeks ago* (last edited 2 weeks ago)

They seem to have held back the "big" locally runnable model.

It's also kinda conservative/old, architecture wise: 16-bit weights, sliding window attention interleaved with global attention. No MTP, no QAT (yet), no tightly integrated vision, no hybrid mamba like Qwen/Deepseek, nothing weird like that. It's especially glaring since we know Google is using an exotic architecture for Gemini, and has basically infinite resources for experimentation.

It also feels kinda "deep fried" like GPT-OSS to me, see: https://github.com/ikawrakow/ik_llama.cpp/issues/1572

it is acting crazy. it can't do anything without the proper chat template, or it goes crazy.


IMO it's not very interesting, especially with so many other models that run really well on desktops.

[–] brucethemoose@lemmy.world 13 points 2 weeks ago* (last edited 2 weeks ago) (1 children)

Also, for any interested, desktop inference and quantization is my autistic interest. Ask my anything.

I don't like Gemma 4 much so far, but if you want to try it anyway:


But TBH I'd point most people to Qwen 3.5/3.6 or Step 3.5 instead. They seem big, but being sparse MoEs, they can run quite quickly on single-GPU desktops: https://huggingface.co/models?other=ik_llama.cpp&sort=modified

[–] TrippinMallard@lemmy.ml 4 points 2 weeks ago (1 children)
[–] brucethemoose@lemmy.world 9 points 2 weeks ago* (last edited 2 weeks ago)

Ughhh, I could go on forever, but to keep it short:

Basically, the devs are Tech Bros. They're scammer-adjacent. I've been in local inference for years, and wouldn't touch ollama if you paid me to. I'd trust Gemini API over them any day.

I'd recommend base llama.cpp or ik_llama.cpp or kobold.cpp, but if you must use an "turnkey" and popular UI, LMStudio is way better.

But the problem is, if you want a performant local LLM, nothing about local inference is really turnkey. It's just too hardware sensitive, and moves too fast.

[–] madcaesar@lemmy.world 6 points 2 weeks ago (1 children)

What could this be used for?

[–] baatliwala@lemmy.world 10 points 2 weeks ago* (last edited 2 weeks ago) (1 children)

Local LLMs, probably even ones you can host on phones. But they won't be as powered of course

[–] madcaesar@lemmy.world 4 points 2 weeks ago (3 children)

Yea I get that, but does anyone have any practical ideas for local LLM?

[–] Imgonnatrythis@sh.itjust.works 17 points 2 weeks ago (1 children)

Literature summarization, data analysis, not being a pawn in corporate data harvesting.

[–] XLE@piefed.social 0 points 2 weeks ago

As long as you don't care if the summaries and analyses are wrong!

[–] felsiq@piefed.zip 10 points 2 weeks ago (1 children)

Home assistant is the big one imo, voice control for a private smart home is useful and low-stakes so hallucinations won’t be the end of the world

[–] leftascenter@jlai.lu 4 points 2 weeks ago

I'm eagerly waiting for a locally run phone assistant. Just for voice control while driving.

[–] baatliwala@lemmy.world 2 points 2 weeks ago* (last edited 2 weeks ago)

In addition to what the others said, some apps allow you to link to an LLM model for additional features.

For eg Immich has prebuilt models you can choose depending on how powerful your PC is, which will give facial recognition and powerful NLP-like search capabilities for your library. So if they think this is model good they can make a new prebuilt one using this as a base. Software like Microsoft Teams uses LLM for better background blurring for video calls, so maybe an open source equivalent can make use of it.

Also you can use it for other stuff like image generation too