I suggest looking at llm arena leaderboards filtered by open weight models. It offers benchmarks at a very complete and statistically detailed level for models, and usually is quite up to date when new models come out. The new Gemma that just came out might be the best for 1x GPU, and if you have a bunch of vram check out the larger Chinese models
Selfhosted
A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.
Rules:
-
Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.
-
No spam posting.
-
Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.
-
Don't duplicate the full text of your blog or github here. Just post the link for folks to click.
-
Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).
-
No trolling.
-
No low-effort posts. This is subjective and will largely be determined by the community member reports.
Resources:
- selfh.st Newsletter and index of selfhosted software and apps
- awesome-selfhosted software
- awesome-sysadmin resources
- Self-Hosted Podcast from Jupiter Broadcasting
Any issues on the community? Report it using the report flag.
Questions? DM the mods!
I find Qwen3.5 is the best at toolcalling and agent use, otherwise Gemma4 is a very solid all-rounder and it should be the first you try. Tbh gpt-oss is still good to this day, are you running into any problems w it?
No problems per se. I just thought that I had not checked for an update for a longer time.
I'm not on there, but you might have more luck in !localllama@sh.itjust.works
You might also want to list the hardware that you plan to use, since that'll constrain what you can reasonably run.
I'd say Qwen 3.5 and Gemma 4 beat GPT OSS in every aspect.
The latest open weights model from google might be a good fit for you. The 26B model works pretty well on my machine, though the performance isn't great (6 tokens per second, CPU only).
How much VRAM?
I'm in the same boat. You'll get better responses if you post your machine specs. I
What are your computer specs?
I did just update my post with the specs. Maybe it takes a while to federate?
I must have not refreshed ignore my comment