this post was submitted on 11 Mar 2026
80 points (95.5% liked)
Technology
82518 readers
4556 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments

This tech was never ready for release.
Here's what's going to happen: this will make the rounds, it'll get added to the fine tune dataset, and all the big AI companies will pretend it's all good.
The issue however is that these questions will be patched, but not the intent, or the latent spaces in the models, or the training data.
Exactly. They won't actually change the models because they don't understand the relationship between the input and output enough to actually target responses like this. So what they will do is add an administrative filter layer on top, but it will always be something can work around because that is the nature of that kind of filter. The whole engine is still accessibile.
That's what regular people never seem to understand (and the AI apologists are hoping you don't know). These models aren't "getting better," they're just filled with more reactive patches over these unintended responses. And as the models scale up, so do the holes that need patching.
It's a never ending game of bad-prompt Whack-a-Mole, all at the cost of our environment and safety, just so the Tech Bros can try to convince venture capitalists that "AGI is definitely just around the corner, trust me, bro," and keep that bubble filled with their own farts.
And the only "improvement" they can do is to manually filter responses and program rote responses to certain specific prompts. Which amounts to actually reducing the amount of LLM that reaches the surface. They are actually reverse engineering these things into more primitive chat bots with algorithmic responses, except that they cost trillions of dollars and require massive amounts of energy to run.
Its like deciding that a Ferrari is not suitable for commuting, so instead of actually building a different car, they just fill the trunk with sand and drag a trailer behind around to slow it down.