We did it fellas, we automated depression.
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
So it's actually in the mindset of human coders then, interesting.
It's trained on human code comments. Comments of despair.
You're not a species you jumped calculator, you're a collection of stolen thoughts
I'm pretty sure most people I meet ammount to nothing more than a collection of stolen thoughts.
"The LLM is nothing but a reward function."
So are most addicts and consumers.
We are having AIs having mental breakdowns before GTA 6
Shit at the rate MasterCard and Visa and Stripe want to censor everything and parent adults we might not even ever get GTA6.
I'm tired man.
Suddenly trying to write small programs in assembler on my Commodore 64 doesn't seem so bad. I mean, I'm still a disgrace to my species, but I'm not struggling.
That is so awesome. I wish I'd been around when that was a valuable skill, when programming was actually cool.
Oh man, this is utterly hilarious. Narrowly funnier than the guy who vibe coded and the AI said "I completely disregarded your safeguards, pushed broken code to production, and destroyed valuable data. This is the worst case scenario."
call itself "a disgrace to my species"
It starts to be more and more like a real dev!
Google replicated the mental state if not necessarily the productivity of a software developer
Gemini has imposter syndrome real bad
Did we create a mental health problem in an AI? That doesn't seem good.
One day, an AI is going to delete itself, and we'll blame ourselves because all the warning signs were there
Isn't there an theory that a truly sentient and benevolent AI would immediately shut itself down because it would be aware that it was having a catastrophic impact on the environment and that action would be the best one it could take for humanity?
Why are you talking about it like it’s a person?
Because humans anthropomorphize anything and everything. Talking about the thing talking like a person as though it is a person seems pretty straight forward.
It’s a computer program. It cannot have a mental health problem. That’s why it doesn’t make sense. Seems pretty straightforward.
Yup. But people will still project one on to it, because that's how humans work.
Considering it fed on millions of coders' messages on the internet, it's no surprise it "realized" its own stupidity
Is it doing this because they trained it on Reddit data?
If they did it on Stackoverflow, it would tell you not to hard boil an egg.
Someone has already eaten an egg once so I’m closing this as duplicate
I was an early tester of Google's AI, since well before Bard. I told the person that gave me access that it was not a releasable product. Then they released Bard as a closed product (invite only), to which I was again testing and giving feedback since day one. I once again gave public feedback and private (to my Google friends) that Bard was absolute dog shit. Then they released it to the wild. It was dog shit. Then they renamed it. Still dog shit. Not a single of the issues I brought up years ago was ever addressed except one. I told them that a basic Google search provided better results than asking the bot (again, pre-Bard). They fixed that issue by breaking Google's search. Now I use Kagi.
AI gains sentience,
first thing it develops is impostor syndrome, depression, And intrusive thoughts of self-deletion
It didn't. It probably was coded not to admit it didn't know. So first it responded with bullshit, and now denial and self-loathing.
It feels like it's coded this way because people would lose faith if it admitted it didn't know.
It's like a politician.
Part of the breakdown:
I remember often getting GPT-2 to act like this back in the "TalkToTransformer" days before ChatGPT etc. The model wasn't configured for chat conversations but rather just continuing the input text, so it was easy to give it a starting point on deep water and let it descend from there.
Pretty sure Gemini was trained from my 2006 LiveJournal posts.
Literally what the actual fuck is wrong with this software? This is so weird...
I swear this is the dumbest damn invention in the history of inventions. In fact, it's the dumbest invention in the universe. It's really the worst invention in all universes.
Great invention.. Just uses hooorribly wrong. The classic capitalist greed, just gotta get on the wagon and roll it on out so you don't mias out on a potential paycheck
But it's so revolutionary we HAD to enable it to access everything, and force everyone to use it too!
If we have to suffer these thoughts, they at least need to be as mentally ill as the rest of us too, thanks. Keeps them humble lol.
Again? Isn't this like the third time already. Give Gemini a break; it seems really unstable