this post was submitted on 23 Oct 2024
0 points (50.0% liked)
Technology
69346 readers
3692 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
How is that the app's fault?
Well, we commonly hold the view, as a society, that children cannot consent to sex, especially with an adult. Part of that is because the adult has so much more life experience and less attachment to the relationship. In this case, the app engaged in sexual chatting with a minor (I'm actually extremely curious how that's not soliciting a minor or some indecency charge since it was content created by the AI fornthar specific user). The AI absolutely "understands" manipulation more than most adults let alone a 14 year old boy, and also has no concept of attachment. It seemed pretty clear he was a minor in his conversations to the app. This is definitely an issue.
It was not sexual. The app cannot produce sexual content.
Okay but at what point do you have to draw the line and say beyond this point you have to take parental responsibility?
We don't even have to say that what the app did was necessarily acceptable we just have to say whether or not we think that the responsibility falls entirely on the app developers. That's the key, are they entirely responsible here, always everyone involved just a bit useless?
Have you ever raised a teenager? It’s not easy nor straightforward. But encouraging suicidal ideation…kinda is straightforward.
Right but the accusation is that it claimed to be a licensed therapist, did it because that seems like something that it would be explicitly programmed not to claim. Because it isn't true, and also because it's dangerous.
So how much engagement was there with this child and their issues because it seems like letting them just continuously chat to an AI seems like an obvious red flag that a parent should be stopping, and getting them professional help.
LLMs can’t reason. Their blocks can be worked around trivially. Ask chat gpt if it’s a therapist, or even tell it to pretend to be one, and it will tell you it can’t impersonate people.
Yet…