Live by the sword π‘οΈ, die by the sword
Technology
Which posts fit here?
Any news that are at least tangentially connected to the technology, social media platforms, informational technologies or tech policy.
Post guidelines
[Opinion] prefix
Opinion (op-ed) articles must use [Opinion] prefix before the title.
Rules
1. English only
Title and associated content has to be in English.
2. Use original link
Post URL should be the original link to the article (even if paywalled) and archived copies left in the body. It allows avoiding duplicate posts when cross-posting.
3. Respectful communication
All communication has to be respectful of differing opinions, viewpoints, and experiences.
4. Inclusivity
Everyone is welcome here regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, education, socio-economic status, nationality, personal appearance, race, caste, color, religion, or sexual identity and orientation.
5. Ad hominem attacks
Any kind of personal attacks are expressly forbidden. If you can't argue your position without attacking a person's character, you already lost the argument.
6. Off-topic tangents
Stay on topic. Keep it relevant.
7. Instance rules may apply
If something is not covered by community rules, but are against lemmy.zip instance rules, they will be enforced.
Companion communities
!globalnews@lemmy.zip
!interestingshare@lemmy.zip
Icon attribution | Banner attribution
If someone is interested in moderating this community, message @brikox@lemmy.zip.
You should read the guy's explanation. It legitimately seems like an honest mistake, but given his work was specifically geared towards avoiding a situation like that I'm not surprised he was let go.
The guy had several explanations rolled into one so it seems more like a dishonest lie than an honest mistake. The guy the article was about had a decent explanation of how it happened though. His blog has AI scraping protections enabled, so when the so called journalist asked an AI to write an article for him citing the blog post, the AI couldn't access it, and did what AI do, made shit up.
His blog has AI scraping protections enabled,
Tell me more.
https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me-part-2/
It's all in here.
That's just one explanation among many. A more reasonable guess is that the Ars writer went to his webpage, then asked an AI extension, which would have total access to an open tabs, to pull out quotes or something similar. LLMs find it hard to not change text, even when instructed to.
There are more egregious examples of the author overestimating AI on the same blog post...
lmao, didn't know the 'ai' tool is that stupid to not handle website blocks/exceptions..
It's not quite like that. The tools used to scrape the web for training data couldn't access the site to scrape the data, so it's not encoded in the model.
The query interface for the model just hallucinates when there's a 'vacuum'.
i was thinking of some "automated browser" type, like if the browser returns an error page saying it's blocked, the LLM would get the "blocked from website"ish error as the page content, and shouldn't it say something around "I'm sorry, I couldn't access the website" instead of "Sure! Here's a summary of that webpage" followed by hallucinated bs. well maybe that's not the case here?
It doesn't say something like that specifically because it isn't an algorithm that receives x input and spits out Y. It's an algorithm that receives x query and spits out the most common variant word that comes after "query" . If there isn't a most common word that makes sense to a human, the AI doesn't know that and so it still gives the most common word in its training set.
If the query is "Juicy" it may output "melons" . If "melons" were not available in its training set it might output "grapes" or "cherries" , but if those weren't available it might output "apple bottom jeans" which would have made sense in 2003 but likely wouldn't make sense to the average kid today who's never heard of juicy couture.
It doesn't understand anything. It can't reason.
Ars saw his old explanation and conducted an internal review that took weeks, and found it lacking. His new explanation is... He isn't giving one, after being asked for comment.
I think he's more a libality at this point.
Not caring for if something is true ("reckless regard for the truth"), opens up libel lawsuits.
You cant just publish made up quotes on reporting a virtual hit piece on someone's reputation.