There are some packages that aren't compatible, that are alternatives that do the same thing, so apt won't let you install them at the same time.
tal
Speaking of debian - anyone here running debian testing as a daily driver?
I do, though usually not early in the release cycle. Normally I use stable, then pull the occasional thing in from testing if I want it and it's not in stable, and switch to testing if there's a bunch of stuff I want.
Right now, I'm on stable because trixie just went stable.
No, because there's no reliable way to distinguish AI-generated spam sites from non-AI-generated spam sites. I'll also add that I don't expect there to be one promptly forthcoming: any attempt to identify them is going to run into improved systems, and that's gonna happen even if the systems aren't explicitly intending to evade detection. If it were easy, Google would have done so years back. I can recognize some now, but the SEO spam crowd that's creating this is trying hard to pollute search engine results, and if someone implements a generalized "block" that's effective, they're going to keep looking for alternatives until they find something that gets through.
On Kagi, I can set the acceptable date range on results to prior to the emergence of LLMs, but that cuts out a lot of material that I want to see. For some searches, that might work, but it's not really a general solution.
You can manually blacklist or deprioritize sites on Kagi. Probably can either run some sort of local proxy or Greasemonkey-style plugin that would let you do so in browser on any search engine. Problem is that there are people making these sites faster than you're going to be banning them.
Kagi's also got a "pin" and a "raised priority" feature for a list of sites, and I suppose could whitelist some "known good" sites. Kagi's "blacklist/deprioritize/prioritize/pin" feature does not have the ability to exchange sites between users (and I imagine that there'd be some privacy issues with doing so) aside from Kagi running a "leaderboard" of the most-blacklisted/deprioritized/prioritized/pinned sites. One could probably do the "proxy" or "plugin" route as well for a variety of websites on other search engines. Any general solution would need to have some level of interchange, since requiring every individual user to maintain a "killfile" on websites is going to be impractical. It may be that the human labor involved in curation is outweighed by how cheap it is to generate new websites; not sure.
At some point, I assume that it may become practical to just make a conservative whitelist of "non-spam" sites that accepts that many useful websites will be excluded because we just can't validate them as not being non-spam. Probably require human curation, which is either going to need volunteer labor or a commercial service.
There's also a secondary problem that if you curate content at the domain level, Web 2.0 sites that permit posting content (Reddit, Wikipedia, the Threadiverse, etc) can have individual users inserting AI-generated spam. So a general solution is probably going to need to permit some sort of sub-domain level filtering for at least major sites.
And there's also the wrinkle that a "trusted good" site or user can become a spammer at some point. Spammers/people who want to run influence operations have been buying high-karma Reddit accounts
and the reputation that comes with them
for quite some years. Domains expire, or their operators change. Reputation has value, and it can be sold. So that also has to be addressed.
This isn't really a qualitative change. I mean, people have hand-crafted spam websites that try to grab searchers before. It's just that the ability to use a computer to do it is way more cost-efficient, brings the cost way down, and thus opens up a lot of opportunity for spam that wouldn't have made sense financially before. So what you're really aiming to do is to get the cost to make a spam website up. One possibility
which I am absolutely confident that TLS certificate issuers would like
would be to have tiers of TLS certificate, some of which are a lot more expensive. Search engine indexers could check and validate the TLS "cost tier" when indexing a site. That will artificially inflate the cost of running a website, and can be done to an arbitrary degree. That's not fantastic, since it also tends to cut out non-spam individual/low-cost websites, but if you're a large company somewhere, the price is basically a rounding error compared to what a spammer needs to make to make his super-cheap-to-generate LLM-generated website worthwhile. Could be a component in a system that takes into account other factors.
You can edit your post title. Might need to undelete the post first, can't recall.
I don't know if pict-rs, the image-hosting thing, has delete functionality, though unless someone is an instance admin or has a saved copy of the URL, I don't know if there's any way to get a list of images.
Reddit when you delete a comment or post it gets deleted in full and does not leave any traces and is not linked to your account.
There have been various services that monitor and back up comments and let people see deleted Reddit comments.
EDIT: Ah, you mentioned that.
Yeah, but at least then you know that you've done that, because it'll display it on your screen.
I’ll just avoid sites and services that require this.
It's not the site wanting to do it. It's a requirement from Parliament in the UK.
https://en.wikipedia.org/wiki/Online_Safety_Act_2023
The Online Safety Act 2023[1][2][3] (c. 50) is an Act of the Parliament of the United Kingdom to regulate online content. It was passed on 26 October 2023 and gives the relevant Secretary of State the power to designate, suppress, and record a wide range of online content that is deemed "illegal" or "harmful to children".[4][5]
The Act creates a new duty of care for online platforms, requiring them to take action against illegal content, or legal content that could be "harmful" to children where children are likely to access it. Platforms failing this duty would be liable to fines of up to £18 million or 10% of their annual turnover, whichever is higher. It also empowers Ofcom to block access to particular websites. It obliges large social media platforms not to remove, and to preserve access to, journalistic or "democratically important" content such as user comments on political parties and issues.
I mean, you don't have to go there if you want, but it's probably Parliament you want to be irritated with, not websites following British laws.
https://pilotinstitute.com/acars-explained/
Modern aircraft have a system called ACARS that sends text messages to ground systems.
Text-based communication reduces the risk of misheard calls, especially if the frequency is jam-packed.
Might have been a good option in this case.
To try to attempt to avoid confusion, the international language of air traffic control is regularly conducted in English, but as can be seen in this incident, that doesn’t necessarily mean that people speaking English as a second or even third language might not get confused from time to time.
I think the situation was dire enough here that maybe it's worth trying, as a fallback interchange language, a common ancestor of the Portuguese airline and the French air traffic control
Latin. Maybe that'd clear things up.
“You have to be very careful,” Simone Fischer-Hübner, a computer science researcher at Karlstad University, told Aftonbladet, warning against using ChatGPT to work with sensitive information.
I mean, sending queries to a search engine or an LLM are about the same in terms of exposing one's queries.
If the guy were complaining about information from an LLM not being cited or something, then I think I could see where he was coming from more.
Building on efforts to ensure quality healthcare accessibility for the people of Illinois, Governor JB Pritzker signed legislation on Friday that protects patients by limiting the use of artificial intelligence (AI) in therapy and psychotherapy services.
Poor ELIZA.
https://en.wikipedia.org/wiki/ELIZA
ELIZA is an early natural language processing computer program developed from 1964 to 1967[1] at MIT by Joseph Weizenbaum.
The most famous script, DOCTOR, simulated a psychotherapist of the Rogerian school (in which the therapist often reflects back the patient's words to the patient),[9][10][11] and used rules, dictated in the script, to respond with non-directional questions to user inputs.
Decades of emacs having M-x doctor.
M-x doctor
I am the psychotherapist. Please, describe your problems. Each time you are finished talking,
type RET twice.
Checking git, the initial work on doctor.el appears to precede its initial commit into a version control system in 1991, but the initial copyright message on doctor.el appears to date to 1985.
https://www.armtecdefense.com/Portals/5/Documents/RR170-188_1023.pdf?ver=2024-09-11-165306-580