Not just "no" but, fuck no.
nickwitha_k
Yup. Need something like EV certs to really verify... And that would only make sense if it's a "no (non-real) screennames" kind of thing.
I'd argue that showing disdain, aggression, and disrespect in communication with AI/LLM things is more likely to be dangerous as one is conditioning themselves to be disdainful, aggressive, and disrespectful when communicating with the same methods used to communicate with other people. Our brains do a great job at association, so, it's basically just training oneself to be an asshole.
Well. I certainly didn't have "Murdoch turns on a right-wing billionaire" on my 2025 hellscape bingo card.
Unfortunately, the US administrative state is still being murdered. So, don't expect act regs or enforcement to be seen out of the US and expect flagrant violation of EU and other regs due to mobster-style protectionism.
Right-wingers never agree to annual payouts except for the rich. It's absolutely a one time thing.
I have to agree with pretty much everything that you've said there. Since I don't use CAD professionally, and I'm not about to suffer through the windows experience voluntarily, I'm pretty much such with FreeCAD and (when I get around to it) CADquery. Hopefully more companies will start supporting Linux and free CAD devs from all the MS fuckery - might even get FreeCAD (or a fork) to be more productive and prioritize things necessary to be competitive for SMB/hobbyists.
I hate the syntax in OpenSCAD. It LOOKS like something object-oriented but it is procedural, causing oh so many footguns, if one expects it to act like OOP.
Oh definitely do. The recent improvements (in the last 1-2 years) have made it much more useable, and sometimes even intuitive.
Depends on your needs. I probably wouldn't consider it good enough yet for commercial but the improvements on 1.0 take care of pretty much all of my needs. The "free" licenses for Fusion360 and OnShape are garbage and feel like nothing more than attempts to get hobbyists and small businesses locked in before changing terms. Plus, last I checked, they pull the same kinda data vacuum bullshit that social media companies did in their terms - "free" license holders should expect any and all of their work to be resold by the companies for profit.
Rationally and in vacuum, anthropomorphizing tools and animals is kinda silly and sometimes dangerous. But human brains don't work do well at context separation and rationality. They are very noisy and prone to conceptual cross-talk.
The reason that this is important is that, as useless as LLMs are at nearly everything they are billed as, they are really good at fooling our brains into thinking that they possess consciousness (there's plenty even on Lemmy that ascribe levels of intelligence to them that are impossible with the technology). Just like knowledge and awareness don't grant immunity to propaganda, our unconscious processes will do their own thing. Humans are social animals and our brains are adapted to act as such, resulting in behaviors that run the gamut from wonderfully bizzare (keeping pets that don't "work") to dangerous (attempting to pet bears or keep chimps as "family").
Things that are perceived by our brains, consciously or unconsciously, are stored with associations to other similar things. So the danger here that I was trying to highlight is that being abusive to a tool, like an LLM, that can trick our brains into associating it with conscious beings, is that that acceptability of abusive behavior towards other people can be indirectly reinforced.
Basically, like I said before, one can unintentionally train themselves into practicing antisocial behaviors.
You do have a good point though that people believing that ChatGPT is a being that they can confide in, etc is very harmful and, itself, likely to lead to antisocial behaviors.
It is human behavior. Humans are irrational as fuck, even the most rational of us. It's best to plan accordingly.